|
@@ -0,0 +1,634 @@
|
|
|
+#!/usr/bin/python
|
|
|
+#
|
|
|
+# Cpu task migration overview toy
|
|
|
+#
|
|
|
+# Copyright (C) 2010 Frederic Weisbecker <fweisbec@gmail.com>
|
|
|
+#
|
|
|
+# perf trace event handlers have been generated by perf trace -g python
|
|
|
+#
|
|
|
+# The whole is licensed under the terms of the GNU GPL License version 2
|
|
|
+
|
|
|
+
|
|
|
+try:
|
|
|
+ import wx
|
|
|
+except ImportError:
|
|
|
+ raise ImportError, "You need to install the wxpython lib for this script"
|
|
|
+
|
|
|
+import os
|
|
|
+import sys
|
|
|
+
|
|
|
+from collections import defaultdict
|
|
|
+from UserList import UserList
|
|
|
+
|
|
|
+sys.path.append(os.environ['PERF_EXEC_PATH'] + \
|
|
|
+ '/scripts/python/Perf-Trace-Util/lib/Perf/Trace')
|
|
|
+
|
|
|
+from perf_trace_context import *
|
|
|
+from Core import *
|
|
|
+
|
|
|
+class RootFrame(wx.Frame):
|
|
|
+ def __init__(self, timeslices, parent = None, id = -1, title = "Migration"):
|
|
|
+ wx.Frame.__init__(self, parent, id, title)
|
|
|
+
|
|
|
+ (self.screen_width, self.screen_height) = wx.GetDisplaySize()
|
|
|
+ self.screen_width -= 10
|
|
|
+ self.screen_height -= 10
|
|
|
+ self.zoom = 0.5
|
|
|
+ self.scroll_scale = 20
|
|
|
+ self.timeslices = timeslices
|
|
|
+ (self.ts_start, self.ts_end) = timeslices.interval()
|
|
|
+ self.update_width_virtual()
|
|
|
+
|
|
|
+ # whole window panel
|
|
|
+ self.panel = wx.Panel(self, size=(self.screen_width, self.screen_height))
|
|
|
+
|
|
|
+ # scrollable container
|
|
|
+ self.scroll = wx.ScrolledWindow(self.panel)
|
|
|
+ self.scroll.SetScrollbars(self.scroll_scale, self.scroll_scale, self.width_virtual / self.scroll_scale, 100 / 10)
|
|
|
+ self.scroll.EnableScrolling(True, True)
|
|
|
+ self.scroll.SetFocus()
|
|
|
+
|
|
|
+ # scrollable drawing area
|
|
|
+ self.scroll_panel = wx.Panel(self.scroll, size=(self.screen_width, self.screen_height / 2))
|
|
|
+ self.scroll_panel.Bind(wx.EVT_PAINT, self.on_paint)
|
|
|
+ self.scroll_panel.Bind(wx.EVT_KEY_DOWN, self.on_key_press)
|
|
|
+ self.scroll_panel.Bind(wx.EVT_LEFT_DOWN, self.on_mouse_down)
|
|
|
+ self.scroll.Bind(wx.EVT_PAINT, self.on_paint)
|
|
|
+
|
|
|
+ self.scroll.Fit()
|
|
|
+ self.Fit()
|
|
|
+
|
|
|
+ self.scroll_panel.SetDimensions(-1, -1, self.width_virtual, -1, wx.SIZE_USE_EXISTING)
|
|
|
+
|
|
|
+ self.max_cpu = -1
|
|
|
+ self.txt = None
|
|
|
+
|
|
|
+ self.Show(True)
|
|
|
+
|
|
|
+ def us_to_px(self, val):
|
|
|
+ return val / (10 ** 3) * self.zoom
|
|
|
+
|
|
|
+ def px_to_us(self, val):
|
|
|
+ return (val / self.zoom) * (10 ** 3)
|
|
|
+
|
|
|
+ def scroll_start(self):
|
|
|
+ (x, y) = self.scroll.GetViewStart()
|
|
|
+ return (x * self.scroll_scale, y * self.scroll_scale)
|
|
|
+
|
|
|
+ def scroll_start_us(self):
|
|
|
+ (x, y) = self.scroll_start()
|
|
|
+ return self.px_to_us(x)
|
|
|
+
|
|
|
+ def update_rectangle_cpu(self, dc, slice, cpu, offset_time):
|
|
|
+ rq = slice.rqs[cpu]
|
|
|
+
|
|
|
+ if slice.total_load != 0:
|
|
|
+ load_rate = rq.load() / float(slice.total_load)
|
|
|
+ else:
|
|
|
+ load_rate = 0
|
|
|
+
|
|
|
+
|
|
|
+ offset_px = self.us_to_px(slice.start - offset_time)
|
|
|
+ width_px = self.us_to_px(slice.end - slice.start)
|
|
|
+ (x, y) = self.scroll_start()
|
|
|
+
|
|
|
+ if width_px == 0:
|
|
|
+ return
|
|
|
+
|
|
|
+ offset_py = 100 + (cpu * 150)
|
|
|
+ width_py = 100
|
|
|
+
|
|
|
+ if cpu in slice.event_cpus:
|
|
|
+ rgb = rq.event.color()
|
|
|
+ if rgb is not None:
|
|
|
+ (r, g, b) = rgb
|
|
|
+ color = wx.Colour(r, g, b)
|
|
|
+ brush = wx.Brush(color, wx.SOLID)
|
|
|
+ dc.SetBrush(brush)
|
|
|
+ dc.DrawRectangle(offset_px, offset_py, width_px, 5)
|
|
|
+ width_py -= 5
|
|
|
+ offset_py += 5
|
|
|
+
|
|
|
+ red_power = int(0xff - (0xff * load_rate))
|
|
|
+ color = wx.Colour(0xff, red_power, red_power)
|
|
|
+ brush = wx.Brush(color, wx.SOLID)
|
|
|
+ dc.SetBrush(brush)
|
|
|
+ dc.DrawRectangle(offset_px, offset_py, width_px, width_py)
|
|
|
+
|
|
|
+ def update_rectangles(self, dc, start, end):
|
|
|
+ if len(self.timeslices) == 0:
|
|
|
+ return
|
|
|
+ start += self.timeslices[0].start
|
|
|
+ end += self.timeslices[0].start
|
|
|
+
|
|
|
+ color = wx.Colour(0, 0, 0)
|
|
|
+ brush = wx.Brush(color, wx.SOLID)
|
|
|
+ dc.SetBrush(brush)
|
|
|
+
|
|
|
+ i = self.timeslices.find_time_slice(start)
|
|
|
+ if i == -1:
|
|
|
+ return
|
|
|
+
|
|
|
+ for i in xrange(i, len(self.timeslices)):
|
|
|
+ timeslice = self.timeslices[i]
|
|
|
+ if timeslice.start > end:
|
|
|
+ return
|
|
|
+
|
|
|
+ for cpu in timeslice.rqs:
|
|
|
+ self.update_rectangle_cpu(dc, timeslice, cpu, self.timeslices[0].start)
|
|
|
+ if cpu > self.max_cpu:
|
|
|
+ self.max_cpu = cpu
|
|
|
+
|
|
|
+ def on_paint(self, event):
|
|
|
+ color = wx.Colour(0xff, 0xff, 0xff)
|
|
|
+ brush = wx.Brush(color, wx.SOLID)
|
|
|
+ dc = wx.PaintDC(self.scroll_panel)
|
|
|
+ dc.SetBrush(brush)
|
|
|
+
|
|
|
+ width = min(self.width_virtual, self.screen_width)
|
|
|
+ (x, y) = self.scroll_start()
|
|
|
+ start = self.px_to_us(x)
|
|
|
+ end = self.px_to_us(x + width)
|
|
|
+ self.update_rectangles(dc, start, end)
|
|
|
+
|
|
|
+ def cpu_from_ypixel(self, y):
|
|
|
+ y -= 100
|
|
|
+ cpu = y / 150
|
|
|
+ height = y % 150
|
|
|
+
|
|
|
+ if cpu < 0 or cpu > self.max_cpu or height > 100:
|
|
|
+ return -1
|
|
|
+
|
|
|
+ return cpu
|
|
|
+
|
|
|
+ def update_summary(self, cpu, t):
|
|
|
+ idx = self.timeslices.find_time_slice(t)
|
|
|
+ if idx == -1:
|
|
|
+ return
|
|
|
+
|
|
|
+ ts = self.timeslices[idx]
|
|
|
+ rq = ts.rqs[cpu]
|
|
|
+ raw = "CPU: %d\n" % cpu
|
|
|
+ raw += "Last event : %s\n" % rq.event.__repr__()
|
|
|
+ raw += "Timestamp : %d.%06d\n" % (ts.start / (10 ** 9), (ts.start % (10 ** 9)) / 1000)
|
|
|
+ raw += "Duration : %6d us\n" % ((ts.end - ts.start) / (10 ** 6))
|
|
|
+ raw += "Load = %d\n" % rq.load()
|
|
|
+ for t in rq.tasks:
|
|
|
+ raw += "%s \n" % thread_name(t)
|
|
|
+
|
|
|
+ if self.txt:
|
|
|
+ self.txt.Destroy()
|
|
|
+ self.txt = wx.StaticText(self.panel, -1, raw, (0, (self.screen_height / 2) + 50))
|
|
|
+
|
|
|
+
|
|
|
+ def on_mouse_down(self, event):
|
|
|
+ (x, y) = event.GetPositionTuple()
|
|
|
+ cpu = self.cpu_from_ypixel(y)
|
|
|
+ if cpu == -1:
|
|
|
+ return
|
|
|
+
|
|
|
+ t = self.px_to_us(x) + self.timeslices[0].start
|
|
|
+
|
|
|
+ self.update_summary(cpu, t)
|
|
|
+
|
|
|
+
|
|
|
+ def update_width_virtual(self):
|
|
|
+ self.width_virtual = self.us_to_px(self.ts_end - self.ts_start)
|
|
|
+
|
|
|
+ def __zoom(self, x):
|
|
|
+ self.update_width_virtual()
|
|
|
+ (xpos, ypos) = self.scroll.GetViewStart()
|
|
|
+ xpos = self.us_to_px(x) / self.scroll_scale
|
|
|
+ self.scroll.SetScrollbars(self.scroll_scale, self.scroll_scale, self.width_virtual / self.scroll_scale, 100 / 10, xpos, ypos)
|
|
|
+ self.Refresh()
|
|
|
+
|
|
|
+ def zoom_in(self):
|
|
|
+ x = self.scroll_start_us()
|
|
|
+ self.zoom *= 2
|
|
|
+ self.__zoom(x)
|
|
|
+
|
|
|
+ def zoom_out(self):
|
|
|
+ x = self.scroll_start_us()
|
|
|
+ self.zoom /= 2
|
|
|
+ self.__zoom(x)
|
|
|
+
|
|
|
+
|
|
|
+ def on_key_press(self, event):
|
|
|
+ key = event.GetRawKeyCode()
|
|
|
+ if key == ord("+"):
|
|
|
+ self.zoom_in()
|
|
|
+ return
|
|
|
+ if key == ord("-"):
|
|
|
+ self.zoom_out()
|
|
|
+ return
|
|
|
+
|
|
|
+ key = event.GetKeyCode()
|
|
|
+ (x, y) = self.scroll.GetViewStart()
|
|
|
+ if key == wx.WXK_RIGHT:
|
|
|
+ self.scroll.Scroll(x + 1, y)
|
|
|
+ elif key == wx.WXK_LEFT:
|
|
|
+ self.scroll.Scroll(x -1, y)
|
|
|
+
|
|
|
+
|
|
|
+threads = { 0 : "idle"}
|
|
|
+
|
|
|
+def thread_name(pid):
|
|
|
+ return "%s:%d" % (threads[pid], pid)
|
|
|
+
|
|
|
+class EventHeaders:
|
|
|
+ def __init__(self, common_cpu, common_secs, common_nsecs,
|
|
|
+ common_pid, common_comm):
|
|
|
+ self.cpu = common_cpu
|
|
|
+ self.secs = common_secs
|
|
|
+ self.nsecs = common_nsecs
|
|
|
+ self.pid = common_pid
|
|
|
+ self.comm = common_comm
|
|
|
+
|
|
|
+ def ts(self):
|
|
|
+ return (self.secs * (10 ** 9)) + self.nsecs
|
|
|
+
|
|
|
+ def ts_format(self):
|
|
|
+ return "%d.%d" % (self.secs, int(self.nsecs / 1000))
|
|
|
+
|
|
|
+
|
|
|
+def taskState(state):
|
|
|
+ states = {
|
|
|
+ 0 : "R",
|
|
|
+ 1 : "S",
|
|
|
+ 2 : "D",
|
|
|
+ 64: "DEAD"
|
|
|
+ }
|
|
|
+
|
|
|
+ if state not in states:
|
|
|
+ print "Unhandled task state %d" % state
|
|
|
+ return ""
|
|
|
+
|
|
|
+ return states[state]
|
|
|
+
|
|
|
+
|
|
|
+class RunqueueEventUnknown:
|
|
|
+ @staticmethod
|
|
|
+ def color():
|
|
|
+ return None
|
|
|
+
|
|
|
+ def __repr__(self):
|
|
|
+ return "unknown"
|
|
|
+
|
|
|
+class RunqueueEventSleep:
|
|
|
+ @staticmethod
|
|
|
+ def color():
|
|
|
+ return (0, 0, 0xff)
|
|
|
+
|
|
|
+ def __init__(self, sleeper):
|
|
|
+ self.sleeper = sleeper
|
|
|
+
|
|
|
+ def __repr__(self):
|
|
|
+ return "%s gone to sleep" % thread_name(self.sleeper)
|
|
|
+
|
|
|
+class RunqueueEventWakeup:
|
|
|
+ @staticmethod
|
|
|
+ def color():
|
|
|
+ return (0xff, 0xff, 0)
|
|
|
+
|
|
|
+ def __init__(self, wakee):
|
|
|
+ self.wakee = wakee
|
|
|
+
|
|
|
+ def __repr__(self):
|
|
|
+ return "%s woke up" % thread_name(self.wakee)
|
|
|
+
|
|
|
+class RunqueueEventFork:
|
|
|
+ @staticmethod
|
|
|
+ def color():
|
|
|
+ return (0, 0xff, 0)
|
|
|
+
|
|
|
+ def __init__(self, child):
|
|
|
+ self.child = child
|
|
|
+
|
|
|
+ def __repr__(self):
|
|
|
+ return "new forked task %s" % thread_name(self.child)
|
|
|
+
|
|
|
+class RunqueueMigrateIn:
|
|
|
+ @staticmethod
|
|
|
+ def color():
|
|
|
+ return (0, 0xf0, 0xff)
|
|
|
+
|
|
|
+ def __init__(self, new):
|
|
|
+ self.new = new
|
|
|
+
|
|
|
+ def __repr__(self):
|
|
|
+ return "task migrated in %s" % thread_name(self.new)
|
|
|
+
|
|
|
+class RunqueueMigrateOut:
|
|
|
+ @staticmethod
|
|
|
+ def color():
|
|
|
+ return (0xff, 0, 0xff)
|
|
|
+
|
|
|
+ def __init__(self, old):
|
|
|
+ self.old = old
|
|
|
+
|
|
|
+ def __repr__(self):
|
|
|
+ return "task migrated out %s" % thread_name(self.old)
|
|
|
+
|
|
|
+class RunqueueSnapshot:
|
|
|
+ def __init__(self, tasks = [0], event = RunqueueEventUnknown()):
|
|
|
+ self.tasks = tuple(tasks)
|
|
|
+ self.event = event
|
|
|
+
|
|
|
+ def sched_switch(self, prev, prev_state, next):
|
|
|
+ event = RunqueueEventUnknown()
|
|
|
+
|
|
|
+ if taskState(prev_state) == "R" and next in self.tasks \
|
|
|
+ and prev in self.tasks:
|
|
|
+ return self
|
|
|
+
|
|
|
+ if taskState(prev_state) != "R":
|
|
|
+ event = RunqueueEventSleep(prev)
|
|
|
+
|
|
|
+ next_tasks = list(self.tasks[:])
|
|
|
+ if prev in self.tasks:
|
|
|
+ if taskState(prev_state) != "R":
|
|
|
+ next_tasks.remove(prev)
|
|
|
+ elif taskState(prev_state) == "R":
|
|
|
+ next_tasks.append(prev)
|
|
|
+
|
|
|
+ if next not in next_tasks:
|
|
|
+ next_tasks.append(next)
|
|
|
+
|
|
|
+ return RunqueueSnapshot(next_tasks, event)
|
|
|
+
|
|
|
+ def migrate_out(self, old):
|
|
|
+ if old not in self.tasks:
|
|
|
+ return self
|
|
|
+ next_tasks = [task for task in self.tasks if task != old]
|
|
|
+
|
|
|
+ return RunqueueSnapshot(next_tasks, RunqueueMigrateOut(old))
|
|
|
+
|
|
|
+ def __migrate_in(self, new, event):
|
|
|
+ if new in self.tasks:
|
|
|
+ self.event = event
|
|
|
+ return self
|
|
|
+ next_tasks = self.tasks[:] + tuple([new])
|
|
|
+
|
|
|
+ return RunqueueSnapshot(next_tasks, event)
|
|
|
+
|
|
|
+ def migrate_in(self, new):
|
|
|
+ return self.__migrate_in(new, RunqueueMigrateIn(new))
|
|
|
+
|
|
|
+ def wake_up(self, new):
|
|
|
+ return self.__migrate_in(new, RunqueueEventWakeup(new))
|
|
|
+
|
|
|
+ def wake_up_new(self, new):
|
|
|
+ return self.__migrate_in(new, RunqueueEventFork(new))
|
|
|
+
|
|
|
+ def load(self):
|
|
|
+ """ Provide the number of tasks on the runqueue.
|
|
|
+ Don't count idle"""
|
|
|
+ return len(self.tasks) - 1
|
|
|
+
|
|
|
+ def __repr__(self):
|
|
|
+ ret = self.tasks.__repr__()
|
|
|
+ ret += self.origin_tostring()
|
|
|
+
|
|
|
+ return ret
|
|
|
+
|
|
|
+class TimeSlice:
|
|
|
+ def __init__(self, start, prev):
|
|
|
+ self.start = start
|
|
|
+ self.prev = prev
|
|
|
+ self.end = start
|
|
|
+ # cpus that triggered the event
|
|
|
+ self.event_cpus = []
|
|
|
+ if prev is not None:
|
|
|
+ self.total_load = prev.total_load
|
|
|
+ self.rqs = prev.rqs.copy()
|
|
|
+ else:
|
|
|
+ self.rqs = defaultdict(RunqueueSnapshot)
|
|
|
+ self.total_load = 0
|
|
|
+
|
|
|
+ def __update_total_load(self, old_rq, new_rq):
|
|
|
+ diff = new_rq.load() - old_rq.load()
|
|
|
+ self.total_load += diff
|
|
|
+
|
|
|
+ def sched_switch(self, ts_list, prev, prev_state, next, cpu):
|
|
|
+ old_rq = self.prev.rqs[cpu]
|
|
|
+ new_rq = old_rq.sched_switch(prev, prev_state, next)
|
|
|
+
|
|
|
+ if old_rq is new_rq:
|
|
|
+ return
|
|
|
+
|
|
|
+ self.rqs[cpu] = new_rq
|
|
|
+ self.__update_total_load(old_rq, new_rq)
|
|
|
+ ts_list.append(self)
|
|
|
+ self.event_cpus = [cpu]
|
|
|
+
|
|
|
+ def migrate(self, ts_list, new, old_cpu, new_cpu):
|
|
|
+ if old_cpu == new_cpu:
|
|
|
+ return
|
|
|
+ old_rq = self.prev.rqs[old_cpu]
|
|
|
+ out_rq = old_rq.migrate_out(new)
|
|
|
+ self.rqs[old_cpu] = out_rq
|
|
|
+ self.__update_total_load(old_rq, out_rq)
|
|
|
+
|
|
|
+ new_rq = self.prev.rqs[new_cpu]
|
|
|
+ in_rq = new_rq.migrate_in(new)
|
|
|
+ self.rqs[new_cpu] = in_rq
|
|
|
+ self.__update_total_load(new_rq, in_rq)
|
|
|
+
|
|
|
+ ts_list.append(self)
|
|
|
+ self.event_cpus = [old_cpu, new_cpu]
|
|
|
+
|
|
|
+ def wake_up(self, ts_list, pid, cpu, fork):
|
|
|
+ old_rq = self.prev.rqs[cpu]
|
|
|
+ if fork:
|
|
|
+ new_rq = old_rq.wake_up_new(pid)
|
|
|
+ else:
|
|
|
+ new_rq = old_rq.wake_up(pid)
|
|
|
+
|
|
|
+ if new_rq is old_rq:
|
|
|
+ return
|
|
|
+ self.rqs[cpu] = new_rq
|
|
|
+ self.__update_total_load(old_rq, new_rq)
|
|
|
+ ts_list.append(self)
|
|
|
+ self.event_cpus = [cpu]
|
|
|
+
|
|
|
+ def next(self, t):
|
|
|
+ self.end = t
|
|
|
+ return TimeSlice(t, self)
|
|
|
+
|
|
|
+class TimeSliceList(UserList):
|
|
|
+ def __init__(self, arg = []):
|
|
|
+ self.data = arg
|
|
|
+
|
|
|
+ def get_time_slice(self, ts):
|
|
|
+ if len(self.data) == 0:
|
|
|
+ slice = TimeSlice(ts, TimeSlice(-1, None))
|
|
|
+ else:
|
|
|
+ slice = self.data[-1].next(ts)
|
|
|
+ return slice
|
|
|
+
|
|
|
+ def find_time_slice(self, ts):
|
|
|
+ start = 0
|
|
|
+ end = len(self.data)
|
|
|
+ found = -1
|
|
|
+ searching = True
|
|
|
+ while searching:
|
|
|
+ if start == end or start == end - 1:
|
|
|
+ searching = False
|
|
|
+
|
|
|
+ i = (end + start) / 2
|
|
|
+ if self.data[i].start <= ts and self.data[i].end >= ts:
|
|
|
+ found = i
|
|
|
+ end = i
|
|
|
+ continue
|
|
|
+
|
|
|
+ if self.data[i].end < ts:
|
|
|
+ start = i
|
|
|
+
|
|
|
+ elif self.data[i].start > ts:
|
|
|
+ end = i
|
|
|
+
|
|
|
+ return found
|
|
|
+
|
|
|
+ def interval(self):
|
|
|
+ if len(self.data) == 0:
|
|
|
+ return (0, 0)
|
|
|
+
|
|
|
+ return (self.data[0].start, self.data[-1].end)
|
|
|
+
|
|
|
+
|
|
|
+class SchedEventProxy:
|
|
|
+ def __init__(self):
|
|
|
+ self.current_tsk = defaultdict(lambda : -1)
|
|
|
+ self.timeslices = TimeSliceList()
|
|
|
+
|
|
|
+ def sched_switch(self, headers, prev_comm, prev_pid, prev_prio, prev_state,
|
|
|
+ next_comm, next_pid, next_prio):
|
|
|
+ """ Ensure the task we sched out this cpu is really the one
|
|
|
+ we logged. Otherwise we may have missed traces """
|
|
|
+
|
|
|
+ on_cpu_task = self.current_tsk[headers.cpu]
|
|
|
+
|
|
|
+ if on_cpu_task != -1 and on_cpu_task != prev_pid:
|
|
|
+ print "Sched switch event rejected ts: %s cpu: %d prev: %s(%d) next: %s(%d)" % \
|
|
|
+ (headers.ts_format(), headers.cpu, prev_comm, prev_pid, next_comm, next_pid)
|
|
|
+
|
|
|
+ threads[prev_pid] = prev_comm
|
|
|
+ threads[next_pid] = next_comm
|
|
|
+ self.current_tsk[headers.cpu] = next_pid
|
|
|
+
|
|
|
+ ts = self.timeslices.get_time_slice(headers.ts())
|
|
|
+ ts.sched_switch(self.timeslices, prev_pid, prev_state, next_pid, headers.cpu)
|
|
|
+
|
|
|
+ def migrate(self, headers, pid, prio, orig_cpu, dest_cpu):
|
|
|
+ ts = self.timeslices.get_time_slice(headers.ts())
|
|
|
+ ts.migrate(self.timeslices, pid, orig_cpu, dest_cpu)
|
|
|
+
|
|
|
+ def wake_up(self, headers, comm, pid, success, target_cpu, fork):
|
|
|
+ if success == 0:
|
|
|
+ return
|
|
|
+ ts = self.timeslices.get_time_slice(headers.ts())
|
|
|
+ ts.wake_up(self.timeslices, pid, target_cpu, fork)
|
|
|
+
|
|
|
+
|
|
|
+def trace_begin():
|
|
|
+ global parser
|
|
|
+ parser = SchedEventProxy()
|
|
|
+
|
|
|
+def trace_end():
|
|
|
+ app = wx.App(False)
|
|
|
+ timeslices = parser.timeslices
|
|
|
+ frame = RootFrame(timeslices)
|
|
|
+ app.MainLoop()
|
|
|
+
|
|
|
+def sched__sched_stat_runtime(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, runtime, vruntime):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_stat_iowait(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, delay):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_stat_sleep(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, delay):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_stat_wait(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, delay):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_process_fork(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ parent_comm, parent_pid, child_comm, child_pid):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_process_wait(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, prio):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_process_exit(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, prio):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_process_free(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, prio):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_migrate_task(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, prio, orig_cpu,
|
|
|
+ dest_cpu):
|
|
|
+ headers = EventHeaders(common_cpu, common_secs, common_nsecs,
|
|
|
+ common_pid, common_comm)
|
|
|
+ parser.migrate(headers, pid, prio, orig_cpu, dest_cpu)
|
|
|
+
|
|
|
+def sched__sched_switch(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ prev_comm, prev_pid, prev_prio, prev_state,
|
|
|
+ next_comm, next_pid, next_prio):
|
|
|
+
|
|
|
+ headers = EventHeaders(common_cpu, common_secs, common_nsecs,
|
|
|
+ common_pid, common_comm)
|
|
|
+ parser.sched_switch(headers, prev_comm, prev_pid, prev_prio, prev_state,
|
|
|
+ next_comm, next_pid, next_prio)
|
|
|
+
|
|
|
+def sched__sched_wakeup_new(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, prio, success,
|
|
|
+ target_cpu):
|
|
|
+ headers = EventHeaders(common_cpu, common_secs, common_nsecs,
|
|
|
+ common_pid, common_comm)
|
|
|
+ parser.wake_up(headers, comm, pid, success, target_cpu, 1)
|
|
|
+
|
|
|
+def sched__sched_wakeup(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, prio, success,
|
|
|
+ target_cpu):
|
|
|
+ headers = EventHeaders(common_cpu, common_secs, common_nsecs,
|
|
|
+ common_pid, common_comm)
|
|
|
+ parser.wake_up(headers, comm, pid, success, target_cpu, 0)
|
|
|
+
|
|
|
+def sched__sched_wait_task(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid, prio):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_kthread_stop_ret(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ ret):
|
|
|
+ pass
|
|
|
+
|
|
|
+def sched__sched_kthread_stop(event_name, context, common_cpu,
|
|
|
+ common_secs, common_nsecs, common_pid, common_comm,
|
|
|
+ comm, pid):
|
|
|
+ pass
|
|
|
+
|
|
|
+def trace_unhandled(event_name, context, common_cpu, common_secs, common_nsecs,
|
|
|
+ common_pid, common_comm):
|
|
|
+ pass
|