mirror of
https://github.com/ARM-software/workload-automation.git
synced 2025-01-18 12:06:08 +00:00
fd8a7e442c
re._pattern_type became re.Pattern in Python 3.
339 lines
12 KiB
Python
339 lines
12 KiB
Python
# Copyright 2015-2018 ARM Limited
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
import re
|
|
import logging
|
|
from itertools import chain
|
|
|
|
from devlib.collector.ftrace import TRACE_MARKER_START, TRACE_MARKER_STOP
|
|
|
|
from wa.utils.misc import isiterable
|
|
from wa.utils.types import numeric
|
|
|
|
|
|
logger = logging.getLogger('trace-cmd')
|
|
|
|
|
|
class TraceCmdEvent(object):
|
|
"""
|
|
A single trace-cmd event. This will appear in the trace cmd report in the format ::
|
|
|
|
<idle>-0 [000] 3284.126993: sched_rq_runnable_load: cpu=0 load=54
|
|
| | | | |___________|
|
|
| | | | |
|
|
thread cpu timestamp name body
|
|
|
|
"""
|
|
|
|
__slots__ = ['thread', 'reporting_cpu_id', 'timestamp', 'name', 'text', 'fields']
|
|
|
|
def __init__(self, thread, cpu_id, ts, name, body, parser=None):
|
|
"""
|
|
parameters:
|
|
|
|
:thread: thread which generated the event
|
|
:cpu: cpu on which the event has occurred
|
|
:ts: timestamp of the event
|
|
:name: the name of the event
|
|
:bodytext: a string with the rest of the event text
|
|
:parser: optionally, a function that will parse body text to populate
|
|
this event's attributes
|
|
|
|
The parser can be any callable that can be invoked with
|
|
|
|
parser(event, text)
|
|
|
|
Where ``event`` is this TraceCmdEvent instance, and ``text`` is the body text to be
|
|
parsed. The parser should updated the passed event instance and not return anything
|
|
(the return value will be ignored). Any exceptions raised by the parser will be silently
|
|
ignored (note that this means that the event's attributes may be partially initialized).
|
|
|
|
"""
|
|
self.thread = thread
|
|
self.reporting_cpu_id = int(cpu_id)
|
|
self.timestamp = numeric(ts)
|
|
self.name = name
|
|
self.text = body
|
|
self.fields = {}
|
|
|
|
if parser:
|
|
try:
|
|
parser(self, self.text)
|
|
except Exception: # pylint: disable=broad-except
|
|
# unknown format assume user does not care or know how to
|
|
# parse self.text
|
|
pass
|
|
|
|
def __getattr__(self, name):
|
|
try:
|
|
return self.fields[name]
|
|
except KeyError:
|
|
raise AttributeError(name)
|
|
|
|
def __str__(self):
|
|
return 'TE({} @ {})'.format(self.name, self.timestamp)
|
|
|
|
__repr__ = __str__
|
|
|
|
|
|
class DroppedEventsEvent(object):
|
|
|
|
__slots__ = ['thread', 'reporting_cpu_id', 'timestamp', 'name', 'text', 'fields']
|
|
|
|
def __init__(self, cpu_id):
|
|
self.thread = None
|
|
self.reporting_cpu_id = None
|
|
self.timestamp = None
|
|
self.name = 'DROPPED EVENTS DETECTED'
|
|
self.text = None
|
|
self.fields = {'cpu_id': int(cpu_id)}
|
|
|
|
def __getattr__(self, name):
|
|
try:
|
|
return self.fields[name]
|
|
except KeyError:
|
|
raise AttributeError(name)
|
|
|
|
def __str__(self):
|
|
return 'DROPPED_EVENTS_ON_CPU{}'.format(self.cpu_id)
|
|
|
|
__repr__ = __str__
|
|
|
|
|
|
def try_convert_to_numeric(v):
|
|
try:
|
|
if isiterable(v):
|
|
return list(map(numeric, v))
|
|
else:
|
|
return numeric(v)
|
|
except ValueError:
|
|
return v
|
|
|
|
|
|
def default_body_parser(event, text):
|
|
"""
|
|
Default parser to attempt to use to parser body text for the event (i.e. after
|
|
the "header" common to all events has been parsed). This assumes that the body is
|
|
a whitespace-separated list of key=value pairs. The parser will attempt to convert
|
|
the value into a numeric type, and failing that, keep it as string.
|
|
|
|
"""
|
|
parts = [e.rsplit(' ', 1) for e in text.strip().split('=')]
|
|
parts = [p.strip() for p in chain.from_iterable(parts)]
|
|
if not len(parts) % 2:
|
|
i = iter(parts)
|
|
for k, v in zip(i, i):
|
|
try:
|
|
v = int(v)
|
|
except ValueError:
|
|
pass
|
|
event.fields[k] = v
|
|
|
|
|
|
def regex_body_parser(regex, flags=0):
|
|
"""
|
|
Creates an event body parser form the specified regular expression (could be an
|
|
``re.RegexObject``, or a string). The regular expression should contain some named
|
|
groups, as those will be extracted as the event attributes (unnamed groups and the
|
|
reset of the match will be ignored).
|
|
|
|
If the specified regex is a string, it will be compiled, in which case ``flags`` may
|
|
be provided for the resulting regex object (see ``re`` standard module documentation).
|
|
If regex is a pre-compiled object, flags will be ignored.
|
|
|
|
"""
|
|
if isinstance(regex, str):
|
|
regex = re.compile(regex, flags)
|
|
|
|
def regex_parser_func(event, text):
|
|
match = regex.search(text)
|
|
if match:
|
|
for k, v in match.groupdict().items():
|
|
try:
|
|
event.fields[k] = int(v)
|
|
except ValueError:
|
|
event.fields[k] = v
|
|
|
|
return regex_parser_func
|
|
|
|
|
|
def sched_switch_parser(event, text):
|
|
"""
|
|
Sched switch output may be presented in a couple of different formats. One is handled
|
|
by a regex. The other format can *almost* be handled by the default parser, if it
|
|
weren't for the ``==>`` that appears in the middle.
|
|
"""
|
|
if text.count('=') == 2: # old format
|
|
regex = re.compile(
|
|
r'(?P<prev_comm>\S.*):(?P<prev_pid>\d+) \[(?P<prev_prio>\d+)\] (?P<status>\S+)'
|
|
r' ==> '
|
|
r'(?P<next_comm>\S.*):(?P<next_pid>\d+) \[(?P<next_prio>\d+)\]'
|
|
)
|
|
parser_func = regex_body_parser(regex)
|
|
return parser_func(event, text)
|
|
else: # there are more than two "=" -- new format
|
|
return default_body_parser(event, text.replace('==>', ''))
|
|
|
|
|
|
def sched_stat_parser(event, text):
|
|
"""
|
|
sched_stat_* events unclude the units, "[ns]", in an otherwise
|
|
regular key=value sequence; so the units need to be stripped out first.
|
|
"""
|
|
return default_body_parser(event, text.replace(' [ns]', ''))
|
|
|
|
|
|
def sched_wakeup_parser(event, text):
|
|
regex = re.compile(r'(?P<comm>\S+):(?P<pid>\d+) \[(?P<prio>\d+)\] success=(?P<success>\d) CPU:(?P<cpu>\d+)')
|
|
parse_func = regex_body_parser(regex)
|
|
return parse_func(event, text)
|
|
|
|
|
|
# Maps event onto the corresponding parser for its body text. A parser may be
|
|
# a callable with signature
|
|
#
|
|
# parser(event, bodytext)
|
|
#
|
|
# a re.RegexObject, or a string (in which case it will be compiled into a
|
|
# regex). In case of a string/regex, its named groups will be used to populate
|
|
# the event's attributes.
|
|
EVENT_PARSER_MAP = {
|
|
'sched_stat_blocked': sched_stat_parser,
|
|
'sched_stat_iowait': sched_stat_parser,
|
|
'sched_stat_runtime': sched_stat_parser,
|
|
'sched_stat_sleep': sched_stat_parser,
|
|
'sched_stat_wait': sched_stat_parser,
|
|
'sched_switch': sched_switch_parser,
|
|
'sched_wakeup': sched_wakeup_parser,
|
|
'sched_wakeup_new': sched_wakeup_parser,
|
|
}
|
|
|
|
TRACE_EVENT_REGEX = re.compile(r'^\s+(?P<thread>\S+.*?\S+)\s+\[(?P<cpu_id>\d+)\]\s+(?P<ts>[\d.]+):\s+'
|
|
r'(?P<name>[^:]+):\s+(?P<body>.*?)\s*$')
|
|
|
|
HEADER_REGEX = re.compile(r'^\s*(?:version|cpus)\s*=\s*([\d.]+)\s*$')
|
|
|
|
DROPPED_EVENTS_REGEX = re.compile(r'CPU:(?P<cpu_id>\d+) \[\d*\s*EVENTS DROPPED\]')
|
|
|
|
EMPTY_CPU_REGEX = re.compile(r'CPU \d+ is empty')
|
|
|
|
|
|
class TraceCmdParser(object):
|
|
"""
|
|
A parser for textual representation of ftrace as reported by trace-cmd
|
|
|
|
"""
|
|
|
|
def __init__(self, filter_markers=True, check_for_markers=True, events=None):
|
|
"""
|
|
Initialize a new trace parser.
|
|
|
|
:param filter_markers: Specifies whether the trace before the start
|
|
marker and after the stop marker should be
|
|
filtered out (so only events between the two
|
|
markers will be reported). This maybe overriden
|
|
based on `check_for_markers` parameter of
|
|
`parse()`
|
|
:param check_for_markers: Check if the start/stop markers are present
|
|
in the trace and ensure that `filter_markers`
|
|
is `False` if they aren't
|
|
:param events: A list of event names to be reported; if not specified,
|
|
all events will be reported.
|
|
|
|
|
|
"""
|
|
self.filter_markers = filter_markers
|
|
self.check_for_markers = check_for_markers
|
|
self.events = events
|
|
|
|
def parse(self, filepath): # pylint: disable=too-many-branches,too-many-locals
|
|
"""
|
|
This is a generator for the trace event stream.
|
|
|
|
:param filepath: The path to the file containg text trace as reported
|
|
by trace-cmd
|
|
"""
|
|
inside_maked_region = False
|
|
# pylint: disable=superfluous-parens
|
|
filters = [re.compile('^{}$'.format(e)) for e in (self.events or [])]
|
|
filter_markers = self.filter_markers
|
|
if filter_markers and self.check_for_markers:
|
|
with open(filepath) as fh:
|
|
for line in fh:
|
|
if TRACE_MARKER_START in line:
|
|
break
|
|
else:
|
|
# maker not found force filtering by marker to False
|
|
filter_markers = False
|
|
|
|
with open(filepath) as fh:
|
|
for line in fh:
|
|
# if processing trace markers, skip marker lines as well as all
|
|
# lines outside marked region
|
|
if filter_markers:
|
|
if not inside_maked_region:
|
|
if TRACE_MARKER_START in line:
|
|
inside_maked_region = True
|
|
continue
|
|
elif TRACE_MARKER_STOP in line:
|
|
inside_maked_region = False
|
|
continue
|
|
|
|
match = DROPPED_EVENTS_REGEX.search(line)
|
|
if match:
|
|
yield DroppedEventsEvent(match.group('cpu_id'))
|
|
continue
|
|
|
|
matched = False
|
|
for rx in [HEADER_REGEX, EMPTY_CPU_REGEX]:
|
|
match = rx.search(line)
|
|
if match:
|
|
logger.debug(line.strip())
|
|
matched = True
|
|
break
|
|
if matched:
|
|
continue
|
|
|
|
match = TRACE_EVENT_REGEX.search(line)
|
|
if not match:
|
|
logger.warning('Invalid trace event: "{}"'.format(line))
|
|
continue
|
|
|
|
event_name = match.group('name')
|
|
|
|
if filters:
|
|
found = False
|
|
for f in filters:
|
|
if f.search(event_name):
|
|
found = True
|
|
break
|
|
if not found:
|
|
continue
|
|
|
|
body_parser = EVENT_PARSER_MAP.get(event_name, default_body_parser)
|
|
if isinstance(body_parser, (str, re.Pattern)): # pylint: disable=protected-access
|
|
body_parser = regex_body_parser(body_parser)
|
|
yield TraceCmdEvent(parser=body_parser, **match.groupdict())
|
|
|
|
|
|
def trace_has_marker(filepath, max_lines_to_check=2000000):
|
|
with open(filepath) as fh:
|
|
for i, line in enumerate(fh):
|
|
if TRACE_MARKER_START in line:
|
|
return True
|
|
if i >= max_lines_to_check:
|
|
break
|
|
return False
|