mirror of
https://github.com/opencv/opencv.git
synced 2024-12-16 10:29:11 +08:00
c0a84dcc85
ts: basic documentation for utility scripts
139 lines
5.7 KiB
Python
Executable File
139 lines
5.7 KiB
Python
Executable File
#!/usr/bin/env python
|
|
""" Print performance test run statistics.
|
|
|
|
Performance data is stored in the GTest log file created by performance tests. Default name is
|
|
`test_details.xml`. It can be changed with the `--gtest_output=xml:<location>/<filename>.xml` test
|
|
option. See https://github.com/opencv/opencv/wiki/HowToUsePerfTests for more details.
|
|
|
|
This script produces configurable performance report tables in text and HTML formats. It allows to
|
|
filter test cases by name and parameter string and select specific performance metrics columns. One
|
|
or multiple test results can be used for input.
|
|
|
|
### Example
|
|
|
|
./report.py -c min,mean,median -f '(LUT|Match).*640' opencv_perf_core.xml opencv_perf_features2d.xml
|
|
|
|
opencv_perf_features2d.xml, opencv_perf_core.xml
|
|
|
|
Name of Test Min Mean Median
|
|
KnnMatch::OCL_BruteForceMatcherFixture::(640x480, 32FC1) 1365.04 ms 1368.18 ms 1368.52 ms
|
|
LUT::OCL_LUTFixture::(640x480, 32FC1) 2.57 ms 2.62 ms 2.64 ms
|
|
LUT::OCL_LUTFixture::(640x480, 32FC4) 21.15 ms 21.25 ms 21.24 ms
|
|
LUT::OCL_LUTFixture::(640x480, 8UC1) 2.22 ms 2.28 ms 2.29 ms
|
|
LUT::OCL_LUTFixture::(640x480, 8UC4) 19.12 ms 19.24 ms 19.19 ms
|
|
LUT::SizePrm::640x480 2.22 ms 2.27 ms 2.29 ms
|
|
Match::OCL_BruteForceMatcherFixture::(640x480, 32FC1) 1364.15 ms 1367.73 ms 1365.45 ms
|
|
RadiusMatch::OCL_BruteForceMatcherFixture::(640x480, 32FC1) 1372.68 ms 1375.52 ms 1375.42 ms
|
|
|
|
### Options
|
|
|
|
-o FMT, --output=FMT - output results in text format (can be 'txt', 'html' or 'auto' - default)
|
|
-u UNITS, --units=UNITS - units for output values (s, ms (default), us, ns or ticks)
|
|
-c COLS, --columns=COLS - comma-separated list of columns to show
|
|
-f REGEX, --filter=REGEX - regex to filter tests
|
|
--show-all - also include empty and "notrun" lines
|
|
"""
|
|
|
|
from __future__ import print_function
|
|
import testlog_parser, sys, os, xml, re, glob
|
|
from table_formatter import *
|
|
from optparse import OptionParser
|
|
|
|
if __name__ == "__main__":
|
|
parser = OptionParser()
|
|
parser.add_option("-o", "--output", dest="format", help="output results in text format (can be 'txt', 'html' or 'auto' - default)", metavar="FMT", default="auto")
|
|
parser.add_option("-u", "--units", dest="units", help="units for output values (s, ms (default), us, ns or ticks)", metavar="UNITS", default="ms")
|
|
parser.add_option("-c", "--columns", dest="columns", help="comma-separated list of columns to show", metavar="COLS", default="")
|
|
parser.add_option("-f", "--filter", dest="filter", help="regex to filter tests", metavar="REGEX", default=None)
|
|
parser.add_option("", "--show-all", action="store_true", dest="showall", default=False, help="also include empty and \"notrun\" lines")
|
|
(options, args) = parser.parse_args()
|
|
|
|
if len(args) < 1:
|
|
print("Usage:\n", os.path.basename(sys.argv[0]), "<log_name1>.xml", file=sys.stderr)
|
|
exit(0)
|
|
|
|
options.generateHtml = detectHtmlOutputType(options.format)
|
|
|
|
# expand wildcards and filter duplicates
|
|
files = []
|
|
files1 = []
|
|
for arg in args:
|
|
if ("*" in arg) or ("?" in arg):
|
|
files1.extend([os.path.abspath(f) for f in glob.glob(arg)])
|
|
else:
|
|
files.append(os.path.abspath(arg))
|
|
seen = set()
|
|
files = [ x for x in files if x not in seen and not seen.add(x)]
|
|
files.extend((set(files1) - set(files)))
|
|
args = files
|
|
|
|
# load test data
|
|
tests = []
|
|
files = []
|
|
for arg in set(args):
|
|
try:
|
|
cases = testlog_parser.parseLogFile(arg)
|
|
if cases:
|
|
files.append(os.path.basename(arg))
|
|
tests.extend(cases)
|
|
except:
|
|
pass
|
|
|
|
if options.filter:
|
|
expr = re.compile(options.filter)
|
|
tests = [t for t in tests if expr.search(str(t))]
|
|
|
|
tbl = table(", ".join(files))
|
|
if options.columns:
|
|
metrics = [s.strip() for s in options.columns.split(",")]
|
|
metrics = [m for m in metrics if m and not m.endswith("%") and m in metrix_table]
|
|
else:
|
|
metrics = None
|
|
if not metrics:
|
|
metrics = ["name", "samples", "outliers", "min", "median", "gmean", "mean", "stddev"]
|
|
if "name" not in metrics:
|
|
metrics.insert(0, "name")
|
|
|
|
for m in metrics:
|
|
if m == "name":
|
|
tbl.newColumn(m, metrix_table[m][0])
|
|
else:
|
|
tbl.newColumn(m, metrix_table[m][0], align = "center")
|
|
|
|
needNewRow = True
|
|
for case in sorted(tests, key=lambda x: str(x)):
|
|
if needNewRow:
|
|
tbl.newRow()
|
|
if not options.showall:
|
|
needNewRow = False
|
|
status = case.get("status")
|
|
if status != "run":
|
|
if status != "notrun":
|
|
needNewRow = True
|
|
for m in metrics:
|
|
if m == "name":
|
|
tbl.newCell(m, str(case))
|
|
else:
|
|
tbl.newCell(m, status, color = "red")
|
|
else:
|
|
needNewRow = True
|
|
for m in metrics:
|
|
val = metrix_table[m][1](case, None, options.units)
|
|
if isinstance(val, float):
|
|
tbl.newCell(m, "%.2f %s" % (val, options.units), val)
|
|
else:
|
|
tbl.newCell(m, val, val)
|
|
if not needNewRow:
|
|
tbl.trimLastRow()
|
|
|
|
# output table
|
|
if options.generateHtml:
|
|
if options.format == "moinwiki":
|
|
tbl.htmlPrintTable(sys.stdout, True)
|
|
else:
|
|
htmlPrintHeader(sys.stdout, "Report %s tests from %s" % (len(tests), ", ".join(files)))
|
|
tbl.htmlPrintTable(sys.stdout)
|
|
htmlPrintFooter(sys.stdout)
|
|
else:
|
|
tbl.consolePrintTable(sys.stdout)
|