1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158
|
from __future__ import print_function
from __future__ import division
import argparse
from enum import Enum
import xml.etree.ElementTree as ET
from collections import OrderedDict
class BenchmarkResult:
"""BenchmarkResult class to hold perf numbers"""
def __init__(self, name, mean_value, mean_lower_bound, mean_upper_bound, \
std_val, std_lower_bound, std_upper_bound):
self.name = name
self.mean_value = mean_value
self.mean_lower_bound = mean_lower_bound
self.mean_upper_bound = mean_upper_bound
self.std_val = std_val
self.std_lower_bound = std_lower_bound
self.std_upper_bound = std_upper_bound
def __str__(self):
return "BenchmarkResults %s\nmean value=%f, lowerBound=%f, upperBound=%f\nstd:%f, lowerBound=%f, upperBound=%f" \
% (self.name, self.mean_value, self.mean_lower_bound, self.mean_upper_bound, \
self.std_val, self.std_lower_bound, self.std_upper_bound)
class TestCase:
"""TestCase class to hold the list of benchmark results"""
def __init__(self, name):
self.name = name
self.results = OrderedDict()
def __str__(self):
return "TestCase %s\nBenchmark result:\n%s" % (self.name, self.results)
class Metric(Enum):
mean = 'mean'
lowMean = 'low_mean'
highMean = 'high_mean'
class Unit(Enum):
nano = 'nano'
micro = 'micro'
milli = 'milli'
sec = 'sec'
def displayUnit(unit):
if unit == Unit.sec:
return "s"
elif unit == Unit.milli:
return "ms"
elif unit == Unit.micro:
return "us"
else:
return "ns"
def nanoToMicro(nano):
return nano / 1000
def nanoToMilli(nano):
# return nano / (1000*1000)
return nano
def nanoToSec(nano):
return nano / (1000*1000*1000)
def convertUnit(nano, unit):
if unit == Unit.sec:
return nanoToSec(nano)
elif unit == Unit.milli:
return nanoToMilli(nano)
elif unit == Unit.micro:
return nanoToMicro(nano)
else:
return nano
parser = argparse.ArgumentParser()
parser.add_argument("--before", help="Path to XML perf log for before comparison.", required=True)
parser.add_argument("--after", help="Path to XML perf log for after comparison.", required=True)
parser.add_argument("--before-label", help="Label for before column.", default='Before')
parser.add_argument("--after-label", help="Label for after column.", default='After')
parser.add_argument("--metric", help="Benchmark metric (mean, low_mean, high_mean).", type=Metric, choices=list(Metric), default=Metric.mean)
parser.add_argument("--unit", help="Benchmark unit (nano, micro, milli, sec).", type=Unit, choices=list(Unit), default=Unit.micro)
args = parser.parse_args()
file_before = args.before
file_after = args.after
unit = args.unit
print('Path to XML perf log for before comparison:', file_before)
print('Path to XML perf log for after comparison:', file_after)
print('Time unit:', unit)
print()
tree_before = ET.parse(file_before)
tree_after = ET.parse(file_after)
root_before = tree_before.getroot()
root_after = tree_after.getroot()
def loadResults(root):
results = OrderedDict()
for test_case in root.iter('TestCase'):
test_name = test_case.attrib['name']
current_test = TestCase(test_name)
for bench_res in test_case.iter('BenchmarkResults'):
bench_name = bench_res.attrib['name']
mean_node = bench_res.find('mean')
mean = float(mean_node.attrib['value'])
mean_lower = float(mean_node.attrib['lowerBound'])
mean_upper = float(mean_node.attrib['upperBound'])
std_node = bench_res.find('standardDeviation')
std = float(std_node.attrib['value'])
std_lower = float(std_node.attrib['lowerBound'])
std_upper = float(std_node.attrib['upperBound'])
current_test.results[bench_name] = BenchmarkResult(bench_name, mean, mean_lower, mean_upper, std, std_lower, std_upper)
results[test_name] = current_test
return results
results_before = loadResults(root_before)
results_after = loadResults(root_after)
for r_before_name, r_before in results_before.items():
title = '{} - time unit: {}'.format(r_before_name, displayUnit(unit))
print('| {} | {} | {} | Speed-up |'.format(title, args.before_label, args.after_label))
print('| --- | --- | --- | --- |')
if r_before_name in results_after:
r_after = results_after[r_before_name]
for b_before_name, b_before in r_before.results.items():
if b_before_name in r_after.results:
b_after = r_after.results[b_before_name]
if args.metric == Metric.lowMean:
metric_before = b_before.mean_lower_bound
metric_after = b_after.mean_lower_bound
elif args.metric == Metric.highMean:
metric_before = b_before.mean_upper_bound
metric_after = b_after.mean_upper_bound
else:
metric_before = b_before.mean_value
metric_after = b_after.mean_value
metric_before = convertUnit(metric_before, unit)
metric_after = convertUnit(metric_after, unit)
speed_up = metric_before / metric_after
print('| {} | {:.2f} | {:.2f} | {:.2f} |'.format(b_before_name, metric_before, metric_after, speed_up))
print()
|