aboutsummaryrefslogtreecommitdiff
path: root/benchtests
diff options
context:
space:
mode:
authorSiddhesh Poyarekar <siddhesh@sourceware.org>2017-08-08 00:55:12 +0530
committerSiddhesh Poyarekar <siddhesh@sourceware.org>2017-08-08 00:55:12 +0530
commitb115e819af637101d9d9b0d26c3685b7236d3fb1 (patch)
tree297a6d9bd37fbd71a302b9121fd1a40e017fabc1 /benchtests
parent7ee38e6040d34bca96ee668efbbd2f56b446319d (diff)
downloadglibc-b115e819af637101d9d9b0d26c3685b7236d3fb1.tar
glibc-b115e819af637101d9d9b0d26c3685b7236d3fb1.tar.gz
glibc-b115e819af637101d9d9b0d26c3685b7236d3fb1.tar.bz2
glibc-b115e819af637101d9d9b0d26c3685b7236d3fb1.zip
benchtests: Allow selecting baseline for compare_string.py
This patch allows one to provide the function name using an optional -base option to compare all other functions against. This is useful when pitching one implementation of a string function against alternatives. In the absence of this option, comparisons are done against the first ifunc in the list. * benchtests/scripts/compare_strings.py (main): Add an optional -base option. (process_results): New argument base_func.
Diffstat (limited to 'benchtests')
-rwxr-xr-xbenchtests/scripts/compare_strings.py28
1 files changed, 18 insertions, 10 deletions
diff --git a/benchtests/scripts/compare_strings.py b/benchtests/scripts/compare_strings.py
index 9d73ec41f4..43e70ebee2 100755
--- a/benchtests/scripts/compare_strings.py
+++ b/benchtests/scripts/compare_strings.py
@@ -21,6 +21,7 @@ Given a string benchmark result file, print a table with comparisons with a
baseline. The baseline is the first function, which typically is the builtin
function.
"""
+
import sys
import os
import json
@@ -74,7 +75,7 @@ def draw_graph(f, v, ifuncs, results):
pylab.savefig('%s-%s.png' % (f, v), bbox_inches='tight')
-def process_results(results, attrs):
+def process_results(results, attrs, base_func):
""" Process results and print them
Args:
@@ -84,6 +85,10 @@ def process_results(results, attrs):
for f in results['functions'].keys():
print('Function: %s' % f)
+ base_index = 0
+ if base_func:
+ base_index = results['functions'][f]['ifuncs'].index(base_func)
+
print('\t'.join(results['functions'][f]['ifuncs']))
v = results['functions'][f]['bench-variant']
print('Variant: %s' % v)
@@ -91,19 +96,17 @@ def process_results(results, attrs):
graph_res = {}
for res in results['functions'][f]['results']:
attr_list = ['%s=%s' % (a, res[a]) for a in attrs]
- first = True
+ i = 0
key = ','.join(attr_list)
sys.stdout.write('%s: \t' % key)
graph_res[key] = res['timings']
for t in res['timings']:
sys.stdout.write ('%.2f' % t)
- if first:
- first = False
- else:
- diff = (res['timings'][0] - t) * 100 / res['timings'][0]
-
+ if i != base_index:
+ diff = (res['timings'][base_index] - t) * 100 / res['timings'][base_index]
sys.stdout.write (' (%.2f%%)' % diff)
sys.stdout.write('\t')
+ i = i + 1
print('')
draw_graph(f, v, results['functions'][f]['ifuncs'], graph_res)
@@ -114,15 +117,20 @@ def main(args):
Take a string benchmark output file and compare timings.
"""
if len(args) < 3:
- print('Usage: %s <input file> <schema file> attr1 [attr2 ...]' % sys.argv[0])
+ print('Usage: %s <input file> <schema file> [-base=ifunc_name] attr1 [attr2 ...]' % sys.argv[0])
sys.exit(os.EX_USAGE)
+ base_func = None
filename = args[0]
schema_filename = args[1]
- attrs = args[2:]
+ if args[2].find('-base=') == 0:
+ base_func = args[2][6:]
+ attrs = args[3:]
+ else:
+ attrs = args[2:]
results = parse_file(filename, schema_filename)
- process_results(results, attrs)
+ process_results(results, attrs, base_func)
if __name__ == '__main__':