summaryrefslogtreecommitdiffstats
path: root/qtip/runner/runner.py
diff options
context:
space:
mode:
Diffstat (limited to 'qtip/runner/runner.py')
-rw-r--r--qtip/runner/runner.py105
1 files changed, 105 insertions, 0 deletions
diff --git a/qtip/runner/runner.py b/qtip/runner/runner.py
new file mode 100644
index 00000000..47795bc3
--- /dev/null
+++ b/qtip/runner/runner.py
@@ -0,0 +1,105 @@
+##############################################################################
+# Copyright (c) 2017 ZTE corp. and others.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+import argparse
+import json
+import os
+from os import path
+import sys
+import time
+
+from qtip.collector.parser import grep
+from qtip.driver.ansible_driver import AnsibleDriver
+from qtip.util.logger import QtipLogger
+
+logger = QtipLogger('runner').get
+
+ALL_BENCHMARKS = ['dpi', 'ramspeed', 'ssl', 'dhrystone', 'whetstone']
+
+
+def parse_args(args):
+ parser = argparse.ArgumentParser()
+ parser.add_argument('-d', '--dest', required=True,
+ help='the destination where results will be stored.')
+ parser.add_argument('-b', '--benchmark', required=True, action='append',
+ help='the benchmark you want to execute.')
+ return parser.parse_args(args)
+
+
+def run_benchmark(result_dir, benchmarks):
+ if not path.isdir(result_dir):
+ os.makedirs(result_dir)
+ driver = AnsibleDriver({'args': {'result_dir': result_dir}})
+ driver.pre_run()
+ return driver.run(benchmarks)
+
+
+def generate_report(result_dir, start_time, stop_time):
+ output = {
+ "plan_name": "compute_qpi",
+ "start_time": start_time,
+ "stop_time": stop_time,
+ "sut": []
+ }
+ output.update(parse_result(result_dir))
+ output.update({'stop_time': stop_time})
+ with open('{0}/result.json'.format(result_dir), 'w+') as f:
+ json.dump(output, f, indent=4, sort_keys=True)
+
+
+def parse_result(result_dir):
+ sut_template = {'sut': []}
+ nodes_list = os.listdir(result_dir)
+ for node in nodes_list:
+ node_output_template = {
+ 'name': node,
+ 'type': 'baremetal',
+ 'qpis': []
+ }
+ qpi_result = {'name': 'compute_qpi', 'benchmarks': []}
+ for benchmark in os.listdir('{0}/{1}'.format(result_dir, node)):
+ benchmark_result = \
+ grep.parse_benchmark_result(
+ '{0}/{1}/{2}'.format(result_dir, node, benchmark))
+ qpi_result['benchmarks'].append(benchmark_result)
+ node_output_template['qpis'].append(qpi_result)
+ sut_template['sut'].append(node_output_template)
+ return sut_template
+
+
+def main(args=sys.argv[1:]):
+ args = parse_args(args)
+
+ if not path.isdir(str(args.dest)):
+ logger.error("The destination {0} you give doesn't exist. "
+ "Please check!".format(args.dest))
+ sys.exit(1)
+
+ if args.benchmark == ['all']:
+ args.benchmark = ALL_BENCHMARKS
+ elif len(set(args.benchmark).difference(ALL_BENCHMARKS)) != 0:
+ logger.error("Please check benchmarks name. The supported benchmarks are"
+ "{0}".format(ALL_BENCHMARKS))
+ logger.info("Start to run benchmark test: {0}.".format(args.benchmark))
+
+ start_time = time.strftime("%Y-%m-%d-%H-%M")
+ logger.info("start_time: {0}".format(start_time))
+ if not args.dest.endswith('/'):
+ args.dest += '/'
+ result_dir = args.dest + start_time
+ ansible_result = run_benchmark(result_dir, args.benchmark)
+ stop_time = time.strftime("%Y-%m-%d-%H-%M")
+ logger.info("stop_time: {0}".format(stop_time))
+ if not ansible_result:
+ logger.error("Bechmarks run failed. Cann't generate any report.")
+ sys.exit(1)
+ generate_report(result_dir, start_time, stop_time)
+
+
+if __name__ == "__main__":
+ main()