diff options
-rw-r--r-- | samples/unixbench.yaml | 35 | ||||
-rw-r--r-- | tests/unit/benchmark/scenarios/compute/test_unixbench.py | 169 | ||||
-rwxr-xr-x | tools/ubuntu-server-cloudimg-modify.sh | 7 | ||||
-rw-r--r-- | yardstick/benchmark/scenarios/compute/unixbench.py | 156 | ||||
-rw-r--r-- | yardstick/benchmark/scenarios/compute/unixbench_benchmark.bash | 46 |
5 files changed, 413 insertions, 0 deletions
diff --git a/samples/unixbench.yaml b/samples/unixbench.yaml new file mode 100644 index 000000000..9af032f23 --- /dev/null +++ b/samples/unixbench.yaml @@ -0,0 +1,35 @@ +--- +# Sample benchmark task config file +# measure CPU performance +# There is one sample scenario for Dhrystone +# Dhrystone (MIPS) - higher results are better, i.e. better integer performance. + +schema: "yardstick:task:0.1" + +scenarios: +- + type: UnixBench + options: + run_mode: 'verbose' + test_type: 'dhry2reg' + host: Chang'e.demo + + runner: + type: Iteration + iterations: 1 + interval: 1 + +context: + name: demo + image: yardstick-trusty-server + flavor: yardstick-flavor + user: ec2-user + + servers: + Chang'e: + floating_ip: true + + networks: + test: + cidr: '10.0.1.0/24' + diff --git a/tests/unit/benchmark/scenarios/compute/test_unixbench.py b/tests/unit/benchmark/scenarios/compute/test_unixbench.py new file mode 100644 index 000000000..0935bcad2 --- /dev/null +++ b/tests/unit/benchmark/scenarios/compute/test_unixbench.py @@ -0,0 +1,169 @@ +#!/usr/bin/env python + +############################################################################## +# Copyright (c) 2015 Huawei Technologies Co.,Ltd and other. +# +# All rights reserved. This program and the accompanying materials +# are made available under the terms of the Apache License, Version 2.0 +# which accompanies this distribution, and is available at +# http://www.apache.org/licenses/LICENSE-2.0 +############################################################################## + +# Unittest for yardstick.benchmark.scenarios.compute.unixbench.Unixbench + +import mock +import unittest +import json + +from yardstick.benchmark.scenarios.compute import unixbench + + +@mock.patch('yardstick.benchmark.scenarios.compute.unixbench.ssh') +class UnixbenchTestCase(unittest.TestCase): + + def setUp(self): + self.ctx = { + "host": { + "ip": "192.168.50.28", + "user": "root", + "key_filename": "mykey.key" + } + } + + def test_unixbench_successful_setup(self, mock_ssh): + + u = unixbench.Unixbench({}, self.ctx) + u.setup() + + mock_ssh.SSH().execute.return_value = (0, '', '') + self.assertIsNotNone(u.client) + self.assertEqual(u.setup_done, True) + + def test_unixbench_successful_no_sla(self, mock_ssh): + + options = { + "test_type": 'dhry2reg', + "run_mode": 'verbose' + } + args = { + "options": options, + } + u = unixbench.Unixbench(args, self.ctx) + result = {} + + u.server = mock_ssh.SSH() + + sample_output = '{"Score":"4425.4"}' + mock_ssh.SSH().execute.return_value = (0, sample_output, '') + + u.run(result) + expected_result = json.loads(sample_output) + self.assertEqual(result, expected_result) + + def test_unixbench_successful_in_quiet_mode(self, mock_ssh): + + options = { + "test_type": 'dhry2reg', + "run_mode": 'quiet', + "copies":1 + } + args = { + "options": options, + } + u = unixbench.Unixbench(args, self.ctx) + result = {} + + u.server = mock_ssh.SSH() + + sample_output = '{"Score":"4425.4"}' + mock_ssh.SSH().execute.return_value = (0, sample_output, '') + + u.run(result) + expected_result = json.loads(sample_output) + self.assertEqual(result, expected_result) + + + def test_unixbench_successful_sla(self, mock_ssh): + + options = { + "test_type": 'dhry2reg', + "run_mode": 'verbose' + } + sla = { + "single_score": '100', + "parallel_score": '500' + } + args = { + "options": options, + "sla": sla + } + u = unixbench.Unixbench(args, self.ctx) + result = {} + + u.server = mock_ssh.SSH() + + sample_output = '{"signle_score":"2251.7","parallel_score":"4395.9"}' + mock_ssh.SSH().execute.return_value = (0, sample_output, '') + + u.run(result) + expected_result = json.loads(sample_output) + self.assertEqual(result, expected_result) + + def test_unixbench_unsuccessful_sla_single_score(self, mock_ssh): + + args = { + "options": {}, + "sla": {"single_score": "500"} + } + u = unixbench.Unixbench(args, self.ctx) + result = {} + + u.server = mock_ssh.SSH() + sample_output = '{"single_score":"200.7","parallel_score":"4395.9"}' + + mock_ssh.SSH().execute.return_value = (0, sample_output, '') + self.assertRaises(AssertionError, u.run, result) + + def test_unixbench_unsuccessful_sla_parallel_score(self, mock_ssh): + + args = { + "options": {}, + "sla": {"parallel_score": "4000"} + } + u = unixbench.Unixbench(args, self.ctx) + result = {} + + u.server = mock_ssh.SSH() + sample_output = '{"signle_score":"2251.7","parallel_score":"3395.9"}' + + mock_ssh.SSH().execute.return_value = (0, sample_output, '') + self.assertRaises(AssertionError, u.run, result) + + def test_unixbench_unsuccessful_script_error(self, mock_ssh): + + options = { + "test_type": 'dhry2reg', + "run_mode": 'verbose' + } + sla = { + "single_score": '100', + "parallel_score": '500' + } + args = { + "options": options, + "sla": sla + } + u = unixbench.Unixbench(args, self.ctx) + result = {} + + u.server = mock_ssh.SSH() + + mock_ssh.SSH().execute.return_value = (1, '', 'FOOBAR') + self.assertRaises(RuntimeError, u.run, result) + + +def main(): + unittest.main() + +if __name__ == '__main__': + main() diff --git a/tools/ubuntu-server-cloudimg-modify.sh b/tools/ubuntu-server-cloudimg-modify.sh index 58fcd9277..11e6051cd 100755 --- a/tools/ubuntu-server-cloudimg-modify.sh +++ b/tools/ubuntu-server-cloudimg-modify.sh @@ -42,14 +42,21 @@ EOF apt-get update apt-get install -y \ fio \ + git \ + gcc \ iperf3 \ linux-tools-common \ linux-tools-generic \ lmbench \ + make \ netperf \ + patch \ + perl \ rt-tests \ stress \ sysstat +git clone https://github.com/kdlucas/byte-unixbench.git /opt/tempT +make --directory /opt/tempT/UnixBench/ # restore symlink ln -sf /run/resolvconf/resolv.conf /etc/resolv.conf diff --git a/yardstick/benchmark/scenarios/compute/unixbench.py b/yardstick/benchmark/scenarios/compute/unixbench.py new file mode 100644 index 000000000..e6318b92e --- /dev/null +++ b/yardstick/benchmark/scenarios/compute/unixbench.py @@ -0,0 +1,156 @@ +############################################################################## +# Copyright (c) 2015 Huawei Technologies Co.,Ltd and other. +# +# All rights reserved. This program and the accompanying materials +# are made available under the terms of the Apache License, Version 2.0 +# which accompanies this distribution, and is available at +# http://www.apache.org/licenses/LICENSE-2.0 +############################################################################## +import pkg_resources +import logging +import json + +import yardstick.ssh as ssh +from yardstick.benchmark.scenarios import base + +LOG = logging.getLogger(__name__) + + +class Unixbench(base.Scenario): + """Execute Unixbench cpu benchmark in a host + The Run script takes a number of options which you can use to customise a + test, and you can specify the names of the tests to run. The full usage + is: + + Run [ -q | -v ] [-i <n> ] [-c <n> [-c <n> ...]] [test ...] + + -i <count> Run <count> iterations for each test -- slower tests + use <count> / 3, but at least 1. Defaults to 10 (3 for + slow tests). + -c <n> Run <n> copies of each test in parallel. + + Parameters for setting unixbench + run_mode - Run in quiet mode or verbose mode + type: string + unit: None + default: None + test_type - The available tests are organised into categories; + type: string + unit: None + default: None + iterations - Run <count> iterations for each test -- slower tests + use <count> / 3, but at least 1. Defaults to 10 (3 for slow tests). + type: int + unit: None + default: None + copies - Run <n> copies of each test in parallel. + type: int + unit: None + default: None + + more info https://github.com/kdlucas/byte-unixbench/blob/master/UnixBench + """ + __scenario_type__ = "UnixBench" + + TARGET_SCRIPT = "unixbench_benchmark.bash" + + def __init__(self, scenario_cfg, context_cfg): + self.scenario_cfg = scenario_cfg + self.context_cfg = context_cfg + self.setup_done = False + + def setup(self): + """scenario setup""" + self.target_script = pkg_resources.resource_filename( + "yardstick.benchmark.scenarios.compute", + Unixbench.TARGET_SCRIPT) + + host = self.context_cfg["host"] + user = host.get("user", "ubuntu") + ip = host.get("ip", None) + key_filename = host.get('key_filename', "~/.ssh/id_rsa") + + LOG.info("user:%s, host:%s", user, ip) + self.client = ssh.SSH(user, ip, key_filename=key_filename) + self.client.wait(timeout=600) + + # copy scripts to host + self.client.run("cat > ~/unixbench_benchmark.sh", + stdin=open(self.target_script, 'rb')) + + self.setup_done = True + + def run(self, result): + """execute the benchmark""" + + if not self.setup_done: + self.setup() + + options = self.scenario_cfg["options"] + + run_mode = options.get("run_mode", None) + LOG.debug("Executing run_mode: %s", run_mode) + cmd_args = "" + if run_mode == "quiet": + cmd_args = "-q" + elif run_mode == "verbose": + cmd_args = "-v" + + option_pair_list = [("iterations", "-i"), + ("copies", "-c")] + for option_pair in option_pair_list: + if option_pair[0] in options: + cmd_args += " %s %s " % (option_pair[1], + options[option_pair[0]]) + + test_type = options.get("test_type", None) + if test_type is not None: + cmd_args += " %s " % (test_type) + + cmd = "sudo bash unixbench_benchmark.sh %s" % (cmd_args) + LOG.debug("Executing command: %s", cmd) + status, stdout, stderr = self.client.execute(cmd) + if status: + raise RuntimeError(stderr) + + result.update(json.loads(stdout)) + + if "sla" in self.scenario_cfg: + sla_error = "" + for t, score in result.items(): + if t not in self.scenario_cfg['sla']: + continue + sla_score = float(self.scenario_cfg['sla'][t]) + score = float(score) + if score < sla_score: + sla_error += "%s score %f < sla:%s_score(%f); " % \ + (t, score, t, sla_score) + assert sla_error == "", sla_error + + +def _test(): # pragma: no cover + """internal test function""" + key_filename = pkg_resources.resource_filename('yardstick.resources', + 'files/yardstick_key') + ctx = { + 'host': { + 'ip': '10.229.47.137', + 'user': 'root', + 'key_filename': key_filename + } + } + + options = { + 'test_type': 'dhrystone', + 'run_mode': 'verbose' + } + + args = {'options': options} + result = {} + + p = Unixbench(args, ctx) + p.run(result) + print result + +if __name__ == '__main__': + _test() diff --git a/yardstick/benchmark/scenarios/compute/unixbench_benchmark.bash b/yardstick/benchmark/scenarios/compute/unixbench_benchmark.bash new file mode 100644 index 000000000..5a5dbc394 --- /dev/null +++ b/yardstick/benchmark/scenarios/compute/unixbench_benchmark.bash @@ -0,0 +1,46 @@ +#!/bin/bash + +############################################################################## +# Copyright (c) 2015 Huawei Technologies Co.,Ltd and others. +# +# All rights reserved. This program and the accompanying materials +# are made available under the terms of the Apache License, Version 2.0 +# which accompanies this distribution, and is available at +# http://www.apache.org/licenses/LICENSE-2.0 +############################################################################## + +set -e + +# Commandline arguments +OPTIONS="$@" +OUTPUT_FILE=/tmp/unixbench-out.log + +# run unixbench test +run_unixbench() +{ + cd /opt/tempT/UnixBench/ + ./Run $OPTIONS > $OUTPUT_FILE +} + +# write the result to stdout in json format +output_json() +{ + single_score=$(awk '/Score/{print $7}' $OUTPUT_FILE | head -1 ) + parallel_score=$(awk '/Score/{print $7}' $OUTPUT_FILE | tail -1 ) + echo -e "{ \ + \"single_score\":\"$single_score\", \ + \"parallel_score\":\"$parallel_score\" \ + }" +} + +# main entry +main() +{ + # run the test + run_unixbench + + # output result + output_json +} + +main |