summaryrefslogtreecommitdiffstats
path: root/tests/unit/benchmark/scenarios
diff options
context:
space:
mode:
authorliang gao <jean.gaoliang@huawei.com>2016-07-18 02:53:13 +0000
committerGerrit Code Review <gerrit@172.30.200.206>2016-07-18 02:53:13 +0000
commit428c72a39bde85d29fae024e1f3d38f5a177f292 (patch)
treebc90af624c0711397b3bb981f7be44d81cfa47c7 /tests/unit/benchmark/scenarios
parente6277570188829687b2ecf48adbd10bc3c9d46bd (diff)
parent1d7bdc10b32892539f8bf69eb9631918d34ffd61 (diff)
Merge "throughput between nodes (in progress)"
Diffstat (limited to 'tests/unit/benchmark/scenarios')
-rwxr-xr-xtests/unit/benchmark/scenarios/networking/test_netperf_node.py127
1 files changed, 127 insertions, 0 deletions
diff --git a/tests/unit/benchmark/scenarios/networking/test_netperf_node.py b/tests/unit/benchmark/scenarios/networking/test_netperf_node.py
new file mode 100755
index 000000000..1c39b292b
--- /dev/null
+++ b/tests/unit/benchmark/scenarios/networking/test_netperf_node.py
@@ -0,0 +1,127 @@
+#!/usr/bin/env python
+
+##############################################################################
+# Copyright (c) 2015 Huawei Technologies Co.,Ltd and others.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+
+# Unittest for
+# yardstick.benchmark.scenarios.networking.netperf_node.NetperfNode
+
+import mock
+import unittest
+import os
+import json
+
+from yardstick.benchmark.scenarios.networking import netperf_node
+
+
+@mock.patch('yardstick.benchmark.scenarios.networking.netperf_node.ssh')
+class NetperfNodeTestCase(unittest.TestCase):
+
+ def setUp(self):
+ self.ctx = {
+ 'host': {
+ 'ip': '192.168.10.10',
+ 'user': 'root',
+ 'password': 'root'
+ },
+ 'target': {
+ 'ip': '192.168.10.11',
+ 'user': 'root',
+ 'password': 'root'
+ }
+ }
+
+ def test_netperf_node_successful_setup(self, mock_ssh):
+
+ p = netperf_node.NetperfNode({}, self.ctx)
+ mock_ssh.SSH().execute.return_value = (0, '', '')
+
+ p.setup()
+ self.assertIsNotNone(p.server)
+ self.assertIsNotNone(p.client)
+ self.assertEqual(p.setup_done, True)
+
+ def test_netperf_node_successful_no_sla(self, mock_ssh):
+
+ options = {}
+ args = {'options': options}
+ result = {}
+
+ p = netperf_node.NetperfNode(args, self.ctx)
+ mock_ssh.SSH().execute.return_value = (0, '', '')
+ p.host = mock_ssh.SSH()
+
+ sample_output = self._read_sample_output()
+ mock_ssh.SSH().execute.return_value = (0, sample_output, '')
+ expected_result = json.loads(sample_output)
+ p.run(result)
+ self.assertEqual(result, expected_result)
+
+ def test_netperf_node_successful_sla(self, mock_ssh):
+
+ options = {}
+ args = {
+ 'options': options,
+ 'sla': {'mean_latency': 100}
+ }
+ result = {}
+
+ p = netperf_node.NetperfNode(args, self.ctx)
+ mock_ssh.SSH().execute.return_value = (0, '', '')
+ p.host = mock_ssh.SSH()
+
+ sample_output = self._read_sample_output()
+ mock_ssh.SSH().execute.return_value = (0, sample_output, '')
+ expected_result = json.loads(sample_output)
+ p.run(result)
+ self.assertEqual(result, expected_result)
+
+ def test_netperf_node_unsuccessful_sla(self, mock_ssh):
+
+ options = {}
+ args = {
+ 'options': options,
+ 'sla': {'mean_latency': 5}
+ }
+ result = {}
+
+ p = netperf_node.NetperfNode(args, self.ctx)
+ mock_ssh.SSH().execute.return_value = (0, '', '')
+ p.host = mock_ssh.SSH()
+
+ sample_output = self._read_sample_output()
+ mock_ssh.SSH().execute.return_value = (0, sample_output, '')
+ self.assertRaises(AssertionError, p.run, result)
+
+ def test_netperf_node_unsuccessful_script_error(self, mock_ssh):
+
+ options = {}
+ args = {'options': options}
+ result = {}
+
+ p = netperf_node.NetperfNode(args, self.ctx)
+ mock_ssh.SSH().execute.return_value = (0, '', '')
+ p.host = mock_ssh.SSH()
+
+ mock_ssh.SSH().execute.return_value = (1, '', 'FOOBAR')
+ self.assertRaises(RuntimeError, p.run, result)
+
+ def _read_sample_output(self):
+ curr_path = os.path.dirname(os.path.abspath(__file__))
+ output = os.path.join(curr_path, 'netperf_sample_output.json')
+ with open(output) as f:
+ sample_output = f.read()
+ return sample_output
+
+
+def main():
+ unittest.main()
+
+if __name__ == '__main__':
+ main()