aboutsummaryrefslogtreecommitdiffstats
path: root/tests/unit/benchmark/runner
diff options
context:
space:
mode:
authorEmma Foley <emma.l.foley@intel.com>2017-11-29 22:56:21 +0000
committerEmma Foley <emma.l.foley@intel.com>2018-02-12 16:57:56 +0000
commitba4e9e6e47bd10ecc803bab920178ea973c2fa86 (patch)
tree54576c199cabe8ccc29d04eedfe0e4f88b2bd598 /tests/unit/benchmark/runner
parentbabe3cc2882e19c6dafdbf41d502d7ba5560635a (diff)
Move tests: unit/benchmark
* Fix pylint errors * Add TODOs Some errors are ignored locally, as they were a symptom of other problems. These issues have been flagged with a TODO, and should be fixed later. Change-Id: I30eb4b0aafe0575d0cddbc946108291f21a98ed8 Jira: YARDSTICK-837 Signed-off-by: Emma Foley <emma.l.foley@intel.com> Signed-off-by: Ross Brattain <ross.b.brattain@intel.com>
Diffstat (limited to 'tests/unit/benchmark/runner')
-rw-r--r--tests/unit/benchmark/runner/__init__.py0
-rw-r--r--tests/unit/benchmark/runner/test_base.py95
-rw-r--r--tests/unit/benchmark/runner/test_search.py204
3 files changed, 0 insertions, 299 deletions
diff --git a/tests/unit/benchmark/runner/__init__.py b/tests/unit/benchmark/runner/__init__.py
deleted file mode 100644
index e69de29bb..000000000
--- a/tests/unit/benchmark/runner/__init__.py
+++ /dev/null
diff --git a/tests/unit/benchmark/runner/test_base.py b/tests/unit/benchmark/runner/test_base.py
deleted file mode 100644
index f47b88e95..000000000
--- a/tests/unit/benchmark/runner/test_base.py
+++ /dev/null
@@ -1,95 +0,0 @@
-#!/usr/bin/env python
-
-##############################################################################
-# Copyright (c) 2017 Huawei Technologies Co.,Ltd and others.
-#
-# All rights reserved. This program and the accompanying materials
-# are made available under the terms of the Apache License, Version 2.0
-# which accompanies this distribution, and is available at
-# http://www.apache.org/licenses/LICENSE-2.0
-##############################################################################
-
-from __future__ import print_function
-from __future__ import absolute_import
-
-import unittest
-import time
-
-from mock import mock
-
-from yardstick.benchmark.runners import base
-from yardstick.benchmark.runners.iteration import IterationRunner
-
-
-class ActionTestCase(unittest.TestCase):
-
- @mock.patch("yardstick.benchmark.runners.base.subprocess")
- def test__execute_shell_command(self, mock_subprocess):
- mock_subprocess.check_output.side_effect = Exception()
-
- self.assertEqual(base._execute_shell_command("")[0], -1)
-
- @mock.patch("yardstick.benchmark.runners.base.subprocess")
- def test__single_action(self, mock_subprocess):
- mock_subprocess.check_output.side_effect = Exception()
-
- base._single_action(0, "echo", mock.MagicMock())
-
- @mock.patch("yardstick.benchmark.runners.base.subprocess")
- def test__periodic_action(self, mock_subprocess):
- mock_subprocess.check_output.side_effect = Exception()
-
- base._periodic_action(0, "echo", mock.MagicMock())
-
-
-class RunnerTestCase(unittest.TestCase):
-
- @mock.patch("yardstick.benchmark.runners.iteration.multiprocessing")
- def test_get_output(self, mock_process):
- runner = IterationRunner({})
- runner.output_queue.put({'case': 'opnfv_yardstick_tc002'})
- runner.output_queue.put({'criteria': 'PASS'})
-
- idle_result = {
- 'case': 'opnfv_yardstick_tc002',
- 'criteria': 'PASS'
- }
-
- for retries in range(1000):
- time.sleep(0.01)
- if not runner.output_queue.empty():
- break
- actual_result = runner.get_output()
- self.assertEqual(idle_result, actual_result)
-
- @mock.patch("yardstick.benchmark.runners.iteration.multiprocessing")
- def test_get_result(self, mock_process):
- runner = IterationRunner({})
- runner.result_queue.put({'case': 'opnfv_yardstick_tc002'})
- runner.result_queue.put({'criteria': 'PASS'})
-
- idle_result = [
- {'case': 'opnfv_yardstick_tc002'},
- {'criteria': 'PASS'}
- ]
-
- for retries in range(1000):
- time.sleep(0.01)
- if not runner.result_queue.empty():
- break
- actual_result = runner.get_result()
- self.assertEqual(idle_result, actual_result)
-
- def test__run_benchmark(self):
- runner = base.Runner(mock.Mock())
-
- with self.assertRaises(NotImplementedError):
- runner._run_benchmark(mock.Mock(), mock.Mock(), mock.Mock(), mock.Mock())
-
-
-def main():
- unittest.main()
-
-
-if __name__ == '__main__':
- main()
diff --git a/tests/unit/benchmark/runner/test_search.py b/tests/unit/benchmark/runner/test_search.py
deleted file mode 100644
index 8fab5a71f..000000000
--- a/tests/unit/benchmark/runner/test_search.py
+++ /dev/null
@@ -1,204 +0,0 @@
-# Copyright (c) 2017 Intel Corporation
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-from __future__ import absolute_import
-
-import unittest
-import mock
-
-from tests.unit import STL_MOCKS
-
-STLClient = mock.MagicMock()
-stl_patch = mock.patch.dict("sys.modules", STL_MOCKS)
-stl_patch.start()
-
-if stl_patch:
- from yardstick.benchmark.runners.search import SearchRunner
- from yardstick.benchmark.runners.search import SearchRunnerHelper
-
-
-class TestSearchRunnerHelper(unittest.TestCase):
-
- def test___call__(self):
- cls = mock.MagicMock()
- aborted = mock.MagicMock()
- scenario_cfg = {
- 'runner': {},
- }
-
- benchmark = cls()
- method = getattr(benchmark, 'my_method')
- helper = SearchRunnerHelper(cls, 'my_method', scenario_cfg, {}, aborted)
-
- with helper.get_benchmark_instance():
- helper()
-
- self.assertEqual(method.call_count, 1)
-
- def test___call___error(self):
- cls = mock.MagicMock()
- aborted = mock.MagicMock()
- scenario_cfg = {
- 'runner': {},
- }
-
- helper = SearchRunnerHelper(cls, 'my_method', scenario_cfg, {}, aborted)
-
- with self.assertRaises(RuntimeError):
- helper()
-
- @mock.patch('yardstick.benchmark.runners.search.time')
- def test_is_not_done(self, mock_time):
- cls = mock.MagicMock()
- aborted = mock.MagicMock()
- scenario_cfg = {
- 'runner': {},
- }
-
- mock_time.time.side_effect = range(1000)
-
- helper = SearchRunnerHelper(cls, 'my_method', scenario_cfg, {}, aborted)
-
- index = -1
- for index in helper.is_not_done():
- if index >= 10:
- break
-
- self.assertGreaterEqual(index, 10)
-
- @mock.patch('yardstick.benchmark.runners.search.time')
- def test_is_not_done_immediate_stop(self, mock_time):
- cls = mock.MagicMock()
- aborted = mock.MagicMock()
- scenario_cfg = {
- 'runner': {
- 'run_step': '',
- },
- }
-
- helper = SearchRunnerHelper(cls, 'my_method', scenario_cfg, {}, aborted)
-
- index = -1
- for index in helper.is_not_done():
- if index >= 10:
- break
-
- self.assertEqual(index, -1)
-
-class TestSearchRunner(unittest.TestCase):
-
- def test__worker_run_once(self):
- def update(*args):
- args[-1].update(data)
-
- data = {
- 'key1': {
- 'inner1': 'value1',
- 'done': 0,
- },
- 'key2': {
- 'done': None,
- },
- }
-
- runner = SearchRunner({})
- runner.worker_helper = mock.MagicMock(side_effect=update)
-
- self.assertFalse(runner._worker_run_once('sequence 1'))
-
- def test__worker_run_once_done(self):
- def update(*args):
- args[-1].update(data)
-
- data = {
- 'key1': {
- 'inner1': 'value1',
- 'done': 0,
- },
- 'key2': {
- 'done': None,
- },
- 'key3': {
- 'done': True,
- },
- 'key4': [],
- 'key5': 'value5',
- }
-
- runner = SearchRunner({})
- runner.worker_helper = mock.MagicMock(side_effect=update)
-
- self.assertTrue(runner._worker_run_once('sequence 1'))
-
- def test__worker_run_once_assertion_error_assert(self):
- runner = SearchRunner({})
- runner.sla_action = 'assert'
- runner.worker_helper = mock.MagicMock(side_effect=AssertionError)
-
- with self.assertRaises(AssertionError):
- runner._worker_run_once('sequence 1')
-
- def test__worker_run_once_assertion_error_monitor(self):
- runner = SearchRunner({})
- runner.sla_action = 'monitor'
- runner.worker_helper = mock.MagicMock(side_effect=AssertionError)
-
- self.assertFalse(runner._worker_run_once('sequence 1'))
-
- def test__worker_run_once_non_assertion_error_none(self):
- runner = SearchRunner({})
- runner.worker_helper = mock.MagicMock(side_effect=RuntimeError)
-
- self.assertTrue(runner._worker_run_once('sequence 1'))
-
- def test__worker_run_once_non_assertion_error(self):
- runner = SearchRunner({})
- runner.sla_action = 'monitor'
- runner.worker_helper = mock.MagicMock(side_effect=RuntimeError)
-
- self.assertFalse(runner._worker_run_once('sequence 1'))
-
- def test__worker_run(self):
- cls = mock.MagicMock()
- scenario_cfg = {
- 'runner': {'interval': 0, 'timeout': 1},
- }
-
- runner = SearchRunner({})
- runner._worker_run_once = mock.MagicMock(side_effect=[0, 0, 1])
-
- runner._worker_run(cls, 'my_method', scenario_cfg, {})
-
- def test__worker_run_immediate_stop(self):
- cls = mock.MagicMock()
- scenario_cfg = {
- 'runner': {
- 'run_step': '',
- },
- }
-
- runner = SearchRunner({})
- runner._worker_run(cls, 'my_method', scenario_cfg, {})
-
- @mock.patch('yardstick.benchmark.runners.search.multiprocessing')
- def test__run_benchmark(self, mock_multi_process):
- cls = mock.MagicMock()
- scenario_cfg = {
- 'runner': {},
- }
-
- runner = SearchRunner({})
- runner._run_benchmark(cls, 'my_method', scenario_cfg, {})
- self.assertEqual(mock_multi_process.Process.call_count, 1)