summaryrefslogtreecommitdiffstats
path: root/tests/unit/benchmark/runner/test_search.py
diff options
context:
space:
mode:
authorEmma Foley <emma.l.foley@intel.com>2017-11-29 22:56:21 +0000
committerEmma Foley <emma.l.foley@intel.com>2018-02-12 16:57:56 +0000
commitba4e9e6e47bd10ecc803bab920178ea973c2fa86 (patch)
tree54576c199cabe8ccc29d04eedfe0e4f88b2bd598 /tests/unit/benchmark/runner/test_search.py
parentbabe3cc2882e19c6dafdbf41d502d7ba5560635a (diff)
Move tests: unit/benchmark
* Fix pylint errors * Add TODOs Some errors are ignored locally, as they were a symptom of other problems. These issues have been flagged with a TODO, and should be fixed later. Change-Id: I30eb4b0aafe0575d0cddbc946108291f21a98ed8 Jira: YARDSTICK-837 Signed-off-by: Emma Foley <emma.l.foley@intel.com> Signed-off-by: Ross Brattain <ross.b.brattain@intel.com>
Diffstat (limited to 'tests/unit/benchmark/runner/test_search.py')
-rw-r--r--tests/unit/benchmark/runner/test_search.py204
1 files changed, 0 insertions, 204 deletions
diff --git a/tests/unit/benchmark/runner/test_search.py b/tests/unit/benchmark/runner/test_search.py
deleted file mode 100644
index 8fab5a71f..000000000
--- a/tests/unit/benchmark/runner/test_search.py
+++ /dev/null
@@ -1,204 +0,0 @@
-# Copyright (c) 2017 Intel Corporation
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-from __future__ import absolute_import
-
-import unittest
-import mock
-
-from tests.unit import STL_MOCKS
-
-STLClient = mock.MagicMock()
-stl_patch = mock.patch.dict("sys.modules", STL_MOCKS)
-stl_patch.start()
-
-if stl_patch:
- from yardstick.benchmark.runners.search import SearchRunner
- from yardstick.benchmark.runners.search import SearchRunnerHelper
-
-
-class TestSearchRunnerHelper(unittest.TestCase):
-
- def test___call__(self):
- cls = mock.MagicMock()
- aborted = mock.MagicMock()
- scenario_cfg = {
- 'runner': {},
- }
-
- benchmark = cls()
- method = getattr(benchmark, 'my_method')
- helper = SearchRunnerHelper(cls, 'my_method', scenario_cfg, {}, aborted)
-
- with helper.get_benchmark_instance():
- helper()
-
- self.assertEqual(method.call_count, 1)
-
- def test___call___error(self):
- cls = mock.MagicMock()
- aborted = mock.MagicMock()
- scenario_cfg = {
- 'runner': {},
- }
-
- helper = SearchRunnerHelper(cls, 'my_method', scenario_cfg, {}, aborted)
-
- with self.assertRaises(RuntimeError):
- helper()
-
- @mock.patch('yardstick.benchmark.runners.search.time')
- def test_is_not_done(self, mock_time):
- cls = mock.MagicMock()
- aborted = mock.MagicMock()
- scenario_cfg = {
- 'runner': {},
- }
-
- mock_time.time.side_effect = range(1000)
-
- helper = SearchRunnerHelper(cls, 'my_method', scenario_cfg, {}, aborted)
-
- index = -1
- for index in helper.is_not_done():
- if index >= 10:
- break
-
- self.assertGreaterEqual(index, 10)
-
- @mock.patch('yardstick.benchmark.runners.search.time')
- def test_is_not_done_immediate_stop(self, mock_time):
- cls = mock.MagicMock()
- aborted = mock.MagicMock()
- scenario_cfg = {
- 'runner': {
- 'run_step': '',
- },
- }
-
- helper = SearchRunnerHelper(cls, 'my_method', scenario_cfg, {}, aborted)
-
- index = -1
- for index in helper.is_not_done():
- if index >= 10:
- break
-
- self.assertEqual(index, -1)
-
-class TestSearchRunner(unittest.TestCase):
-
- def test__worker_run_once(self):
- def update(*args):
- args[-1].update(data)
-
- data = {
- 'key1': {
- 'inner1': 'value1',
- 'done': 0,
- },
- 'key2': {
- 'done': None,
- },
- }
-
- runner = SearchRunner({})
- runner.worker_helper = mock.MagicMock(side_effect=update)
-
- self.assertFalse(runner._worker_run_once('sequence 1'))
-
- def test__worker_run_once_done(self):
- def update(*args):
- args[-1].update(data)
-
- data = {
- 'key1': {
- 'inner1': 'value1',
- 'done': 0,
- },
- 'key2': {
- 'done': None,
- },
- 'key3': {
- 'done': True,
- },
- 'key4': [],
- 'key5': 'value5',
- }
-
- runner = SearchRunner({})
- runner.worker_helper = mock.MagicMock(side_effect=update)
-
- self.assertTrue(runner._worker_run_once('sequence 1'))
-
- def test__worker_run_once_assertion_error_assert(self):
- runner = SearchRunner({})
- runner.sla_action = 'assert'
- runner.worker_helper = mock.MagicMock(side_effect=AssertionError)
-
- with self.assertRaises(AssertionError):
- runner._worker_run_once('sequence 1')
-
- def test__worker_run_once_assertion_error_monitor(self):
- runner = SearchRunner({})
- runner.sla_action = 'monitor'
- runner.worker_helper = mock.MagicMock(side_effect=AssertionError)
-
- self.assertFalse(runner._worker_run_once('sequence 1'))
-
- def test__worker_run_once_non_assertion_error_none(self):
- runner = SearchRunner({})
- runner.worker_helper = mock.MagicMock(side_effect=RuntimeError)
-
- self.assertTrue(runner._worker_run_once('sequence 1'))
-
- def test__worker_run_once_non_assertion_error(self):
- runner = SearchRunner({})
- runner.sla_action = 'monitor'
- runner.worker_helper = mock.MagicMock(side_effect=RuntimeError)
-
- self.assertFalse(runner._worker_run_once('sequence 1'))
-
- def test__worker_run(self):
- cls = mock.MagicMock()
- scenario_cfg = {
- 'runner': {'interval': 0, 'timeout': 1},
- }
-
- runner = SearchRunner({})
- runner._worker_run_once = mock.MagicMock(side_effect=[0, 0, 1])
-
- runner._worker_run(cls, 'my_method', scenario_cfg, {})
-
- def test__worker_run_immediate_stop(self):
- cls = mock.MagicMock()
- scenario_cfg = {
- 'runner': {
- 'run_step': '',
- },
- }
-
- runner = SearchRunner({})
- runner._worker_run(cls, 'my_method', scenario_cfg, {})
-
- @mock.patch('yardstick.benchmark.runners.search.multiprocessing')
- def test__run_benchmark(self, mock_multi_process):
- cls = mock.MagicMock()
- scenario_cfg = {
- 'runner': {},
- }
-
- runner = SearchRunner({})
- runner._run_benchmark(cls, 'my_method', scenario_cfg, {})
- self.assertEqual(mock_multi_process.Process.call_count, 1)