diff options
Diffstat (limited to 'yardstick/tests/unit/benchmark/runner/test_search.py')
-rw-r--r-- | yardstick/tests/unit/benchmark/runner/test_search.py | 202 |
1 files changed, 202 insertions, 0 deletions
diff --git a/yardstick/tests/unit/benchmark/runner/test_search.py b/yardstick/tests/unit/benchmark/runner/test_search.py new file mode 100644 index 000000000..4e5b4fe77 --- /dev/null +++ b/yardstick/tests/unit/benchmark/runner/test_search.py @@ -0,0 +1,202 @@ +# Copyright (c) 2017 Intel Corporation +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import time + +import mock +import unittest + +from yardstick.benchmark.runners.search import SearchRunner +from yardstick.benchmark.runners.search import SearchRunnerHelper + + +class TestSearchRunnerHelper(unittest.TestCase): + + def test___call__(self): + cls = mock.MagicMock() + aborted = mock.MagicMock() + scenario_cfg = { + 'runner': {}, + } + + benchmark = cls() + method = getattr(benchmark, 'my_method') + helper = SearchRunnerHelper( + cls, 'my_method', scenario_cfg, {}, aborted) + + with helper.get_benchmark_instance(): + helper() + + self.assertEqual(method.call_count, 1) + + def test___call___error(self): + cls = mock.MagicMock() + aborted = mock.MagicMock() + scenario_cfg = { + 'runner': {}, + } + + helper = SearchRunnerHelper( + cls, 'my_method', scenario_cfg, {}, aborted) + + with self.assertRaises(RuntimeError): + helper() + + @mock.patch.object(time, 'sleep') + @mock.patch.object(time, 'time') + def test_is_not_done(self, mock_time, *args): + cls = mock.MagicMock() + aborted = mock.MagicMock() + scenario_cfg = { + 'runner': {}, + } + + mock_time.side_effect = range(1000) + + helper = SearchRunnerHelper( + cls, 'my_method', scenario_cfg, {}, aborted) + + index = -1 + for index in helper.is_not_done(): + if index >= 10: + break + + self.assertGreaterEqual(index, 10) + + @mock.patch.object(time, 'sleep') + def test_is_not_done_immediate_stop(self, *args): + cls = mock.MagicMock() + aborted = mock.MagicMock() + scenario_cfg = { + 'runner': { + 'run_step': '', + }, + } + + helper = SearchRunnerHelper( + cls, 'my_method', scenario_cfg, {}, aborted) + + index = -1 + for index in helper.is_not_done(): + if index >= 10: + break + + self.assertEqual(index, -1) + + +class TestSearchRunner(unittest.TestCase): + + def test__worker_run_once(self): + def update(*args): + args[-1].update(data) + + data = { + 'key1': { + 'inner1': 'value1', + 'done': 0, + }, + 'key2': { + 'done': None, + }, + } + + runner = SearchRunner({}) + runner.worker_helper = mock.MagicMock(side_effect=update) + + self.assertFalse(runner._worker_run_once('sequence 1')) + + def test__worker_run_once_done(self): + def update(*args): + args[-1].update(data) + + data = { + 'key1': { + 'inner1': 'value1', + 'done': 0, + }, + 'key2': { + 'done': None, + }, + 'key3': { + 'done': True, + }, + 'key4': [], + 'key5': 'value5', + } + + runner = SearchRunner({}) + runner.worker_helper = mock.MagicMock(side_effect=update) + + self.assertTrue(runner._worker_run_once('sequence 1')) + + def test__worker_run_once_assertion_error_assert(self): + runner = SearchRunner({}) + runner.sla_action = 'assert' + runner.worker_helper = mock.MagicMock(side_effect=AssertionError) + + with self.assertRaises(AssertionError): + runner._worker_run_once('sequence 1') + + def test__worker_run_once_assertion_error_monitor(self): + runner = SearchRunner({}) + runner.sla_action = 'monitor' + runner.worker_helper = mock.MagicMock(side_effect=AssertionError) + + self.assertFalse(runner._worker_run_once('sequence 1')) + + def test__worker_run_once_non_assertion_error_none(self): + runner = SearchRunner({}) + runner.worker_helper = mock.MagicMock(side_effect=RuntimeError) + + self.assertTrue(runner._worker_run_once('sequence 1')) + + def test__worker_run_once_non_assertion_error(self): + runner = SearchRunner({}) + runner.sla_action = 'monitor' + runner.worker_helper = mock.MagicMock(side_effect=RuntimeError) + + self.assertFalse(runner._worker_run_once('sequence 1')) + + def test__worker_run(self): + cls = mock.MagicMock() + scenario_cfg = { + 'runner': {'interval': 0, 'timeout': 1}, + } + + runner = SearchRunner({}) + runner._worker_run_once = mock.MagicMock(side_effect=[0, 0, 1]) + + runner._worker_run(cls, 'my_method', scenario_cfg, {}) + + def test__worker_run_immediate_stop(self): + cls = mock.MagicMock() + scenario_cfg = { + 'runner': { + 'run_step': '', + }, + } + + runner = SearchRunner({}) + runner._worker_run(cls, 'my_method', scenario_cfg, {}) + + @mock.patch('yardstick.benchmark.runners.search.multiprocessing') + def test__run_benchmark(self, mock_multi_process): + cls = mock.MagicMock() + scenario_cfg = { + 'runner': {}, + } + + runner = SearchRunner({}) + runner._run_benchmark(cls, 'my_method', scenario_cfg, {}) + self.assertEqual(mock_multi_process.Process.call_count, 1) |