diff options
Diffstat (limited to 'xtesting')
-rw-r--r-- | xtesting/behaveframework.py | 123 |
1 files changed, 0 insertions, 123 deletions
diff --git a/xtesting/behaveframework.py b/xtesting/behaveframework.py deleted file mode 100644 index 651240d..0000000 --- a/xtesting/behaveframework.py +++ /dev/null @@ -1,123 +0,0 @@ -#!/usr/bin/env python -# Copyright 2021 Orange -# -# Licensed under the Apache License, Version 2.0 (the "License"); you may -# not use this file except in compliance with the License. You may obtain -# a copy of the License at -# -# http://www.apache.org/licenses/LICENSE-2.0 -# -# Unless required by applicable law or agreed to in writing, software -# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT -# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the -# License for the specific language governing permissions and limitations -# under the License. -# - -"""Define classes required to run any Behave test suites.""" - -from __future__ import division - -import logging -import os -import time - -import json -import six - -from behave.__main__ import main as behave_main - -from xtesting.core import testcase - -__author__ = "Deepak Chandella <deepak.chandella@orange.com>" - - -class BehaveFramework(testcase.TestCase): - """BehaveFramework runner.""" - # pylint: disable=too-many-instance-attributes - - __logger = logging.getLogger(__name__) - dir_results = "/var/lib/xtesting/results" - - def __init__(self, **kwargs): - super().__init__(**kwargs) - self.json_file = os.path.join(self.res_dir, 'output.json') - self.total_tests = 0 - self.pass_tests = 0 - self.fail_tests = 0 - self.skip_tests = 0 - self.response = None - - def parse_results(self): - """Parse output.json and get the details in it.""" - with open(self.json_file) as stream_: - self.response = json.load(stream_) - if self.response: - self.total_tests = len(self.response) - for item in self.response: - if item['status'] == 'passed': - self.pass_tests += 1 - elif item['status'] == 'failed': - self.fail_tests += 1 - elif item['status'] == 'skipped': - self.skip_tests += 1 - self.result = 100 * ( - self.pass_tests / self.total_tests) - self.details = {} - self.details['total_tests'] = self.total_tests - self.details['pass_tests'] = self.pass_tests - self.details['fail_tests'] = self.fail_tests - self.details['skip_tests'] = self.skip_tests - self.details['tests'] = self.response - - def run(self, **kwargs): - """Run the BehaveFramework feature files - - Here are the steps: - * create the output directories if required, - * run behave features with parameters - * get the results in output.json, - - Args: - kwargs: Arbitrary keyword arguments. - - Returns: - EX_OK if all suites ran well. - EX_RUN_ERROR otherwise. - """ - try: - suites = kwargs["suites"] - tags = kwargs.get("tags", []) - console = kwargs["console"] if "console" in kwargs else False - except KeyError: - self.__logger.exception("Mandatory args were not passed") - return self.EX_RUN_ERROR - if not os.path.exists(self.res_dir): - try: - os.makedirs(self.res_dir) - except Exception: # pylint: disable=broad-except - self.__logger.exception("Cannot create %s", self.res_dir) - return self.EX_RUN_ERROR - config = ['--tags=' + ','.join(tags), - '--junit', '--junit-directory={}'.format(self.res_dir), - '--format=json', '--outfile={}'.format(self.json_file)] - if six.PY3: - html_file = os.path.join(self.res_dir, 'output.html') - config += ['--format=behave_html_formatter:HTMLFormatter', - '--outfile={}'.format(html_file)] - if console: - config += ['--format=pretty', - '--outfile=-'] - for feature in suites: - config.append(feature) - self.start_time = time.time() - behave_main(config) - self.stop_time = time.time() - - try: - self.parse_results() - self.__logger.info("Results were successfully parsed") - except Exception: # pylint: disable=broad-except - self.__logger.exception("Cannot parse results") - return self.EX_RUN_ERROR - return self.EX_OK |