diff options
Diffstat (limited to 'functest')
21 files changed, 291 insertions, 545 deletions
diff --git a/functest/ci/config_functest.yaml b/functest/ci/config_functest.yaml index 2820f8b0..f6cb14cb 100755 --- a/functest/ci/config_functest.yaml +++ b/functest/ci/config_functest.yaml @@ -20,8 +20,8 @@ general: repo_sdnvpn: /home/opnfv/repos/sdnvpn repo_sfc: /home/opnfv/repos/sfc dir_repo_onos: /home/opnfv/repos/onos - dir_repo_promise: /home/opnfv/repos/promise - dir_repo_doctor: /home/opnfv/repos/doctor + repo_promise: /home/opnfv/repos/promise + repo_doctor: /home/opnfv/repos/doctor repo_copper: /home/opnfv/repos/copper dir_repo_ovno: /home/opnfv/repos/ovno repo_parser: /home/opnfv/repos/parser diff --git a/functest/ci/exec_test.sh b/functest/ci/exec_test.sh index 54a7c624..b288fe36 100755 --- a/functest/ci/exec_test.sh +++ b/functest/ci/exec_test.sh @@ -90,22 +90,12 @@ function run_test(){ --osusername ${OS_USERNAME} \ ${args} ;; - "vims") - python ${FUNCTEST_TEST_DIR}/vnf/ims/vims.py $clean_flag $report - ;; "onos") python ${FUNCTEST_TEST_DIR}/sdn/onos/teston/onos.py ;; "onos_sfc") python ${FUNCTEST_TEST_DIR}/sdn/onos/teston/onos.py -t sfc ;; - "promise") - python ${FUNCTEST_TEST_DIR}/features/promise.py $report - sleep 10 # to let the instances terminate - ;; - "doctor") - python ${FUNCTEST_TEST_DIR}/features/doctor.py $report - ;; "ovno") # suite under rewritting for colorado # no need to run anything until refactoring done @@ -116,9 +106,6 @@ function run_test(){ source ${FUNCTEST_CONF_DIR}/stackrc python ${FUNCTEST_TEST_DIR}/security_scan/security_scan.py --config ${FUNCTEST_TEST_DIR}/security_scan/config.ini ;; - "copper") - python ${FUNCTEST_TEST_DIR}/features/copper.py $report - ;; "moon") python ${REPOS_DIR}/moon/tests/run_tests.py $report ;; diff --git a/functest/ci/logging.json b/functest/ci/logging.json index 3f454e8f..2a2399d3 100644 --- a/functest/ci/logging.json +++ b/functest/ci/logging.json @@ -1,29 +1,29 @@ -{
- "version": 1,
- "disable_existing_loggers": false,
- "formatters": {
- "standard": {
- "format": "%(asctime)s - %(name)s - %(levelname)s - %(message)s"
- }
- },
- "handlers": {
- "console": {
- "level": "INFO",
- "class": "logging.StreamHandler",
- "formatter": "standard"
- },
- "file": {
- "level": "DEBUG",
- "class": "logging.FileHandler",
- "formatter": "standard",
- "filename": "/home/opnfv/functest/results/functest.log"
- }
- },
- "loggers": {
- "": {
- "handlers": ["console", "file"],
- "level": "DEBUG",
- "propagate": "yes"
- }
- }
-}
+{ + "version": 1, + "disable_existing_loggers": false, + "formatters": { + "standard": { + "format": "%(asctime)s - %(name)s - %(levelname)s - %(message)s" + } + }, + "handlers": { + "console": { + "level": "INFO", + "class": "logging.StreamHandler", + "formatter": "standard" + }, + "file": { + "level": "DEBUG", + "class": "logging.FileHandler", + "formatter": "standard", + "filename": "/home/opnfv/functest/results/functest.log" + } + }, + "loggers": { + "": { + "handlers": ["console", "file"], + "level": "DEBUG", + "propagate": "yes" + } + } +} diff --git a/functest/ci/prepare_env.py b/functest/ci/prepare_env.py index 8bbdf18b..b3e59020 100755 --- a/functest/ci/prepare_env.py +++ b/functest/ci/prepare_env.py @@ -261,6 +261,14 @@ def install_tempest(): error_msg="Problem while installing Tempest.") +def create_flavor(): + os_utils.get_or_create_flavor('m1.tiny', + '512', + '1', + '1', + public=True) + + def check_environment(): msg_not_active = "The Functest environment is not installed." if not os.path.isfile(CONST.env_active): @@ -290,6 +298,7 @@ def main(**kwargs): verify_deployment() install_rally() install_tempest() + create_flavor() with open(CONST.env_active, "w") as env_file: env_file.write("1") diff --git a/functest/ci/run_tests.py b/functest/ci/run_tests.py index ef080016..320102dd 100755 --- a/functest/ci/run_tests.py +++ b/functest/ci/run_tests.py @@ -148,7 +148,11 @@ def run_test(test, tier_name, testcases=None): module = importlib.import_module(run_dict['module']) cls = getattr(module, run_dict['class']) test_case = cls() - result = test_case.run() + try: + kwargs = run_dict['args'] + result = test_case.run(**kwargs) + except KeyError: + result = test_case.run() if result == testcase_base.TestcaseBase.EX_OK: if GlobalVariables.REPORT_FLAG: test_case.publish_report() diff --git a/functest/ci/testcases.yaml b/functest/ci/testcases.yaml index f1cce489..12d791f0 100755 --- a/functest/ci/testcases.yaml +++ b/functest/ci/testcases.yaml @@ -99,6 +99,10 @@ tiers: run: module: 'functest.opnfv_tests.sdn.odl.odl' class: 'ODLTests' + args: + suites: + - /home/opnfv/repos/odl_test/csit/suites/integration/basic + - /home/opnfv/repos/odl_test/csit/suites/openstack/neutron - name: onos @@ -185,6 +189,9 @@ tiers: dependencies: installer: '(fuel)|(joid)' scenario: '' + run: + module: 'functest.opnfv_tests.features.promise' + class: 'Promise' - name: doctor @@ -195,6 +202,9 @@ tiers: dependencies: installer: '(apex)|(fuel)|(joid)' scenario: '^((?!fdio).)*$' + run: + module: 'functest.opnfv_tests.features.doctor' + class: 'Doctor' - name: bgpvpn @@ -218,18 +228,18 @@ tiers: dependencies: installer: 'apex' scenario: '^((?!fdio).)*$' - - - name: copper - criteria: 'status == "PASS"' - blocking: false - description: >- - Test suite for policy management based on OpenStack Congress - dependencies: - installer: '(apex)|(joid)' - scenario: '^((?!fdio|lxd).)*$' - run: - module: 'functest.opnfv_tests.features.copper' - class: 'Copper' +# - +# name: copper +# criteria: 'status == "PASS"' +# blocking: false +# description: >- +# Test suite for policy management based on OpenStack Congress +# dependencies: +# installer: '(apex)|(joid)' +# scenario: '^((?!fdio|lxd).)*$' +# run: +# module: 'functest.opnfv_tests.features.copper' +# class: 'Copper' - name: moon criteria: 'status == "PASS"' diff --git a/functest/core/pytest_suite_runner.py b/functest/core/pytest_suite_runner.py index 1eed92b5..c168d7d9 100644 --- a/functest/core/pytest_suite_runner.py +++ b/functest/core/pytest_suite_runner.py @@ -41,14 +41,18 @@ class PyTestSuiteRunner(base.TestcaseBase): for test, message in result.failures: self.logger.error(str(test) + " FAILED with " + message) + # a result can be PASS or FAIL + # But in this case it means that the Execution was OK + # we shall distinguish Execution Error from FAIL results + # TestcaseBase.EX_RUN_ERROR means that the test case was not run + # not that it was run but the result was FAIL + exit_code = base.TestcaseBase.EX_OK if ((result.errors and len(result.errors) > 0) or (result.failures and len(result.failures) > 0)): self.logger.info("%s FAILED" % self.case_name) self.criteria = 'FAIL' - exit_code = base.TestcaseBase.EX_RUN_ERROR else: self.logger.info("%s OK" % self.case_name) - exit_code = base.TestcaseBase.EX_OK self.criteria = 'PASS' self.details = {} diff --git a/functest/opnfv_tests/features/doctor.py b/functest/opnfv_tests/features/doctor.py index dbd803a6..4d295a67 100755 --- a/functest/opnfv_tests/features/doctor.py +++ b/functest/opnfv_tests/features/doctor.py @@ -1,6 +1,6 @@ #!/usr/bin/python # -# Copyright (c) 2015 All rights reserved +# Copyright (c) 2017 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at @@ -13,77 +13,12 @@ # 0.2: measure test duration and publish results under json format # # -import argparse -import os -import time +import functest.core.feature_base as base -import functest.utils.functest_logger as ft_logger -import functest.utils.functest_utils as functest_utils -import functest.utils.functest_constants as ft_constants -parser = argparse.ArgumentParser() -parser.add_argument("-r", "--report", - help="Create json result file", - action="store_true") -args = parser.parse_args() - -functest_yaml = functest_utils.get_functest_yaml() - -DOCTOR_REPO_DIR = ft_constants.DOCTOR_REPO_DIR -RESULTS_DIR = ft_constants.FUNCTEST_RESULTS_DIR - -logger = ft_logger.Logger("doctor").getLogger() - - -def main(): - exit_code = -1 - - # if the image name is explicitly set for the doctor suite, set it as - # enviroment variable - if 'doctor' in functest_yaml and 'image_name' in functest_yaml['doctor']: - os.environ["IMAGE_NAME"] = functest_yaml['doctor']['image_name'] - - cmd = 'cd %s/tests && ./run.sh' % DOCTOR_REPO_DIR - log_file = RESULTS_DIR + "/doctor.log" - - start_time = time.time() - - ret = functest_utils.execute_command(cmd, - info=True, - output_file=log_file) - - stop_time = time.time() - duration = round(stop_time - start_time, 1) - if ret == 0: - logger.info("Doctor test case OK") - test_status = 'OK' - exit_code = 0 - else: - logger.info("Doctor test case FAILED") - test_status = 'NOK' - - details = { - 'timestart': start_time, - 'duration': duration, - 'status': test_status, - } - status = "FAIL" - if details['status'] == "OK": - status = "PASS" - functest_utils.logger_test_results("Doctor", - "doctor-notification", - status, details) - if args.report: - functest_utils.push_results_to_db("doctor", - "doctor-notification", - start_time, - stop_time, - status, - details) - logger.info("Doctor results pushed to DB") - - exit(exit_code) - - -if __name__ == '__main__': - main() +class Doctor(base.FeatureBase): + def __init__(self): + super(Doctor, self).__init__(project='doctor', + case='doctor-notification', + repo='dir_repo_doctor') + self.cmd = 'cd %s/tests && ./run.sh' % self.repo diff --git a/functest/opnfv_tests/features/promise.py b/functest/opnfv_tests/features/promise.py index 0a30e3a1..15636fbf 100755 --- a/functest/opnfv_tests/features/promise.py +++ b/functest/opnfv_tests/features/promise.py @@ -1,252 +1,24 @@ #!/usr/bin/python # -# Copyright (c) 2015 All rights reserved -# This program and the accompanying materials -# are made available under the terms of the Apache License, Version 2.0 -# which accompanies this distribution, and is available at +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # -# Maintainer : jose.lausuch@ericsson.com +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. # -import argparse -import json -import os -import subprocess -import time +import functest.core.feature_base as base -import functest.utils.functest_logger as ft_logger -import functest.utils.functest_utils as ft_utils -import functest.utils.openstack_utils as openstack_utils -import keystoneclient.v2_0.client as ksclient -from neutronclient.v2_0 import client as ntclient -import novaclient.client as nvclient -import functest.utils.functest_constants as ft_constants -parser = argparse.ArgumentParser() - -parser.add_argument("-d", "--debug", help="Debug mode", action="store_true") -parser.add_argument("-r", "--report", - help="Create json result file", - action="store_true") -args = parser.parse_args() - - -PROMISE_REPO_DIR = ft_constants.PROMISE_REPO_DIR -RESULTS_DIR = ft_constants.FUNCTEST_RESULTS_DIR - -PROMISE_TENANT_NAME = ft_constants.PROMISE_TENANT_NAME -TENANT_DESCRIPTION = ft_constants.TENANT_DESCRIPTION -PROMISE_USER_NAME = ft_constants.PROMISE_USER_NAME -PROMISE_USER_PWD = ft_constants.PROMISE_USER_PWD -PROMISE_IMAGE_NAME = ft_constants.PROMISE_IMAGE_NAME -PROMISE_FLAVOR_NAME = ft_constants.PROMISE_FLAVOR_NAME -PROMISE_FLAVOR_VCPUS = ft_constants.PROMISE_FLAVOR_VCPUS -PROMISE_FLAVOR_RAM = ft_constants.PROMISE_FLAVOR_RAM -PROMISE_FLAVOR_DISK = ft_constants.PROMISE_FLAVOR_DISK - - -GLANCE_IMAGE_FILENAME = ft_constants.GLANCE_IMAGE_FILENAME -GLANCE_IMAGE_FORMAT = ft_constants.GLANCE_IMAGE_FORMAT -GLANCE_IMAGE_PATH = os.path.join(ft_constants.FUNCTEST_DATA_DIR, - GLANCE_IMAGE_FILENAME) - -PROMISE_NET_NAME = ft_constants.PROMISE_NET_NAME -PROMISE_SUBNET_NAME = ft_constants.PROMISE_SUBNET_NAME -PROMISE_SUBNET_CIDR = ft_constants.PROMISE_SUBNET_CIDR -PROMISE_ROUTER_NAME = ft_constants.PROMISE_ROUTER_NAME - - -""" logging configuration """ -logger = ft_logger.Logger("promise").getLogger() - - -def main(): - exit_code = -1 - start_time = time.time() - ks_creds = openstack_utils.get_credentials("keystone") - nv_creds = openstack_utils.get_credentials("nova") - nt_creds = openstack_utils.get_credentials("neutron") - - keystone = ksclient.Client(**ks_creds) - - user_id = openstack_utils.get_user_id(keystone, ks_creds['username']) - if user_id == '': - logger.error("Error : Failed to get id of %s user" % - ks_creds['username']) - exit(-1) - - logger.info("Creating tenant '%s'..." % PROMISE_TENANT_NAME) - tenant_id = openstack_utils.create_tenant( - keystone, PROMISE_TENANT_NAME, TENANT_DESCRIPTION) - if not tenant_id: - logger.error("Error : Failed to create %s tenant" - % PROMISE_TENANT_NAME) - exit(-1) - logger.debug("Tenant '%s' created successfully." % PROMISE_TENANT_NAME) - - roles_name = ["admin", "Admin"] - role_id = '' - for role_name in roles_name: - if role_id == '': - role_id = openstack_utils.get_role_id(keystone, role_name) - - if role_id == '': - logger.error("Error : Failed to get id for %s role" % role_name) - exit(-1) - - logger.info("Adding role '%s' to tenant '%s'..." - % (role_id, PROMISE_TENANT_NAME)) - if not openstack_utils.add_role_user(keystone, user_id, - role_id, tenant_id): - logger.error("Error : Failed to add %s on tenant %s" % - (ks_creds['username'], PROMISE_TENANT_NAME)) - exit(-1) - logger.debug("Role added successfully.") - - logger.info("Creating user '%s'..." % PROMISE_USER_NAME) - user_id = openstack_utils.create_user( - keystone, PROMISE_USER_NAME, PROMISE_USER_PWD, None, tenant_id) - - if not user_id: - logger.error("Error : Failed to create %s user" % PROMISE_USER_NAME) - exit(-1) - logger.debug("User '%s' created successfully." % PROMISE_USER_NAME) - - logger.info("Updating OpenStack credentials...") - ks_creds.update({ - "username": PROMISE_TENANT_NAME, - "password": PROMISE_TENANT_NAME, - "tenant_name": PROMISE_TENANT_NAME, - }) - - nt_creds.update({ - "tenant_name": PROMISE_TENANT_NAME, - }) - - nv_creds.update({ - "project_id": PROMISE_TENANT_NAME, - }) - - glance = openstack_utils.get_glance_client() - nova = nvclient.Client("2", **nv_creds) - - logger.info("Creating image '%s' from '%s'..." % (PROMISE_IMAGE_NAME, - GLANCE_IMAGE_PATH)) - image_id = openstack_utils.create_glance_image(glance, - PROMISE_IMAGE_NAME, - GLANCE_IMAGE_PATH) - if not image_id: - logger.error("Failed to create the Glance image...") - exit(-1) - logger.debug("Image '%s' with ID '%s' created successfully." - % (PROMISE_IMAGE_NAME, image_id)) - flavor_id = openstack_utils.get_flavor_id(nova, PROMISE_FLAVOR_NAME) - if flavor_id == '': - logger.info("Creating flavor '%s'..." % PROMISE_FLAVOR_NAME) - flavor_id = openstack_utils.create_flavor(nova, - PROMISE_FLAVOR_NAME, - PROMISE_FLAVOR_RAM, - PROMISE_FLAVOR_DISK, - PROMISE_FLAVOR_VCPUS) - if not flavor_id: - logger.error("Failed to create the Flavor...") - exit(-1) - logger.debug("Flavor '%s' with ID '%s' created successfully." % - (PROMISE_FLAVOR_NAME, flavor_id)) - else: - logger.debug("Using existing flavor '%s' with ID '%s'..." - % (PROMISE_FLAVOR_NAME, flavor_id)) - - neutron = ntclient.Client(**nt_creds) - - network_dic = openstack_utils.create_network_full(neutron, - PROMISE_NET_NAME, - PROMISE_SUBNET_NAME, - PROMISE_ROUTER_NAME, - PROMISE_SUBNET_CIDR) - if not network_dic: - logger.error("Failed to create the private network...") - exit(-1) - - logger.info("Exporting environment variables...") - os.environ["NODE_ENV"] = "functest" - os.environ["OS_PASSWORD"] = PROMISE_USER_PWD - os.environ["OS_TEST_IMAGE"] = image_id - os.environ["OS_TEST_FLAVOR"] = flavor_id - os.environ["OS_TEST_NETWORK"] = network_dic["net_id"] - os.environ["OS_TENANT_NAME"] = PROMISE_TENANT_NAME - os.environ["OS_USERNAME"] = PROMISE_USER_NAME - - os.chdir(PROMISE_REPO_DIR + '/source/') - results_file_name = os.path.join(RESULTS_DIR, 'promise-results.json') - results_file = open(results_file_name, 'w+') - cmd = 'npm run -s test -- --reporter json' - - logger.info("Running command: %s" % cmd) - ret = subprocess.call(cmd, shell=True, stdout=results_file, - stderr=subprocess.STDOUT) - results_file.close() - - if ret == 0: - logger.info("The test succeeded.") - # test_status = 'OK' - else: - logger.info("The command '%s' failed." % cmd) - # test_status = "Failed" - - # Print output of file - with open(results_file_name, 'r') as results_file: - data = results_file.read() - logger.debug("\n%s" % data) - json_data = json.loads(data) - - suites = json_data["stats"]["suites"] - tests = json_data["stats"]["tests"] - passes = json_data["stats"]["passes"] - pending = json_data["stats"]["pending"] - failures = json_data["stats"]["failures"] - start_time_json = json_data["stats"]["start"] - end_time = json_data["stats"]["end"] - duration = float(json_data["stats"]["duration"]) / float(1000) - - logger.info("\n" - "****************************************\n" - " Promise test report\n\n" - "****************************************\n" - " Suites: \t%s\n" - " Tests: \t%s\n" - " Passes: \t%s\n" - " Pending: \t%s\n" - " Failures:\t%s\n" - " Start: \t%s\n" - " End: \t%s\n" - " Duration:\t%s\n" - "****************************************\n\n" - % (suites, tests, passes, pending, failures, - start_time_json, end_time, duration)) - - if args.report: - stop_time = time.time() - json_results = {"timestart": start_time, "duration": duration, - "tests": int(tests), "failures": int(failures)} - logger.debug("Promise Results json: " + str(json_results)) - - # criteria for Promise in Release B was 100% of tests OK - status = "FAIL" - if int(tests) > 32 and int(failures) < 1: - status = "PASS" - exit_code = 0 - - ft_utils.push_results_to_db("promise", - "promise", - start_time, - stop_time, - status, - json_results) - - exit(exit_code) - - -if __name__ == '__main__': - main() +class Promise(base.FeatureBase): + def __init__(self): + super(Promise, self).__init__(project='promise', + case='promise', + repo='dir_repo_promise') + dir_promise_functest = '{}/promise/test/functest'.format(self.repo) + self.cmd = 'cd %s && python ./run_tests.py' % dir_promise_functest diff --git a/functest/opnfv_tests/openstack/healthcheck/healthcheck.sh b/functest/opnfv_tests/openstack/healthcheck/healthcheck.sh index 57aa0c70..7fa957c0 100755 --- a/functest/opnfv_tests/openstack/healthcheck/healthcheck.sh +++ b/functest/opnfv_tests/openstack/healthcheck/healthcheck.sh @@ -23,17 +23,17 @@ echo "">$LOG_FILE exec 1<>$LOG_FILE info () { - echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healtcheck - INFO - " "$*" | tee -a $LOG_FILE 1>&2 + echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healthcheck - INFO - " "$*" | tee -a $LOG_FILE 1>&2 } debug () { if [[ "${CI_DEBUG,,}" == "true" ]]; then - echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healtcheck - DEBUG - " "$*" | tee -a $LOG_FILE 1>&2 + echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healthcheck - DEBUG - " "$*" | tee -a $LOG_FILE 1>&2 fi } error () { - echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healtcheck - ERROR - " "$*" | tee -a $LOG_FILE 1>&2 + echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healthcheck - ERROR - " "$*" | tee -a $LOG_FILE 1>&2 exit 1 } @@ -125,16 +125,16 @@ kernel_img=$(cat ${YAML_FILE} | shyaml get-value healthcheck.kernel_image 2> /de ramdisk_img=$(cat ${YAML_FILE} | shyaml get-value healthcheck.ramdisk_image 2> /dev/null || true) extra_properties=$(cat ${YAML_FILE} | shyaml get-value healthcheck.extra_properties 2> /dev/null || true) -# Test if we need to create a 3part image +# Test if we need to create a 3party image if [ "X$kernel_img" != "X" ] then - img_id=$(glance image-create --name ${kernel_image} --disk-format aki \ + img_id=$(openstack image create ${kernel_image} --disk-format aki \ --container-format bare < ${kernel_img} | awk '$2 == "id" { print $4 }') extra_opts="--property kernel_id=${img_id}" if [ "X$ramdisk_img" != "X" ] then - img_id=$(glance image-create --name ${ramdisk_image} --disk-format ari \ + img_id=$(openstack image create ${ramdisk_image} --disk-format ari \ --container-format bare < ${ramdisk_img} | awk '$2 == "id" { print $4 }') extra_opts="$extra_opts --property ramdisk_id=${img_id}" fi @@ -152,10 +152,10 @@ fi debug "image extra_properties=${extra_properties}" -eval glance image-create --name ${image_1} --disk-format ${disk_format} --container-format bare \ +eval openstack image create ${image_1} --disk-format ${disk_format} --container-format bare \ ${extra_opts} < ${disk_img} debug "image '${image_1}' created." -eval glance image-create --name ${image_2} --disk-format ${disk_format} --container-format bare \ +eval openstack image create ${image_2} --disk-format ${disk_format} --container-format bare \ ${extra_opts} < ${disk_img} debug "image '${image_2}' created." info "... Glance OK!" diff --git a/functest/opnfv_tests/openstack/rally/rally.py b/functest/opnfv_tests/openstack/rally/rally.py index de553116..f984c368 100644 --- a/functest/opnfv_tests/openstack/rally/rally.py +++ b/functest/opnfv_tests/openstack/rally/rally.py @@ -94,11 +94,8 @@ class RallyBase(testcase_base.TestcaseBase): else: task_args['netid'] = '' - auth_url = CONST.OS_AUTH_URL - if auth_url is not None: - task_args['request_url'] = auth_url.rsplit(":", 1)[0] - else: - task_args['request_url'] = '' + # get keystone auth endpoint + task_args['request_url'] = CONST.OS_AUTH_URL or '' return task_args @@ -533,10 +530,9 @@ class RallyBase(testcase_base.TestcaseBase): return testcase_base.TestcaseBase.EX_RUN_ERROR self.stop_time = time.time() - if self.criteria == "PASS": - return testcase_base.TestcaseBase.EX_OK - else: - return testcase_base.TestcaseBase.EX_TESTCASE_FAILED + # If we are here, it means that the test case was successfully executed + # criteria is managed by the criteria Field + return testcase_base.TestcaseBase.EX_OK class RallySanity(RallyBase): diff --git a/functest/opnfv_tests/openstack/tempest/conf_utils.py b/functest/opnfv_tests/openstack/tempest/conf_utils.py index 88085d9a..0c9064a1 100644 --- a/functest/opnfv_tests/openstack/tempest/conf_utils.py +++ b/functest/opnfv_tests/openstack/tempest/conf_utils.py @@ -11,10 +11,12 @@ import ConfigParser import os import re import shutil +import subprocess import opnfv.utils.constants as releng_constants from functest.utils.constants import CONST +import functest.utils.functest_logger as ft_logger import functest.utils.functest_utils as ft_utils import functest.utils.openstack_utils as os_utils @@ -39,8 +41,74 @@ TEMPEST_LIST = os.path.join(TEMPEST_RESULTS_DIR, 'test_list.txt') CI_INSTALLER_TYPE = CONST.INSTALLER_TYPE CI_INSTALLER_IP = CONST.INSTALLER_IP +""" logging configuration """ +logger = ft_logger.Logger("Tempest").getLogger() -def configure_tempest(logger, deployment_dir, IMAGE_ID=None, FLAVOR_ID=None): + +def get_verifier_id(): + """ + Returns verifer id for current Tempest + """ + cmd = ("rally verify list-verifiers | awk '/" + + CONST.tempest_deployment_name + + "/ {print $2}'") + p = subprocess.Popen(cmd, shell=True, + stdout=subprocess.PIPE, + stderr=subprocess.STDOUT) + deployment_uuid = p.stdout.readline().rstrip() + if deployment_uuid == "": + logger.error("Tempest verifier not found.") + raise Exception('Error with command:%s' % cmd) + return deployment_uuid + + +def get_verifier_deployment_id(): + """ + Returns deployment id for active Rally deployment + """ + cmd = ("rally deployment list | awk '/" + + CONST.rally_deployment_name + + "/ {print $2}'") + p = subprocess.Popen(cmd, shell=True, + stdout=subprocess.PIPE, + stderr=subprocess.STDOUT) + deployment_uuid = p.stdout.readline().rstrip() + if deployment_uuid == "": + logger.error("Rally deployment not found.") + raise Exception('Error with command:%s' % cmd) + return deployment_uuid + + +def get_verifier_repo_dir(verifier_id): + """ + Returns installed verfier repo directory for Tempest + """ + if not verifier_id: + verifier_id = get_verifier_id() + + return os.path.join(CONST.dir_rally_inst, + 'verification', + 'verifier-{}'.format(verifier_id), + 'repo') + + +def get_verifier_deployment_dir(verifier_id, deployment_id): + """ + Returns Rally deployment directory for current verifier + """ + if not verifier_id: + verifier_id = get_verifier_id() + + if not deployment_id: + deployment_id = get_verifier_deployment_id() + + return os.path.join(CONST.dir_rally_inst, + 'verification', + 'verifier-{}'.format(verifier_id), + 'for-deployment-{}'.format(deployment_id)) + + +def configure_tempest(deployment_dir, IMAGE_ID=None, FLAVOR_ID=None): """ Add/update needed parameters into tempest.conf file generated by Rally """ @@ -104,18 +172,21 @@ def configure_tempest(logger, deployment_dir, IMAGE_ID=None, FLAVOR_ID=None): config.write(config_file) # Copy tempest.conf to /home/opnfv/functest/results/tempest/ + if not os.path.exists(TEMPEST_RESULTS_DIR): + os.makedirs(TEMPEST_RESULTS_DIR) + shutil.copyfile(tempest_conf_file, os.path.join(TEMPEST_RESULTS_DIR, 'tempest.conf')) return releng_constants.EXIT_OK -def configure_tempest_multisite(logger, deployment_dir): +def configure_tempest_multisite(deployment_dir): """ Add/update needed parameters into tempest.conf file generated by Rally """ logger.debug("configure the tempest") - configure_tempest(logger, deployment_dir) + configure_tempest(deployment_dir) logger.debug("Finding tempest.conf file...") tempest_conf_old = os.path.join(deployment_dir, 'tempest.conf') diff --git a/functest/opnfv_tests/openstack/tempest/tempest.py b/functest/opnfv_tests/openstack/tempest/tempest.py index 0014b718..9c19a147 100644 --- a/functest/opnfv_tests/openstack/tempest/tempest.py +++ b/functest/opnfv_tests/openstack/tempest/tempest.py @@ -35,74 +35,15 @@ class TempestCommon(testcase_base.TestcaseBase): self.OPTION = "" self.FLAVOR_ID = None self.IMAGE_ID = None - self.VERIFIER_ID = self.get_verifier_id() - self.VERIFIER_REPO_DIR = self.get_verifier_repo_dir() - self.DEPLOYMENT_ID = self.get_verifier_deployment_id() - self.DEPLOYMENT_DIR = self.get_verifier_deployment_dir() + self.VERIFIER_ID = conf_utils.get_verifier_id() + self.VERIFIER_REPO_DIR = conf_utils.get_verifier_repo_dir( + self.VERIFIER_ID) + self.DEPLOYMENT_ID = conf_utils.get_verifier_deployment_id() + self.DEPLOYMENT_DIR = conf_utils.get_verifier_deployment_dir( + self.VERIFIER_ID, self.DEPLOYMENT_ID) self.VERIFICATION_ID = None @staticmethod - def get_verifier_id(): - """ - Returns verifer id for current Tempest - """ - cmd = ("rally verify list-verifiers | awk '/" + - CONST.tempest_deployment_name + - "/ {print $2}'") - p = subprocess.Popen(cmd, shell=True, - stdout=subprocess.PIPE, - stderr=subprocess.STDOUT) - deployment_uuid = p.stdout.readline().rstrip() - if deployment_uuid == "": - logger.error("Tempest verifier not found.") - raise Exception('Error with command:%s' % cmd) - return deployment_uuid - - @staticmethod - def get_verifier_deployment_id(): - """ - Returns deployment id for active Rally deployment - """ - cmd = ("rally deployment list | awk '/" + - CONST.rally_deployment_name + - "/ {print $2}'") - p = subprocess.Popen(cmd, shell=True, - stdout=subprocess.PIPE, - stderr=subprocess.STDOUT) - deployment_uuid = p.stdout.readline().rstrip() - if deployment_uuid == "": - logger.error("Rally deployment not found.") - raise Exception('Error with command:%s' % cmd) - return deployment_uuid - - def get_verifier_repo_dir(self): - """ - Returns installed verfier repo directory for Tempest - """ - if not self.VERIFIER_ID: - self.VERIFIER_ID = self.get_verifier_id() - - return os.path.join(CONST.dir_rally_inst, - 'verification', - 'verifier-{}'.format(self.VERIFIER_ID), - 'repo') - - def get_verifier_deployment_dir(self): - """ - Returns Rally deployment directory for current verifier - """ - if not self.VERIFIER_ID: - self.VERIFIER_ID = self.get_verifier_id() - - if not self.DEPLOYMENT_ID: - self.DEPLOYMENT_ID = self.get_verifier_deployment_id() - - return os.path.join(CONST.dir_rally_inst, - 'verification', - 'verifier-{}'.format(self.VERIFIER_ID), - 'for-deployment-{}'.format(self.DEPLOYMENT_ID)) - - @staticmethod def read_file(filename): with open(filename) as src: return [line.strip() for line in src.readlines()] @@ -297,10 +238,23 @@ class TempestCommon(testcase_base.TestcaseBase): num_success = new_line[2] elif 'Skipped' in new_line: num_skipped = new_line[2] + elif 'Failures' in new_line: + num_failures = new_line[2] try: num_executed = int(num_tests) - int(num_skipped) success_rate = 100 * int(num_success) / int(num_executed) + with open(os.path.join(conf_utils.TEMPEST_RESULTS_DIR, + "tempest.log"), 'r') as logfile: + output = logfile.read() + + error_logs = "" + for match in re.findall('(.*?)[. ]*FAILED', output): + error_logs += match + + self.details = {"tests": int(num_tests), + "failures": int(num_failures), + "errors": error_logs} except Exception: success_rate = 0 @@ -321,8 +275,7 @@ class TempestCommon(testcase_base.TestcaseBase): if res != testcase_base.TestcaseBase.EX_OK: return res - res = conf_utils.configure_tempest(logger, - self.DEPLOYMENT_DIR, + res = conf_utils.configure_tempest(self.DEPLOYMENT_DIR, self.IMAGE_ID, self.FLAVOR_ID) if res != testcase_base.TestcaseBase.EX_OK: @@ -341,10 +294,9 @@ class TempestCommon(testcase_base.TestcaseBase): self.stop_time = time.time() - if self.criteria == "PASS": - return testcase_base.TestcaseBase.EX_OK - else: - return testcase_base.TestcaseBase.EX_TESTCASE_FAILED + # If we are here, it means that the test case was successfully executed + # criteria is managed by the criteria Field + return testcase_base.TestcaseBase.EX_OK class TempestSmokeSerial(TempestCommon): @@ -380,7 +332,7 @@ class TempestMultisite(TempestCommon): self.case_name = "multisite" self.MODE = "feature_multisite" self.OPTION = "--concurrency 1" - conf_utils.configure_tempest_multisite(logger, self.DEPLOYMENT_DIR) + conf_utils.configure_tempest_multisite(self.DEPLOYMENT_DIR) class TempestCustom(TempestCommon): diff --git a/functest/opnfv_tests/sdn/odl/odl.py b/functest/opnfv_tests/sdn/odl/odl.py index 339c305e..9502be93 100755 --- a/functest/opnfv_tests/sdn/odl/odl.py +++ b/functest/opnfv_tests/sdn/odl/odl.py @@ -54,6 +54,7 @@ class ODLTests(testcase_base.TestcaseBase): "csit/suites/openstack/neutron") basic_suite_dir = os.path.join(odl_test_repo, "csit/suites/integration/basic") + default_suites = [basic_suite_dir, neutron_suite_dir] res_dir = '/home/opnfv/functest/results/odl/' logger = ft_logger.Logger("opendaylight").getLogger() @@ -89,8 +90,7 @@ class ODLTests(testcase_base.TestcaseBase): self.details['description'] = result.suite.name self.details['tests'] = visitor.get_data() - def main(self, **kwargs): - dirs = [self.basic_suite_dir, self.neutron_suite_dir] + def main(self, suites=default_suites, **kwargs): try: odlusername = kwargs['odlusername'] odlpassword = kwargs['odlpassword'] @@ -117,7 +117,7 @@ class ODLTests(testcase_base.TestcaseBase): stdout_file = os.path.join(self.res_dir, 'stdout.txt') output_dir = os.path.join(self.res_dir, 'output.xml') with open(stdout_file, 'w+') as stdout: - robot.run(*dirs, variable=variables, + robot.run(*suites, variable=variables, output=output_dir, log='NONE', report='NONE', @@ -140,8 +140,13 @@ class ODLTests(testcase_base.TestcaseBase): else: return self.EX_RUN_ERROR - def run(self): + def run(self, **kwargs): try: + suites = self.default_suites + try: + suites = kwargs["suites"] + except KeyError: + pass keystone_url = op_utils.get_endpoint(service_type='identity') neutron_url = op_utils.get_endpoint(service_type='network') kwargs = {'keystoneip': urlparse.urlparse(keystone_url).hostname} @@ -178,7 +183,7 @@ class ODLTests(testcase_base.TestcaseBase): self.logger.exception("Cannot run ODL testcases.") return self.EX_RUN_ERROR - return self.main(**kwargs) + return self.main(suites, **kwargs) class ODLParser(): @@ -228,7 +233,7 @@ if __name__ == '__main__': parser = ODLParser() args = parser.parse_args(sys.argv[1:]) try: - result = odl.main(**args) + result = odl.main(ODLTests.default_suites, **args) if result != testcase_base.TestcaseBase.EX_OK: sys.exit(result) if args['pushtodb']: diff --git a/functest/opnfv_tests/vnf/aaa/aaa.py b/functest/opnfv_tests/vnf/aaa/aaa.py index f1c265f4..f1c265f4 100644..100755 --- a/functest/opnfv_tests/vnf/aaa/aaa.py +++ b/functest/opnfv_tests/vnf/aaa/aaa.py diff --git a/functest/tests/unit/odl/test_odl.py b/functest/tests/unit/odl/test_odl.py index 568fdc82..810b3592 100644 --- a/functest/tests/unit/odl/test_odl.py +++ b/functest/tests/unit/odl/test_odl.py @@ -346,6 +346,28 @@ class ODLTesting(unittest.TestCase): self.test.main = mock.Mock(return_value=status) self.assertEqual(self.test.run(), status) self.test.main.assert_called_once_with( + odl.ODLTests.default_suites, + keystoneip=self._keystone_ip, neutronip=self._neutron_ip, + odlip=odlip, odlpassword=self._odl_password, + odlrestconfport=odlrestconfport, + odlusername=self._odl_username, odlwebport=odlwebport, + ospassword=self._os_password, ostenantname=self._os_tenantname, + osusername=self._os_username) + + def _test_run_defining_multiple_suites( + self, suites, + status=testcase_base.TestcaseBase.EX_OK, + exception=None, odlip="127.0.0.3", odlwebport="8080", + odlrestconfport="8181"): + with mock.patch('functest.utils.openstack_utils.get_endpoint', + side_effect=self._fake_url_for): + if exception: + self.test.main = mock.Mock(side_effect=exception) + else: + self.test.main = mock.Mock(return_value=status) + self.assertEqual(self.test.run(suites=suites), status) + self.test.main.assert_called_once_with( + suites, keystoneip=self._keystone_ip, neutronip=self._neutron_ip, odlip=odlip, odlpassword=self._odl_password, odlrestconfport=odlrestconfport, @@ -394,6 +416,14 @@ class ODLTesting(unittest.TestCase): odlip=self._sdn_controller_ip, odlwebport=self._odl_webport) + def test_run_redefining_suites(self): + os.environ["SDN_CONTROLLER_IP"] = self._sdn_controller_ip + self._test_run_defining_multiple_suites( + [odl.ODLTests.basic_suite_dir], + testcase_base.TestcaseBase.EX_OK, + odlip=self._sdn_controller_ip, + odlwebport=self._odl_webport) + def test_run_fuel(self): os.environ["INSTALLER_TYPE"] = "fuel" self._test_run(testcase_base.TestcaseBase.EX_OK, diff --git a/functest/tests/unit/test_logging.ini b/functest/tests/unit/test_logging.ini index 3d5b947c..492767d1 100644 --- a/functest/tests/unit/test_logging.ini +++ b/functest/tests/unit/test_logging.ini @@ -1,27 +1,27 @@ -[loggers]
-keys=root,functest_logger
-
-[logger_root]
-level=DEBUG
-handlers=console
-
-[logger_functest_logger]
-level=DEBUG
-handlers=console
-qualname=functest.utils.functest_logger
-propagate=0
-
-[handlers]
-keys=console
-
-[handler_console]
-class=StreamHandler
-level=INFO
-formatter=standard
-args=(sys.stdout,)
-
-[formatters]
-keys=standard
-
-[formatter_standard]
+[loggers] +keys=root,functest_logger + +[logger_root] +level=DEBUG +handlers=console + +[logger_functest_logger] +level=DEBUG +handlers=console +qualname=functest.utils.functest_logger +propagate=0 + +[handlers] +keys=console + +[handler_console] +class=StreamHandler +level=INFO +formatter=standard +args=(sys.stdout,) + +[formatters] +keys=standard + +[formatter_standard] format=%(asctime)s - %(name)s - %(levelname)s - %(message)s
\ No newline at end of file diff --git a/functest/tests/unit/utils/test_openstack_tacker.py b/functest/tests/unit/utils/test_openstack_tacker.py index a8330c0e..dc717258 100644 --- a/functest/tests/unit/utils/test_openstack_tacker.py +++ b/functest/tests/unit/utils/test_openstack_tacker.py @@ -146,8 +146,7 @@ class OSTackerTesting(unittest.TestCase): tosca_file=None) self.assertEqual(resp, self.createvnfd) - @mock.patch('functest.utils.openstack_tacker.logger.error') - def test_create_vnfd_default(self, mock_logger_error): + def test_create_vnfd_default(self): with mock.patch.object(self.tacker_client, 'create_vnfd', return_value=self.createvnfd), \ mock.patch('__builtin__.open', mock.mock_open(read_data='1')) \ @@ -155,16 +154,15 @@ class OSTackerTesting(unittest.TestCase): resp = openstack_tacker.create_vnfd(self.tacker_client, tosca_file=self.tosca_file) m.assert_called_once_with(self.tosca_file) - mock_logger_error.assert_called_once_with('1') self.assertEqual(resp, self.createvnfd) - @mock.patch('functest.utils.openstack_tacker.logger.exception') - def test_create_vnfd_exception(self, mock_logger_excep): + @mock.patch('functest.utils.openstack_tacker.logger.error') + def test_create_vnfd_exception(self, mock_logger_error): with mock.patch.object(self.tacker_client, 'create_vnfd', side_effect=Exception): resp = openstack_tacker.create_vnfd(self.tacker_client, tosca_file=self.tosca_file) - mock_logger_excep.assert_called_once_with(test_utils. + mock_logger_error.assert_called_once_with(test_utils. SubstrMatch("Error" " [create" "_vnfd(" diff --git a/functest/utils/constants.py b/functest/utils/constants.py index 2e8eb3f4..2e8eb3f4 100644..100755 --- a/functest/utils/constants.py +++ b/functest/utils/constants.py diff --git a/functest/utils/functest_constants.py b/functest/utils/functest_constants.py index 3cc744c6..0a9a03b1 100644 --- a/functest/utils/functest_constants.py +++ b/functest/utils/functest_constants.py @@ -195,35 +195,6 @@ ONOS_INSTALLER_MASTER_USERNAME = get_value( ONOS_INSTALLER_MASTER_PASSWORD = get_value( 'ONOS.environment.installer_master_password', 'ONOS_INSTALLER_MASTER_PASSWORD') -PROMISE_REPO_DIR = get_value('general.dir.dir_repo_promise', - 'PROMISE_REPO_DIR') -PROMISE_TENANT_NAME = get_value('promise.tenant_name', - 'PROMISE_TENANT_NAME') -TENANT_DESCRIPTION = get_value('promise.tenant_description', - 'TENANT_DESCRIPTION') -PROMISE_USER_NAME = get_value('promise.user_name', 'PROMISE_USER_NAME') -PROMISE_USER_PWD = get_value('promise.user_pwd', 'PROMISE_USER_PWD') -PROMISE_IMAGE_NAME = get_value('promise.image_name', - 'PROMISE_IMAGE_NAME') -PROMISE_FLAVOR_NAME = get_value('promise.flavor_name', - 'PROMISE_FLAVOR_NAME') -PROMISE_FLAVOR_VCPUS = get_value('promise.flavor_vcpus', - 'PROMISE_FLAVOR_VCPUS') -PROMISE_FLAVOR_RAM = get_value('promise.flavor_ram', - 'PROMISE_FLAVOR_RAM') -PROMISE_FLAVOR_DISK = get_value('promise.flavor_disk', - 'PROMISE_FLAVOR_DISK') -PROMISE_NET_NAME = get_value('promise.network_name', 'PROMISE_NET_NAME') -PROMISE_SUBNET_NAME = get_value('promise.subnet_name', - 'PROMISE_SUBNET_NAME') -PROMISE_SUBNET_CIDR = get_value('promise.subnet_cidr', - 'PROMISE_SUBNET_CIDR') -PROMISE_ROUTER_NAME = get_value('promise.router_name', - 'PROMISE_ROUTER_NAME') -DOCTOR_REPO_DIR = get_value('general.dir.dir_repo_doctor', - 'DOCTOR_REPO_DIR') -COPPER_REPO_DIR = get_value('general.dir.repo_copper', - 'COPPER_REPO_DIR') EXAMPLE_INSTANCE_NAME = get_value('example.vm_name', 'EXAMPLE_INSTANCE_NAME') EXAMPLE_FLAVOR = get_value('example.flavor', 'EXAMPLE_FLAVOR') diff --git a/functest/utils/openstack_tacker.py b/functest/utils/openstack_tacker.py index f17b421e..c7ac89af 100644 --- a/functest/utils/openstack_tacker.py +++ b/functest/utils/openstack_tacker.py @@ -20,9 +20,9 @@ import time logger = ft_logger.Logger("tacker_utils").getLogger() -def get_tacker_client(): - creds_tacker = os_utils.get_credentials() - return tackerclient.Client(**creds_tacker) +def get_tacker_client(other_creds={}): + sess = os_utils.get_session(other_creds) + return tackerclient.Client(session=sess) # ********************************************* @@ -74,12 +74,12 @@ def create_vnfd(tacker_client, tosca_file=None): if tosca_file is not None: with open(tosca_file) as tosca_fd: vnfd_body = tosca_fd.read() - logger.error(vnfd_body) + logger.info('VNFD template:\n{0}'.format(vnfd_body)) return tacker_client.create_vnfd( body={"vnfd": {"attributes": {"vnfd": vnfd_body}}}) except Exception, e: - logger.exception("Error [create_vnfd(tacker_client, '%s')]: %s" - % (tosca_file, e)) + logger.error("Error [create_vnfd(tacker_client, '%s')]: %s" + % (tosca_file, e)) return None @@ -124,7 +124,8 @@ def create_vnf(tacker_client, vnf_name, vnfd_id=None, vnfd_name=None): vnf_body['vnf']['vnfd_id'] = get_vnfd_id(tacker_client, vnfd_name) return tacker_client.create_vnf(body=vnf_body) except Exception, e: - logger.error("error [create_vnf(tacker_client, '%s', '%s', '%s')]: %s" + logger.error("error [create_vnf(tacker_client," + " '%s', '%s', '%s')]: %s" % (vnf_name, vnfd_id, vnfd_name, e)) return None @@ -206,7 +207,8 @@ def create_sfc(tacker_client, sfc_name, for name in chain_vnf_names] return tacker_client.create_sfc(body=sfc_body) except Exception, e: - logger.error("error [create_sfc(tacker_client, '%s', '%s', '%s')]: %s" + logger.error("error [create_sfc(tacker_client," + " '%s', '%s', '%s')]: %s" % (sfc_name, chain_vnf_ids, chain_vnf_names, e)) return None @@ -263,8 +265,8 @@ def create_sfc_classifier(tacker_client, sfc_clf_name, sfc_id=None, tacker_client, sfc_name) return tacker_client.create_sfc_classifier(body=sfc_clf_body) except Exception, e: - logger.error("error [create_sfc_classifier(tacker_client, '%s', '%s'," - " '%s', '%s')]: '%s'" + logger.error("error [create_sfc_classifier(tacker_client," + " '%s', '%s','%s', '%s')]: '%s'" % (sfc_clf_name, sfc_id, sfc_name, str(match), e)) return None |