aboutsummaryrefslogtreecommitdiffstats
path: root/functest
diff options
context:
space:
mode:
Diffstat (limited to 'functest')
-rwxr-xr-xfunctest/ci/config_functest.yaml4
-rwxr-xr-xfunctest/ci/exec_test.sh13
-rw-r--r--functest/ci/logging.json58
-rwxr-xr-xfunctest/ci/prepare_env.py9
-rwxr-xr-xfunctest/ci/run_tests.py6
-rwxr-xr-xfunctest/ci/testcases.yaml34
-rw-r--r--functest/core/pytest_suite_runner.py8
-rwxr-xr-xfunctest/opnfv_tests/features/doctor.py81
-rwxr-xr-xfunctest/opnfv_tests/features/promise.py260
-rwxr-xr-xfunctest/opnfv_tests/openstack/healthcheck/healthcheck.sh16
-rw-r--r--functest/opnfv_tests/openstack/rally/rally.py14
-rw-r--r--functest/opnfv_tests/openstack/tempest/conf_utils.py77
-rw-r--r--functest/opnfv_tests/openstack/tempest/tempest.py96
-rwxr-xr-xfunctest/opnfv_tests/sdn/odl/odl.py17
-rwxr-xr-x[-rw-r--r--]functest/opnfv_tests/vnf/aaa/aaa.py0
-rw-r--r--functest/tests/unit/odl/test_odl.py30
-rw-r--r--functest/tests/unit/test_logging.ini52
-rw-r--r--functest/tests/unit/utils/test_openstack_tacker.py10
-rwxr-xr-x[-rw-r--r--]functest/utils/constants.py0
-rw-r--r--functest/utils/functest_constants.py29
-rw-r--r--functest/utils/openstack_tacker.py22
21 files changed, 291 insertions, 545 deletions
diff --git a/functest/ci/config_functest.yaml b/functest/ci/config_functest.yaml
index 2820f8b0..f6cb14cb 100755
--- a/functest/ci/config_functest.yaml
+++ b/functest/ci/config_functest.yaml
@@ -20,8 +20,8 @@ general:
repo_sdnvpn: /home/opnfv/repos/sdnvpn
repo_sfc: /home/opnfv/repos/sfc
dir_repo_onos: /home/opnfv/repos/onos
- dir_repo_promise: /home/opnfv/repos/promise
- dir_repo_doctor: /home/opnfv/repos/doctor
+ repo_promise: /home/opnfv/repos/promise
+ repo_doctor: /home/opnfv/repos/doctor
repo_copper: /home/opnfv/repos/copper
dir_repo_ovno: /home/opnfv/repos/ovno
repo_parser: /home/opnfv/repos/parser
diff --git a/functest/ci/exec_test.sh b/functest/ci/exec_test.sh
index 54a7c624..b288fe36 100755
--- a/functest/ci/exec_test.sh
+++ b/functest/ci/exec_test.sh
@@ -90,22 +90,12 @@ function run_test(){
--osusername ${OS_USERNAME} \
${args}
;;
- "vims")
- python ${FUNCTEST_TEST_DIR}/vnf/ims/vims.py $clean_flag $report
- ;;
"onos")
python ${FUNCTEST_TEST_DIR}/sdn/onos/teston/onos.py
;;
"onos_sfc")
python ${FUNCTEST_TEST_DIR}/sdn/onos/teston/onos.py -t sfc
;;
- "promise")
- python ${FUNCTEST_TEST_DIR}/features/promise.py $report
- sleep 10 # to let the instances terminate
- ;;
- "doctor")
- python ${FUNCTEST_TEST_DIR}/features/doctor.py $report
- ;;
"ovno")
# suite under rewritting for colorado
# no need to run anything until refactoring done
@@ -116,9 +106,6 @@ function run_test(){
source ${FUNCTEST_CONF_DIR}/stackrc
python ${FUNCTEST_TEST_DIR}/security_scan/security_scan.py --config ${FUNCTEST_TEST_DIR}/security_scan/config.ini
;;
- "copper")
- python ${FUNCTEST_TEST_DIR}/features/copper.py $report
- ;;
"moon")
python ${REPOS_DIR}/moon/tests/run_tests.py $report
;;
diff --git a/functest/ci/logging.json b/functest/ci/logging.json
index 3f454e8f..2a2399d3 100644
--- a/functest/ci/logging.json
+++ b/functest/ci/logging.json
@@ -1,29 +1,29 @@
-{
- "version": 1,
- "disable_existing_loggers": false,
- "formatters": {
- "standard": {
- "format": "%(asctime)s - %(name)s - %(levelname)s - %(message)s"
- }
- },
- "handlers": {
- "console": {
- "level": "INFO",
- "class": "logging.StreamHandler",
- "formatter": "standard"
- },
- "file": {
- "level": "DEBUG",
- "class": "logging.FileHandler",
- "formatter": "standard",
- "filename": "/home/opnfv/functest/results/functest.log"
- }
- },
- "loggers": {
- "": {
- "handlers": ["console", "file"],
- "level": "DEBUG",
- "propagate": "yes"
- }
- }
-}
+{
+ "version": 1,
+ "disable_existing_loggers": false,
+ "formatters": {
+ "standard": {
+ "format": "%(asctime)s - %(name)s - %(levelname)s - %(message)s"
+ }
+ },
+ "handlers": {
+ "console": {
+ "level": "INFO",
+ "class": "logging.StreamHandler",
+ "formatter": "standard"
+ },
+ "file": {
+ "level": "DEBUG",
+ "class": "logging.FileHandler",
+ "formatter": "standard",
+ "filename": "/home/opnfv/functest/results/functest.log"
+ }
+ },
+ "loggers": {
+ "": {
+ "handlers": ["console", "file"],
+ "level": "DEBUG",
+ "propagate": "yes"
+ }
+ }
+}
diff --git a/functest/ci/prepare_env.py b/functest/ci/prepare_env.py
index 8bbdf18b..b3e59020 100755
--- a/functest/ci/prepare_env.py
+++ b/functest/ci/prepare_env.py
@@ -261,6 +261,14 @@ def install_tempest():
error_msg="Problem while installing Tempest.")
+def create_flavor():
+ os_utils.get_or_create_flavor('m1.tiny',
+ '512',
+ '1',
+ '1',
+ public=True)
+
+
def check_environment():
msg_not_active = "The Functest environment is not installed."
if not os.path.isfile(CONST.env_active):
@@ -290,6 +298,7 @@ def main(**kwargs):
verify_deployment()
install_rally()
install_tempest()
+ create_flavor()
with open(CONST.env_active, "w") as env_file:
env_file.write("1")
diff --git a/functest/ci/run_tests.py b/functest/ci/run_tests.py
index ef080016..320102dd 100755
--- a/functest/ci/run_tests.py
+++ b/functest/ci/run_tests.py
@@ -148,7 +148,11 @@ def run_test(test, tier_name, testcases=None):
module = importlib.import_module(run_dict['module'])
cls = getattr(module, run_dict['class'])
test_case = cls()
- result = test_case.run()
+ try:
+ kwargs = run_dict['args']
+ result = test_case.run(**kwargs)
+ except KeyError:
+ result = test_case.run()
if result == testcase_base.TestcaseBase.EX_OK:
if GlobalVariables.REPORT_FLAG:
test_case.publish_report()
diff --git a/functest/ci/testcases.yaml b/functest/ci/testcases.yaml
index f1cce489..12d791f0 100755
--- a/functest/ci/testcases.yaml
+++ b/functest/ci/testcases.yaml
@@ -99,6 +99,10 @@ tiers:
run:
module: 'functest.opnfv_tests.sdn.odl.odl'
class: 'ODLTests'
+ args:
+ suites:
+ - /home/opnfv/repos/odl_test/csit/suites/integration/basic
+ - /home/opnfv/repos/odl_test/csit/suites/openstack/neutron
-
name: onos
@@ -185,6 +189,9 @@ tiers:
dependencies:
installer: '(fuel)|(joid)'
scenario: ''
+ run:
+ module: 'functest.opnfv_tests.features.promise'
+ class: 'Promise'
-
name: doctor
@@ -195,6 +202,9 @@ tiers:
dependencies:
installer: '(apex)|(fuel)|(joid)'
scenario: '^((?!fdio).)*$'
+ run:
+ module: 'functest.opnfv_tests.features.doctor'
+ class: 'Doctor'
-
name: bgpvpn
@@ -218,18 +228,18 @@ tiers:
dependencies:
installer: 'apex'
scenario: '^((?!fdio).)*$'
- -
- name: copper
- criteria: 'status == "PASS"'
- blocking: false
- description: >-
- Test suite for policy management based on OpenStack Congress
- dependencies:
- installer: '(apex)|(joid)'
- scenario: '^((?!fdio|lxd).)*$'
- run:
- module: 'functest.opnfv_tests.features.copper'
- class: 'Copper'
+# -
+# name: copper
+# criteria: 'status == "PASS"'
+# blocking: false
+# description: >-
+# Test suite for policy management based on OpenStack Congress
+# dependencies:
+# installer: '(apex)|(joid)'
+# scenario: '^((?!fdio|lxd).)*$'
+# run:
+# module: 'functest.opnfv_tests.features.copper'
+# class: 'Copper'
-
name: moon
criteria: 'status == "PASS"'
diff --git a/functest/core/pytest_suite_runner.py b/functest/core/pytest_suite_runner.py
index 1eed92b5..c168d7d9 100644
--- a/functest/core/pytest_suite_runner.py
+++ b/functest/core/pytest_suite_runner.py
@@ -41,14 +41,18 @@ class PyTestSuiteRunner(base.TestcaseBase):
for test, message in result.failures:
self.logger.error(str(test) + " FAILED with " + message)
+ # a result can be PASS or FAIL
+ # But in this case it means that the Execution was OK
+ # we shall distinguish Execution Error from FAIL results
+ # TestcaseBase.EX_RUN_ERROR means that the test case was not run
+ # not that it was run but the result was FAIL
+ exit_code = base.TestcaseBase.EX_OK
if ((result.errors and len(result.errors) > 0)
or (result.failures and len(result.failures) > 0)):
self.logger.info("%s FAILED" % self.case_name)
self.criteria = 'FAIL'
- exit_code = base.TestcaseBase.EX_RUN_ERROR
else:
self.logger.info("%s OK" % self.case_name)
- exit_code = base.TestcaseBase.EX_OK
self.criteria = 'PASS'
self.details = {}
diff --git a/functest/opnfv_tests/features/doctor.py b/functest/opnfv_tests/features/doctor.py
index dbd803a6..4d295a67 100755
--- a/functest/opnfv_tests/features/doctor.py
+++ b/functest/opnfv_tests/features/doctor.py
@@ -1,6 +1,6 @@
#!/usr/bin/python
#
-# Copyright (c) 2015 All rights reserved
+# Copyright (c) 2017 All rights reserved
# This program and the accompanying materials
# are made available under the terms of the Apache License, Version 2.0
# which accompanies this distribution, and is available at
@@ -13,77 +13,12 @@
# 0.2: measure test duration and publish results under json format
#
#
-import argparse
-import os
-import time
+import functest.core.feature_base as base
-import functest.utils.functest_logger as ft_logger
-import functest.utils.functest_utils as functest_utils
-import functest.utils.functest_constants as ft_constants
-parser = argparse.ArgumentParser()
-parser.add_argument("-r", "--report",
- help="Create json result file",
- action="store_true")
-args = parser.parse_args()
-
-functest_yaml = functest_utils.get_functest_yaml()
-
-DOCTOR_REPO_DIR = ft_constants.DOCTOR_REPO_DIR
-RESULTS_DIR = ft_constants.FUNCTEST_RESULTS_DIR
-
-logger = ft_logger.Logger("doctor").getLogger()
-
-
-def main():
- exit_code = -1
-
- # if the image name is explicitly set for the doctor suite, set it as
- # enviroment variable
- if 'doctor' in functest_yaml and 'image_name' in functest_yaml['doctor']:
- os.environ["IMAGE_NAME"] = functest_yaml['doctor']['image_name']
-
- cmd = 'cd %s/tests && ./run.sh' % DOCTOR_REPO_DIR
- log_file = RESULTS_DIR + "/doctor.log"
-
- start_time = time.time()
-
- ret = functest_utils.execute_command(cmd,
- info=True,
- output_file=log_file)
-
- stop_time = time.time()
- duration = round(stop_time - start_time, 1)
- if ret == 0:
- logger.info("Doctor test case OK")
- test_status = 'OK'
- exit_code = 0
- else:
- logger.info("Doctor test case FAILED")
- test_status = 'NOK'
-
- details = {
- 'timestart': start_time,
- 'duration': duration,
- 'status': test_status,
- }
- status = "FAIL"
- if details['status'] == "OK":
- status = "PASS"
- functest_utils.logger_test_results("Doctor",
- "doctor-notification",
- status, details)
- if args.report:
- functest_utils.push_results_to_db("doctor",
- "doctor-notification",
- start_time,
- stop_time,
- status,
- details)
- logger.info("Doctor results pushed to DB")
-
- exit(exit_code)
-
-
-if __name__ == '__main__':
- main()
+class Doctor(base.FeatureBase):
+ def __init__(self):
+ super(Doctor, self).__init__(project='doctor',
+ case='doctor-notification',
+ repo='dir_repo_doctor')
+ self.cmd = 'cd %s/tests && ./run.sh' % self.repo
diff --git a/functest/opnfv_tests/features/promise.py b/functest/opnfv_tests/features/promise.py
index 0a30e3a1..15636fbf 100755
--- a/functest/opnfv_tests/features/promise.py
+++ b/functest/opnfv_tests/features/promise.py
@@ -1,252 +1,24 @@
#!/usr/bin/python
#
-# Copyright (c) 2015 All rights reserved
-# This program and the accompanying materials
-# are made available under the terms of the Apache License, Version 2.0
-# which accompanies this distribution, and is available at
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
-# Maintainer : jose.lausuch@ericsson.com
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
#
-import argparse
-import json
-import os
-import subprocess
-import time
+import functest.core.feature_base as base
-import functest.utils.functest_logger as ft_logger
-import functest.utils.functest_utils as ft_utils
-import functest.utils.openstack_utils as openstack_utils
-import keystoneclient.v2_0.client as ksclient
-from neutronclient.v2_0 import client as ntclient
-import novaclient.client as nvclient
-import functest.utils.functest_constants as ft_constants
-parser = argparse.ArgumentParser()
-
-parser.add_argument("-d", "--debug", help="Debug mode", action="store_true")
-parser.add_argument("-r", "--report",
- help="Create json result file",
- action="store_true")
-args = parser.parse_args()
-
-
-PROMISE_REPO_DIR = ft_constants.PROMISE_REPO_DIR
-RESULTS_DIR = ft_constants.FUNCTEST_RESULTS_DIR
-
-PROMISE_TENANT_NAME = ft_constants.PROMISE_TENANT_NAME
-TENANT_DESCRIPTION = ft_constants.TENANT_DESCRIPTION
-PROMISE_USER_NAME = ft_constants.PROMISE_USER_NAME
-PROMISE_USER_PWD = ft_constants.PROMISE_USER_PWD
-PROMISE_IMAGE_NAME = ft_constants.PROMISE_IMAGE_NAME
-PROMISE_FLAVOR_NAME = ft_constants.PROMISE_FLAVOR_NAME
-PROMISE_FLAVOR_VCPUS = ft_constants.PROMISE_FLAVOR_VCPUS
-PROMISE_FLAVOR_RAM = ft_constants.PROMISE_FLAVOR_RAM
-PROMISE_FLAVOR_DISK = ft_constants.PROMISE_FLAVOR_DISK
-
-
-GLANCE_IMAGE_FILENAME = ft_constants.GLANCE_IMAGE_FILENAME
-GLANCE_IMAGE_FORMAT = ft_constants.GLANCE_IMAGE_FORMAT
-GLANCE_IMAGE_PATH = os.path.join(ft_constants.FUNCTEST_DATA_DIR,
- GLANCE_IMAGE_FILENAME)
-
-PROMISE_NET_NAME = ft_constants.PROMISE_NET_NAME
-PROMISE_SUBNET_NAME = ft_constants.PROMISE_SUBNET_NAME
-PROMISE_SUBNET_CIDR = ft_constants.PROMISE_SUBNET_CIDR
-PROMISE_ROUTER_NAME = ft_constants.PROMISE_ROUTER_NAME
-
-
-""" logging configuration """
-logger = ft_logger.Logger("promise").getLogger()
-
-
-def main():
- exit_code = -1
- start_time = time.time()
- ks_creds = openstack_utils.get_credentials("keystone")
- nv_creds = openstack_utils.get_credentials("nova")
- nt_creds = openstack_utils.get_credentials("neutron")
-
- keystone = ksclient.Client(**ks_creds)
-
- user_id = openstack_utils.get_user_id(keystone, ks_creds['username'])
- if user_id == '':
- logger.error("Error : Failed to get id of %s user" %
- ks_creds['username'])
- exit(-1)
-
- logger.info("Creating tenant '%s'..." % PROMISE_TENANT_NAME)
- tenant_id = openstack_utils.create_tenant(
- keystone, PROMISE_TENANT_NAME, TENANT_DESCRIPTION)
- if not tenant_id:
- logger.error("Error : Failed to create %s tenant"
- % PROMISE_TENANT_NAME)
- exit(-1)
- logger.debug("Tenant '%s' created successfully." % PROMISE_TENANT_NAME)
-
- roles_name = ["admin", "Admin"]
- role_id = ''
- for role_name in roles_name:
- if role_id == '':
- role_id = openstack_utils.get_role_id(keystone, role_name)
-
- if role_id == '':
- logger.error("Error : Failed to get id for %s role" % role_name)
- exit(-1)
-
- logger.info("Adding role '%s' to tenant '%s'..."
- % (role_id, PROMISE_TENANT_NAME))
- if not openstack_utils.add_role_user(keystone, user_id,
- role_id, tenant_id):
- logger.error("Error : Failed to add %s on tenant %s" %
- (ks_creds['username'], PROMISE_TENANT_NAME))
- exit(-1)
- logger.debug("Role added successfully.")
-
- logger.info("Creating user '%s'..." % PROMISE_USER_NAME)
- user_id = openstack_utils.create_user(
- keystone, PROMISE_USER_NAME, PROMISE_USER_PWD, None, tenant_id)
-
- if not user_id:
- logger.error("Error : Failed to create %s user" % PROMISE_USER_NAME)
- exit(-1)
- logger.debug("User '%s' created successfully." % PROMISE_USER_NAME)
-
- logger.info("Updating OpenStack credentials...")
- ks_creds.update({
- "username": PROMISE_TENANT_NAME,
- "password": PROMISE_TENANT_NAME,
- "tenant_name": PROMISE_TENANT_NAME,
- })
-
- nt_creds.update({
- "tenant_name": PROMISE_TENANT_NAME,
- })
-
- nv_creds.update({
- "project_id": PROMISE_TENANT_NAME,
- })
-
- glance = openstack_utils.get_glance_client()
- nova = nvclient.Client("2", **nv_creds)
-
- logger.info("Creating image '%s' from '%s'..." % (PROMISE_IMAGE_NAME,
- GLANCE_IMAGE_PATH))
- image_id = openstack_utils.create_glance_image(glance,
- PROMISE_IMAGE_NAME,
- GLANCE_IMAGE_PATH)
- if not image_id:
- logger.error("Failed to create the Glance image...")
- exit(-1)
- logger.debug("Image '%s' with ID '%s' created successfully."
- % (PROMISE_IMAGE_NAME, image_id))
- flavor_id = openstack_utils.get_flavor_id(nova, PROMISE_FLAVOR_NAME)
- if flavor_id == '':
- logger.info("Creating flavor '%s'..." % PROMISE_FLAVOR_NAME)
- flavor_id = openstack_utils.create_flavor(nova,
- PROMISE_FLAVOR_NAME,
- PROMISE_FLAVOR_RAM,
- PROMISE_FLAVOR_DISK,
- PROMISE_FLAVOR_VCPUS)
- if not flavor_id:
- logger.error("Failed to create the Flavor...")
- exit(-1)
- logger.debug("Flavor '%s' with ID '%s' created successfully." %
- (PROMISE_FLAVOR_NAME, flavor_id))
- else:
- logger.debug("Using existing flavor '%s' with ID '%s'..."
- % (PROMISE_FLAVOR_NAME, flavor_id))
-
- neutron = ntclient.Client(**nt_creds)
-
- network_dic = openstack_utils.create_network_full(neutron,
- PROMISE_NET_NAME,
- PROMISE_SUBNET_NAME,
- PROMISE_ROUTER_NAME,
- PROMISE_SUBNET_CIDR)
- if not network_dic:
- logger.error("Failed to create the private network...")
- exit(-1)
-
- logger.info("Exporting environment variables...")
- os.environ["NODE_ENV"] = "functest"
- os.environ["OS_PASSWORD"] = PROMISE_USER_PWD
- os.environ["OS_TEST_IMAGE"] = image_id
- os.environ["OS_TEST_FLAVOR"] = flavor_id
- os.environ["OS_TEST_NETWORK"] = network_dic["net_id"]
- os.environ["OS_TENANT_NAME"] = PROMISE_TENANT_NAME
- os.environ["OS_USERNAME"] = PROMISE_USER_NAME
-
- os.chdir(PROMISE_REPO_DIR + '/source/')
- results_file_name = os.path.join(RESULTS_DIR, 'promise-results.json')
- results_file = open(results_file_name, 'w+')
- cmd = 'npm run -s test -- --reporter json'
-
- logger.info("Running command: %s" % cmd)
- ret = subprocess.call(cmd, shell=True, stdout=results_file,
- stderr=subprocess.STDOUT)
- results_file.close()
-
- if ret == 0:
- logger.info("The test succeeded.")
- # test_status = 'OK'
- else:
- logger.info("The command '%s' failed." % cmd)
- # test_status = "Failed"
-
- # Print output of file
- with open(results_file_name, 'r') as results_file:
- data = results_file.read()
- logger.debug("\n%s" % data)
- json_data = json.loads(data)
-
- suites = json_data["stats"]["suites"]
- tests = json_data["stats"]["tests"]
- passes = json_data["stats"]["passes"]
- pending = json_data["stats"]["pending"]
- failures = json_data["stats"]["failures"]
- start_time_json = json_data["stats"]["start"]
- end_time = json_data["stats"]["end"]
- duration = float(json_data["stats"]["duration"]) / float(1000)
-
- logger.info("\n"
- "****************************************\n"
- " Promise test report\n\n"
- "****************************************\n"
- " Suites: \t%s\n"
- " Tests: \t%s\n"
- " Passes: \t%s\n"
- " Pending: \t%s\n"
- " Failures:\t%s\n"
- " Start: \t%s\n"
- " End: \t%s\n"
- " Duration:\t%s\n"
- "****************************************\n\n"
- % (suites, tests, passes, pending, failures,
- start_time_json, end_time, duration))
-
- if args.report:
- stop_time = time.time()
- json_results = {"timestart": start_time, "duration": duration,
- "tests": int(tests), "failures": int(failures)}
- logger.debug("Promise Results json: " + str(json_results))
-
- # criteria for Promise in Release B was 100% of tests OK
- status = "FAIL"
- if int(tests) > 32 and int(failures) < 1:
- status = "PASS"
- exit_code = 0
-
- ft_utils.push_results_to_db("promise",
- "promise",
- start_time,
- stop_time,
- status,
- json_results)
-
- exit(exit_code)
-
-
-if __name__ == '__main__':
- main()
+class Promise(base.FeatureBase):
+ def __init__(self):
+ super(Promise, self).__init__(project='promise',
+ case='promise',
+ repo='dir_repo_promise')
+ dir_promise_functest = '{}/promise/test/functest'.format(self.repo)
+ self.cmd = 'cd %s && python ./run_tests.py' % dir_promise_functest
diff --git a/functest/opnfv_tests/openstack/healthcheck/healthcheck.sh b/functest/opnfv_tests/openstack/healthcheck/healthcheck.sh
index 57aa0c70..7fa957c0 100755
--- a/functest/opnfv_tests/openstack/healthcheck/healthcheck.sh
+++ b/functest/opnfv_tests/openstack/healthcheck/healthcheck.sh
@@ -23,17 +23,17 @@ echo "">$LOG_FILE
exec 1<>$LOG_FILE
info () {
- echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healtcheck - INFO - " "$*" | tee -a $LOG_FILE 1>&2
+ echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healthcheck - INFO - " "$*" | tee -a $LOG_FILE 1>&2
}
debug () {
if [[ "${CI_DEBUG,,}" == "true" ]]; then
- echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healtcheck - DEBUG - " "$*" | tee -a $LOG_FILE 1>&2
+ echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healthcheck - DEBUG - " "$*" | tee -a $LOG_FILE 1>&2
fi
}
error () {
- echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healtcheck - ERROR - " "$*" | tee -a $LOG_FILE 1>&2
+ echo -e "$(date '+%Y-%m-%d %H:%M:%S,%3N') - healthcheck - ERROR - " "$*" | tee -a $LOG_FILE 1>&2
exit 1
}
@@ -125,16 +125,16 @@ kernel_img=$(cat ${YAML_FILE} | shyaml get-value healthcheck.kernel_image 2> /de
ramdisk_img=$(cat ${YAML_FILE} | shyaml get-value healthcheck.ramdisk_image 2> /dev/null || true)
extra_properties=$(cat ${YAML_FILE} | shyaml get-value healthcheck.extra_properties 2> /dev/null || true)
-# Test if we need to create a 3part image
+# Test if we need to create a 3party image
if [ "X$kernel_img" != "X" ]
then
- img_id=$(glance image-create --name ${kernel_image} --disk-format aki \
+ img_id=$(openstack image create ${kernel_image} --disk-format aki \
--container-format bare < ${kernel_img} | awk '$2 == "id" { print $4 }')
extra_opts="--property kernel_id=${img_id}"
if [ "X$ramdisk_img" != "X" ]
then
- img_id=$(glance image-create --name ${ramdisk_image} --disk-format ari \
+ img_id=$(openstack image create ${ramdisk_image} --disk-format ari \
--container-format bare < ${ramdisk_img} | awk '$2 == "id" { print $4 }')
extra_opts="$extra_opts --property ramdisk_id=${img_id}"
fi
@@ -152,10 +152,10 @@ fi
debug "image extra_properties=${extra_properties}"
-eval glance image-create --name ${image_1} --disk-format ${disk_format} --container-format bare \
+eval openstack image create ${image_1} --disk-format ${disk_format} --container-format bare \
${extra_opts} < ${disk_img}
debug "image '${image_1}' created."
-eval glance image-create --name ${image_2} --disk-format ${disk_format} --container-format bare \
+eval openstack image create ${image_2} --disk-format ${disk_format} --container-format bare \
${extra_opts} < ${disk_img}
debug "image '${image_2}' created."
info "... Glance OK!"
diff --git a/functest/opnfv_tests/openstack/rally/rally.py b/functest/opnfv_tests/openstack/rally/rally.py
index de553116..f984c368 100644
--- a/functest/opnfv_tests/openstack/rally/rally.py
+++ b/functest/opnfv_tests/openstack/rally/rally.py
@@ -94,11 +94,8 @@ class RallyBase(testcase_base.TestcaseBase):
else:
task_args['netid'] = ''
- auth_url = CONST.OS_AUTH_URL
- if auth_url is not None:
- task_args['request_url'] = auth_url.rsplit(":", 1)[0]
- else:
- task_args['request_url'] = ''
+ # get keystone auth endpoint
+ task_args['request_url'] = CONST.OS_AUTH_URL or ''
return task_args
@@ -533,10 +530,9 @@ class RallyBase(testcase_base.TestcaseBase):
return testcase_base.TestcaseBase.EX_RUN_ERROR
self.stop_time = time.time()
- if self.criteria == "PASS":
- return testcase_base.TestcaseBase.EX_OK
- else:
- return testcase_base.TestcaseBase.EX_TESTCASE_FAILED
+ # If we are here, it means that the test case was successfully executed
+ # criteria is managed by the criteria Field
+ return testcase_base.TestcaseBase.EX_OK
class RallySanity(RallyBase):
diff --git a/functest/opnfv_tests/openstack/tempest/conf_utils.py b/functest/opnfv_tests/openstack/tempest/conf_utils.py
index 88085d9a..0c9064a1 100644
--- a/functest/opnfv_tests/openstack/tempest/conf_utils.py
+++ b/functest/opnfv_tests/openstack/tempest/conf_utils.py
@@ -11,10 +11,12 @@ import ConfigParser
import os
import re
import shutil
+import subprocess
import opnfv.utils.constants as releng_constants
from functest.utils.constants import CONST
+import functest.utils.functest_logger as ft_logger
import functest.utils.functest_utils as ft_utils
import functest.utils.openstack_utils as os_utils
@@ -39,8 +41,74 @@ TEMPEST_LIST = os.path.join(TEMPEST_RESULTS_DIR, 'test_list.txt')
CI_INSTALLER_TYPE = CONST.INSTALLER_TYPE
CI_INSTALLER_IP = CONST.INSTALLER_IP
+""" logging configuration """
+logger = ft_logger.Logger("Tempest").getLogger()
-def configure_tempest(logger, deployment_dir, IMAGE_ID=None, FLAVOR_ID=None):
+
+def get_verifier_id():
+ """
+ Returns verifer id for current Tempest
+ """
+ cmd = ("rally verify list-verifiers | awk '/" +
+ CONST.tempest_deployment_name +
+ "/ {print $2}'")
+ p = subprocess.Popen(cmd, shell=True,
+ stdout=subprocess.PIPE,
+ stderr=subprocess.STDOUT)
+ deployment_uuid = p.stdout.readline().rstrip()
+ if deployment_uuid == "":
+ logger.error("Tempest verifier not found.")
+ raise Exception('Error with command:%s' % cmd)
+ return deployment_uuid
+
+
+def get_verifier_deployment_id():
+ """
+ Returns deployment id for active Rally deployment
+ """
+ cmd = ("rally deployment list | awk '/" +
+ CONST.rally_deployment_name +
+ "/ {print $2}'")
+ p = subprocess.Popen(cmd, shell=True,
+ stdout=subprocess.PIPE,
+ stderr=subprocess.STDOUT)
+ deployment_uuid = p.stdout.readline().rstrip()
+ if deployment_uuid == "":
+ logger.error("Rally deployment not found.")
+ raise Exception('Error with command:%s' % cmd)
+ return deployment_uuid
+
+
+def get_verifier_repo_dir(verifier_id):
+ """
+ Returns installed verfier repo directory for Tempest
+ """
+ if not verifier_id:
+ verifier_id = get_verifier_id()
+
+ return os.path.join(CONST.dir_rally_inst,
+ 'verification',
+ 'verifier-{}'.format(verifier_id),
+ 'repo')
+
+
+def get_verifier_deployment_dir(verifier_id, deployment_id):
+ """
+ Returns Rally deployment directory for current verifier
+ """
+ if not verifier_id:
+ verifier_id = get_verifier_id()
+
+ if not deployment_id:
+ deployment_id = get_verifier_deployment_id()
+
+ return os.path.join(CONST.dir_rally_inst,
+ 'verification',
+ 'verifier-{}'.format(verifier_id),
+ 'for-deployment-{}'.format(deployment_id))
+
+
+def configure_tempest(deployment_dir, IMAGE_ID=None, FLAVOR_ID=None):
"""
Add/update needed parameters into tempest.conf file generated by Rally
"""
@@ -104,18 +172,21 @@ def configure_tempest(logger, deployment_dir, IMAGE_ID=None, FLAVOR_ID=None):
config.write(config_file)
# Copy tempest.conf to /home/opnfv/functest/results/tempest/
+ if not os.path.exists(TEMPEST_RESULTS_DIR):
+ os.makedirs(TEMPEST_RESULTS_DIR)
+
shutil.copyfile(tempest_conf_file,
os.path.join(TEMPEST_RESULTS_DIR, 'tempest.conf'))
return releng_constants.EXIT_OK
-def configure_tempest_multisite(logger, deployment_dir):
+def configure_tempest_multisite(deployment_dir):
"""
Add/update needed parameters into tempest.conf file generated by Rally
"""
logger.debug("configure the tempest")
- configure_tempest(logger, deployment_dir)
+ configure_tempest(deployment_dir)
logger.debug("Finding tempest.conf file...")
tempest_conf_old = os.path.join(deployment_dir, 'tempest.conf')
diff --git a/functest/opnfv_tests/openstack/tempest/tempest.py b/functest/opnfv_tests/openstack/tempest/tempest.py
index 0014b718..9c19a147 100644
--- a/functest/opnfv_tests/openstack/tempest/tempest.py
+++ b/functest/opnfv_tests/openstack/tempest/tempest.py
@@ -35,74 +35,15 @@ class TempestCommon(testcase_base.TestcaseBase):
self.OPTION = ""
self.FLAVOR_ID = None
self.IMAGE_ID = None
- self.VERIFIER_ID = self.get_verifier_id()
- self.VERIFIER_REPO_DIR = self.get_verifier_repo_dir()
- self.DEPLOYMENT_ID = self.get_verifier_deployment_id()
- self.DEPLOYMENT_DIR = self.get_verifier_deployment_dir()
+ self.VERIFIER_ID = conf_utils.get_verifier_id()
+ self.VERIFIER_REPO_DIR = conf_utils.get_verifier_repo_dir(
+ self.VERIFIER_ID)
+ self.DEPLOYMENT_ID = conf_utils.get_verifier_deployment_id()
+ self.DEPLOYMENT_DIR = conf_utils.get_verifier_deployment_dir(
+ self.VERIFIER_ID, self.DEPLOYMENT_ID)
self.VERIFICATION_ID = None
@staticmethod
- def get_verifier_id():
- """
- Returns verifer id for current Tempest
- """
- cmd = ("rally verify list-verifiers | awk '/" +
- CONST.tempest_deployment_name +
- "/ {print $2}'")
- p = subprocess.Popen(cmd, shell=True,
- stdout=subprocess.PIPE,
- stderr=subprocess.STDOUT)
- deployment_uuid = p.stdout.readline().rstrip()
- if deployment_uuid == "":
- logger.error("Tempest verifier not found.")
- raise Exception('Error with command:%s' % cmd)
- return deployment_uuid
-
- @staticmethod
- def get_verifier_deployment_id():
- """
- Returns deployment id for active Rally deployment
- """
- cmd = ("rally deployment list | awk '/" +
- CONST.rally_deployment_name +
- "/ {print $2}'")
- p = subprocess.Popen(cmd, shell=True,
- stdout=subprocess.PIPE,
- stderr=subprocess.STDOUT)
- deployment_uuid = p.stdout.readline().rstrip()
- if deployment_uuid == "":
- logger.error("Rally deployment not found.")
- raise Exception('Error with command:%s' % cmd)
- return deployment_uuid
-
- def get_verifier_repo_dir(self):
- """
- Returns installed verfier repo directory for Tempest
- """
- if not self.VERIFIER_ID:
- self.VERIFIER_ID = self.get_verifier_id()
-
- return os.path.join(CONST.dir_rally_inst,
- 'verification',
- 'verifier-{}'.format(self.VERIFIER_ID),
- 'repo')
-
- def get_verifier_deployment_dir(self):
- """
- Returns Rally deployment directory for current verifier
- """
- if not self.VERIFIER_ID:
- self.VERIFIER_ID = self.get_verifier_id()
-
- if not self.DEPLOYMENT_ID:
- self.DEPLOYMENT_ID = self.get_verifier_deployment_id()
-
- return os.path.join(CONST.dir_rally_inst,
- 'verification',
- 'verifier-{}'.format(self.VERIFIER_ID),
- 'for-deployment-{}'.format(self.DEPLOYMENT_ID))
-
- @staticmethod
def read_file(filename):
with open(filename) as src:
return [line.strip() for line in src.readlines()]
@@ -297,10 +238,23 @@ class TempestCommon(testcase_base.TestcaseBase):
num_success = new_line[2]
elif 'Skipped' in new_line:
num_skipped = new_line[2]
+ elif 'Failures' in new_line:
+ num_failures = new_line[2]
try:
num_executed = int(num_tests) - int(num_skipped)
success_rate = 100 * int(num_success) / int(num_executed)
+ with open(os.path.join(conf_utils.TEMPEST_RESULTS_DIR,
+ "tempest.log"), 'r') as logfile:
+ output = logfile.read()
+
+ error_logs = ""
+ for match in re.findall('(.*?)[. ]*FAILED', output):
+ error_logs += match
+
+ self.details = {"tests": int(num_tests),
+ "failures": int(num_failures),
+ "errors": error_logs}
except Exception:
success_rate = 0
@@ -321,8 +275,7 @@ class TempestCommon(testcase_base.TestcaseBase):
if res != testcase_base.TestcaseBase.EX_OK:
return res
- res = conf_utils.configure_tempest(logger,
- self.DEPLOYMENT_DIR,
+ res = conf_utils.configure_tempest(self.DEPLOYMENT_DIR,
self.IMAGE_ID,
self.FLAVOR_ID)
if res != testcase_base.TestcaseBase.EX_OK:
@@ -341,10 +294,9 @@ class TempestCommon(testcase_base.TestcaseBase):
self.stop_time = time.time()
- if self.criteria == "PASS":
- return testcase_base.TestcaseBase.EX_OK
- else:
- return testcase_base.TestcaseBase.EX_TESTCASE_FAILED
+ # If we are here, it means that the test case was successfully executed
+ # criteria is managed by the criteria Field
+ return testcase_base.TestcaseBase.EX_OK
class TempestSmokeSerial(TempestCommon):
@@ -380,7 +332,7 @@ class TempestMultisite(TempestCommon):
self.case_name = "multisite"
self.MODE = "feature_multisite"
self.OPTION = "--concurrency 1"
- conf_utils.configure_tempest_multisite(logger, self.DEPLOYMENT_DIR)
+ conf_utils.configure_tempest_multisite(self.DEPLOYMENT_DIR)
class TempestCustom(TempestCommon):
diff --git a/functest/opnfv_tests/sdn/odl/odl.py b/functest/opnfv_tests/sdn/odl/odl.py
index 339c305e..9502be93 100755
--- a/functest/opnfv_tests/sdn/odl/odl.py
+++ b/functest/opnfv_tests/sdn/odl/odl.py
@@ -54,6 +54,7 @@ class ODLTests(testcase_base.TestcaseBase):
"csit/suites/openstack/neutron")
basic_suite_dir = os.path.join(odl_test_repo,
"csit/suites/integration/basic")
+ default_suites = [basic_suite_dir, neutron_suite_dir]
res_dir = '/home/opnfv/functest/results/odl/'
logger = ft_logger.Logger("opendaylight").getLogger()
@@ -89,8 +90,7 @@ class ODLTests(testcase_base.TestcaseBase):
self.details['description'] = result.suite.name
self.details['tests'] = visitor.get_data()
- def main(self, **kwargs):
- dirs = [self.basic_suite_dir, self.neutron_suite_dir]
+ def main(self, suites=default_suites, **kwargs):
try:
odlusername = kwargs['odlusername']
odlpassword = kwargs['odlpassword']
@@ -117,7 +117,7 @@ class ODLTests(testcase_base.TestcaseBase):
stdout_file = os.path.join(self.res_dir, 'stdout.txt')
output_dir = os.path.join(self.res_dir, 'output.xml')
with open(stdout_file, 'w+') as stdout:
- robot.run(*dirs, variable=variables,
+ robot.run(*suites, variable=variables,
output=output_dir,
log='NONE',
report='NONE',
@@ -140,8 +140,13 @@ class ODLTests(testcase_base.TestcaseBase):
else:
return self.EX_RUN_ERROR
- def run(self):
+ def run(self, **kwargs):
try:
+ suites = self.default_suites
+ try:
+ suites = kwargs["suites"]
+ except KeyError:
+ pass
keystone_url = op_utils.get_endpoint(service_type='identity')
neutron_url = op_utils.get_endpoint(service_type='network')
kwargs = {'keystoneip': urlparse.urlparse(keystone_url).hostname}
@@ -178,7 +183,7 @@ class ODLTests(testcase_base.TestcaseBase):
self.logger.exception("Cannot run ODL testcases.")
return self.EX_RUN_ERROR
- return self.main(**kwargs)
+ return self.main(suites, **kwargs)
class ODLParser():
@@ -228,7 +233,7 @@ if __name__ == '__main__':
parser = ODLParser()
args = parser.parse_args(sys.argv[1:])
try:
- result = odl.main(**args)
+ result = odl.main(ODLTests.default_suites, **args)
if result != testcase_base.TestcaseBase.EX_OK:
sys.exit(result)
if args['pushtodb']:
diff --git a/functest/opnfv_tests/vnf/aaa/aaa.py b/functest/opnfv_tests/vnf/aaa/aaa.py
index f1c265f4..f1c265f4 100644..100755
--- a/functest/opnfv_tests/vnf/aaa/aaa.py
+++ b/functest/opnfv_tests/vnf/aaa/aaa.py
diff --git a/functest/tests/unit/odl/test_odl.py b/functest/tests/unit/odl/test_odl.py
index 568fdc82..810b3592 100644
--- a/functest/tests/unit/odl/test_odl.py
+++ b/functest/tests/unit/odl/test_odl.py
@@ -346,6 +346,28 @@ class ODLTesting(unittest.TestCase):
self.test.main = mock.Mock(return_value=status)
self.assertEqual(self.test.run(), status)
self.test.main.assert_called_once_with(
+ odl.ODLTests.default_suites,
+ keystoneip=self._keystone_ip, neutronip=self._neutron_ip,
+ odlip=odlip, odlpassword=self._odl_password,
+ odlrestconfport=odlrestconfport,
+ odlusername=self._odl_username, odlwebport=odlwebport,
+ ospassword=self._os_password, ostenantname=self._os_tenantname,
+ osusername=self._os_username)
+
+ def _test_run_defining_multiple_suites(
+ self, suites,
+ status=testcase_base.TestcaseBase.EX_OK,
+ exception=None, odlip="127.0.0.3", odlwebport="8080",
+ odlrestconfport="8181"):
+ with mock.patch('functest.utils.openstack_utils.get_endpoint',
+ side_effect=self._fake_url_for):
+ if exception:
+ self.test.main = mock.Mock(side_effect=exception)
+ else:
+ self.test.main = mock.Mock(return_value=status)
+ self.assertEqual(self.test.run(suites=suites), status)
+ self.test.main.assert_called_once_with(
+ suites,
keystoneip=self._keystone_ip, neutronip=self._neutron_ip,
odlip=odlip, odlpassword=self._odl_password,
odlrestconfport=odlrestconfport,
@@ -394,6 +416,14 @@ class ODLTesting(unittest.TestCase):
odlip=self._sdn_controller_ip,
odlwebport=self._odl_webport)
+ def test_run_redefining_suites(self):
+ os.environ["SDN_CONTROLLER_IP"] = self._sdn_controller_ip
+ self._test_run_defining_multiple_suites(
+ [odl.ODLTests.basic_suite_dir],
+ testcase_base.TestcaseBase.EX_OK,
+ odlip=self._sdn_controller_ip,
+ odlwebport=self._odl_webport)
+
def test_run_fuel(self):
os.environ["INSTALLER_TYPE"] = "fuel"
self._test_run(testcase_base.TestcaseBase.EX_OK,
diff --git a/functest/tests/unit/test_logging.ini b/functest/tests/unit/test_logging.ini
index 3d5b947c..492767d1 100644
--- a/functest/tests/unit/test_logging.ini
+++ b/functest/tests/unit/test_logging.ini
@@ -1,27 +1,27 @@
-[loggers]
-keys=root,functest_logger
-
-[logger_root]
-level=DEBUG
-handlers=console
-
-[logger_functest_logger]
-level=DEBUG
-handlers=console
-qualname=functest.utils.functest_logger
-propagate=0
-
-[handlers]
-keys=console
-
-[handler_console]
-class=StreamHandler
-level=INFO
-formatter=standard
-args=(sys.stdout,)
-
-[formatters]
-keys=standard
-
-[formatter_standard]
+[loggers]
+keys=root,functest_logger
+
+[logger_root]
+level=DEBUG
+handlers=console
+
+[logger_functest_logger]
+level=DEBUG
+handlers=console
+qualname=functest.utils.functest_logger
+propagate=0
+
+[handlers]
+keys=console
+
+[handler_console]
+class=StreamHandler
+level=INFO
+formatter=standard
+args=(sys.stdout,)
+
+[formatters]
+keys=standard
+
+[formatter_standard]
format=%(asctime)s - %(name)s - %(levelname)s - %(message)s \ No newline at end of file
diff --git a/functest/tests/unit/utils/test_openstack_tacker.py b/functest/tests/unit/utils/test_openstack_tacker.py
index a8330c0e..dc717258 100644
--- a/functest/tests/unit/utils/test_openstack_tacker.py
+++ b/functest/tests/unit/utils/test_openstack_tacker.py
@@ -146,8 +146,7 @@ class OSTackerTesting(unittest.TestCase):
tosca_file=None)
self.assertEqual(resp, self.createvnfd)
- @mock.patch('functest.utils.openstack_tacker.logger.error')
- def test_create_vnfd_default(self, mock_logger_error):
+ def test_create_vnfd_default(self):
with mock.patch.object(self.tacker_client, 'create_vnfd',
return_value=self.createvnfd), \
mock.patch('__builtin__.open', mock.mock_open(read_data='1')) \
@@ -155,16 +154,15 @@ class OSTackerTesting(unittest.TestCase):
resp = openstack_tacker.create_vnfd(self.tacker_client,
tosca_file=self.tosca_file)
m.assert_called_once_with(self.tosca_file)
- mock_logger_error.assert_called_once_with('1')
self.assertEqual(resp, self.createvnfd)
- @mock.patch('functest.utils.openstack_tacker.logger.exception')
- def test_create_vnfd_exception(self, mock_logger_excep):
+ @mock.patch('functest.utils.openstack_tacker.logger.error')
+ def test_create_vnfd_exception(self, mock_logger_error):
with mock.patch.object(self.tacker_client, 'create_vnfd',
side_effect=Exception):
resp = openstack_tacker.create_vnfd(self.tacker_client,
tosca_file=self.tosca_file)
- mock_logger_excep.assert_called_once_with(test_utils.
+ mock_logger_error.assert_called_once_with(test_utils.
SubstrMatch("Error"
" [create"
"_vnfd("
diff --git a/functest/utils/constants.py b/functest/utils/constants.py
index 2e8eb3f4..2e8eb3f4 100644..100755
--- a/functest/utils/constants.py
+++ b/functest/utils/constants.py
diff --git a/functest/utils/functest_constants.py b/functest/utils/functest_constants.py
index 3cc744c6..0a9a03b1 100644
--- a/functest/utils/functest_constants.py
+++ b/functest/utils/functest_constants.py
@@ -195,35 +195,6 @@ ONOS_INSTALLER_MASTER_USERNAME = get_value(
ONOS_INSTALLER_MASTER_PASSWORD = get_value(
'ONOS.environment.installer_master_password',
'ONOS_INSTALLER_MASTER_PASSWORD')
-PROMISE_REPO_DIR = get_value('general.dir.dir_repo_promise',
- 'PROMISE_REPO_DIR')
-PROMISE_TENANT_NAME = get_value('promise.tenant_name',
- 'PROMISE_TENANT_NAME')
-TENANT_DESCRIPTION = get_value('promise.tenant_description',
- 'TENANT_DESCRIPTION')
-PROMISE_USER_NAME = get_value('promise.user_name', 'PROMISE_USER_NAME')
-PROMISE_USER_PWD = get_value('promise.user_pwd', 'PROMISE_USER_PWD')
-PROMISE_IMAGE_NAME = get_value('promise.image_name',
- 'PROMISE_IMAGE_NAME')
-PROMISE_FLAVOR_NAME = get_value('promise.flavor_name',
- 'PROMISE_FLAVOR_NAME')
-PROMISE_FLAVOR_VCPUS = get_value('promise.flavor_vcpus',
- 'PROMISE_FLAVOR_VCPUS')
-PROMISE_FLAVOR_RAM = get_value('promise.flavor_ram',
- 'PROMISE_FLAVOR_RAM')
-PROMISE_FLAVOR_DISK = get_value('promise.flavor_disk',
- 'PROMISE_FLAVOR_DISK')
-PROMISE_NET_NAME = get_value('promise.network_name', 'PROMISE_NET_NAME')
-PROMISE_SUBNET_NAME = get_value('promise.subnet_name',
- 'PROMISE_SUBNET_NAME')
-PROMISE_SUBNET_CIDR = get_value('promise.subnet_cidr',
- 'PROMISE_SUBNET_CIDR')
-PROMISE_ROUTER_NAME = get_value('promise.router_name',
- 'PROMISE_ROUTER_NAME')
-DOCTOR_REPO_DIR = get_value('general.dir.dir_repo_doctor',
- 'DOCTOR_REPO_DIR')
-COPPER_REPO_DIR = get_value('general.dir.repo_copper',
- 'COPPER_REPO_DIR')
EXAMPLE_INSTANCE_NAME = get_value('example.vm_name',
'EXAMPLE_INSTANCE_NAME')
EXAMPLE_FLAVOR = get_value('example.flavor', 'EXAMPLE_FLAVOR')
diff --git a/functest/utils/openstack_tacker.py b/functest/utils/openstack_tacker.py
index f17b421e..c7ac89af 100644
--- a/functest/utils/openstack_tacker.py
+++ b/functest/utils/openstack_tacker.py
@@ -20,9 +20,9 @@ import time
logger = ft_logger.Logger("tacker_utils").getLogger()
-def get_tacker_client():
- creds_tacker = os_utils.get_credentials()
- return tackerclient.Client(**creds_tacker)
+def get_tacker_client(other_creds={}):
+ sess = os_utils.get_session(other_creds)
+ return tackerclient.Client(session=sess)
# *********************************************
@@ -74,12 +74,12 @@ def create_vnfd(tacker_client, tosca_file=None):
if tosca_file is not None:
with open(tosca_file) as tosca_fd:
vnfd_body = tosca_fd.read()
- logger.error(vnfd_body)
+ logger.info('VNFD template:\n{0}'.format(vnfd_body))
return tacker_client.create_vnfd(
body={"vnfd": {"attributes": {"vnfd": vnfd_body}}})
except Exception, e:
- logger.exception("Error [create_vnfd(tacker_client, '%s')]: %s"
- % (tosca_file, e))
+ logger.error("Error [create_vnfd(tacker_client, '%s')]: %s"
+ % (tosca_file, e))
return None
@@ -124,7 +124,8 @@ def create_vnf(tacker_client, vnf_name, vnfd_id=None, vnfd_name=None):
vnf_body['vnf']['vnfd_id'] = get_vnfd_id(tacker_client, vnfd_name)
return tacker_client.create_vnf(body=vnf_body)
except Exception, e:
- logger.error("error [create_vnf(tacker_client, '%s', '%s', '%s')]: %s"
+ logger.error("error [create_vnf(tacker_client,"
+ " '%s', '%s', '%s')]: %s"
% (vnf_name, vnfd_id, vnfd_name, e))
return None
@@ -206,7 +207,8 @@ def create_sfc(tacker_client, sfc_name,
for name in chain_vnf_names]
return tacker_client.create_sfc(body=sfc_body)
except Exception, e:
- logger.error("error [create_sfc(tacker_client, '%s', '%s', '%s')]: %s"
+ logger.error("error [create_sfc(tacker_client,"
+ " '%s', '%s', '%s')]: %s"
% (sfc_name, chain_vnf_ids, chain_vnf_names, e))
return None
@@ -263,8 +265,8 @@ def create_sfc_classifier(tacker_client, sfc_clf_name, sfc_id=None,
tacker_client, sfc_name)
return tacker_client.create_sfc_classifier(body=sfc_clf_body)
except Exception, e:
- logger.error("error [create_sfc_classifier(tacker_client, '%s', '%s',"
- " '%s', '%s')]: '%s'"
+ logger.error("error [create_sfc_classifier(tacker_client,"
+ " '%s', '%s','%s', '%s')]: '%s'"
% (sfc_clf_name, sfc_id, sfc_name, str(match), e))
return None