diff options
Diffstat (limited to 'testcases/features/sfc')
-rw-r--r-- | testcases/features/sfc/SSHUtils.py | 120 | ||||
-rwxr-xr-x | testcases/features/sfc/compute_presetup_CI.bash | 27 | ||||
-rwxr-xr-x | testcases/features/sfc/correct_classifier.bash | 37 | ||||
-rwxr-xr-x | testcases/features/sfc/delete.sh | 15 | ||||
-rw-r--r-- | testcases/features/sfc/ovs_utils.py | 117 | ||||
-rwxr-xr-x | testcases/features/sfc/prepare_odl_sfc.bash | 38 | ||||
-rwxr-xr-x | testcases/features/sfc/prepare_odl_sfc.py | 96 | ||||
-rwxr-xr-x | testcases/features/sfc/server_presetup_CI.bash | 13 | ||||
-rwxr-xr-x | testcases/features/sfc/sfc.py | 545 | ||||
-rwxr-xr-x | testcases/features/sfc/sfc_change_classi.bash | 7 | ||||
-rwxr-xr-x | testcases/features/sfc/sfc_colorado1.py | 596 | ||||
-rwxr-xr-x | testcases/features/sfc/sfc_tacker.bash | 31 | ||||
-rwxr-xr-x | testcases/features/sfc/tacker_client_install.sh | 43 | ||||
-rw-r--r-- | testcases/features/sfc/test-vnfd1.yaml | 31 | ||||
-rw-r--r-- | testcases/features/sfc/test-vnfd2.yaml | 31 |
15 files changed, 0 insertions, 1747 deletions
diff --git a/testcases/features/sfc/SSHUtils.py b/testcases/features/sfc/SSHUtils.py deleted file mode 100644 index 9c8c2c72..00000000 --- a/testcases/features/sfc/SSHUtils.py +++ /dev/null @@ -1,120 +0,0 @@ -############################################################################## -# Copyright (c) 2015 Ericsson AB and others. -# Authors: George Paraskevopoulos (geopar@intracom-telecom.com) -# Jose Lausuch (jose.lausuch@ericsson.com) -# All rights reserved. This program and the accompanying materials -# are made available under the terms of the Apache License, Version 2.0 -# which accompanies this distribution, and is available at -# http://www.apache.org/licenses/LICENSE-2.0 -############################################################################## - - -import paramiko -import functest.utils.functest_logger as rl -import os - -logger = rl.Logger('SSHUtils').getLogger() - - -def get_ssh_client(hostname, username, password=None, proxy=None): - client = None - try: - if proxy is None: - client = paramiko.SSHClient() - else: - client = ProxyHopClient() - client.configure_jump_host(proxy['ip'], - proxy['username'], - proxy['password']) - - if client is None: - raise Exception('Could not connect to client') - - client.set_missing_host_key_policy(paramiko.AutoAddPolicy()) - client.connect(hostname, - username=username, - password=password) - return client - except Exception, e: - logger.error(e) - return None - - -def get_file(ssh_conn, src, dest): - try: - sftp = ssh_conn.open_sftp() - sftp.get(src, dest) - return True - except Exception, e: - logger.error("Error [get_file(ssh_conn, '%s', '%s']: %s" % - (src, dest, e)) - return None - - -def put_file(ssh_conn, src, dest): - try: - sftp = ssh_conn.open_sftp() - sftp.put(src, dest) - return True - except Exception, e: - logger.error("Error [put_file(ssh_conn, '%s', '%s']: %s" % - (src, dest, e)) - return None - - -class ProxyHopClient(paramiko.SSHClient): - ''' - Connect to a remote server using a proxy hop - ''' - def __init__(self, *args, **kwargs): - self.logger = rl.Logger("ProxyHopClient").getLogger() - self.proxy_ssh = None - self.proxy_transport = None - self.proxy_channel = None - self.proxy_ip = None - self.proxy_ssh_key = None - self.local_ssh_key = os.path.join(os.getcwd(), 'id_rsa') - super(ProxyHopClient, self).__init__(*args, **kwargs) - - def configure_jump_host(self, jh_ip, jh_user, jh_pass, - jh_ssh_key='/root/.ssh/id_rsa'): - self.proxy_ip = jh_ip - self.proxy_ssh_key = jh_ssh_key - self.proxy_ssh = paramiko.SSHClient() - self.proxy_ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy()) - self.proxy_ssh.connect(jh_ip, - username=jh_user, - password=jh_pass) - self.proxy_transport = self.proxy_ssh.get_transport() - - def connect(self, hostname, port=22, username='root', password=None, - pkey=None, key_filename=None, timeout=None, allow_agent=True, - look_for_keys=True, compress=False, sock=None, gss_auth=False, - gss_kex=False, gss_deleg_creds=True, gss_host=None, - banner_timeout=None): - try: - if self.proxy_ssh is None: - raise Exception('You must configure the jump ' - 'host before calling connect') - - get_file_res = get_file(self.proxy_ssh, - self.proxy_ssh_key, - self.local_ssh_key) - if get_file_res is None: - raise Exception('Could\'t fetch SSH key from jump host') - proxy_key = (paramiko.RSAKey - .from_private_key_file(self.local_ssh_key)) - - self.proxy_channel = self.proxy_transport.open_channel( - "direct-tcpip", - (hostname, 22), - (self.proxy_ip, 22)) - - self.set_missing_host_key_policy(paramiko.AutoAddPolicy()) - super(ProxyHopClient, self).connect(hostname, - username=username, - pkey=proxy_key, - sock=self.proxy_channel) - os.remove(self.local_ssh_key) - except Exception, e: - self.logger.error(e) diff --git a/testcases/features/sfc/compute_presetup_CI.bash b/testcases/features/sfc/compute_presetup_CI.bash deleted file mode 100755 index 36148aa1..00000000 --- a/testcases/features/sfc/compute_presetup_CI.bash +++ /dev/null @@ -1,27 +0,0 @@ -#!/bin/bash - -# This script must be use with vxlan-gpe + nsh. Once we have eth + nsh support -# in ODL, we will not need it anymore - -set -e -ssh_options='-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no' -BASEDIR=`dirname $0` -INSTALLER_IP=${INSTALLER_IP:-10.20.0.2} - -pushd $BASEDIR -#ip=`sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'fuel node'|grep compute|\ -#awk '{print $10}' | head -1` - -ip=$1 -echo $ip -#sshpass -p r00tme scp $ssh_options correct_classifier.bash ${INSTALLER_IP}:/root -#sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'scp correct_classifier.bash '"$ip"':/root' - -sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'ssh root@'"$ip"' ifconfig br-int up' -output=$(sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'ssh root@'"$ip"' ip route | \ -cut -d" " -f1 | grep 11.0.0.0' ; exit 0) - -if [ -z "$output" ]; then -sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'ssh root@'"$ip"' ip route add 11.0.0.0/24 \ -dev br-int' -fi diff --git a/testcases/features/sfc/correct_classifier.bash b/testcases/features/sfc/correct_classifier.bash deleted file mode 100755 index fb08af5c..00000000 --- a/testcases/features/sfc/correct_classifier.bash +++ /dev/null @@ -1,37 +0,0 @@ -#!/bin/bash - -#This scripts correct the current ODL bug which does not detect -#when SFF and classifier are in the same swtich - -nsp=`ovs-ofctl -O Openflow13 dump-flows br-int table=11 | \ -grep "NXM_NX_NSP" | head -1 | cut -d',' -f13 | cut -d':' -f2 \ -| cut -d'-' -f1` - -ip=`ovs-ofctl -O Openflow13 dump-flows br-int table=11 | \ -grep NXM_NX_NSH_C1 | head -1 | cut -d':' -f5 | cut -d'-' -f1` - -output_port=`ovs-ofctl -O Openflow13 show br-int | \ -grep vxgpe | cut -d'(' -f1` - -output_port2=`echo $output_port` - -echo "This is the nsp =$(($nsp))" -echo "This is the ip=$ip" -echo "This is the vxlan-gpe port=$output_port2" - -ovs-ofctl -O Openflow13 del-flows br-int "table=11,tcp,reg0=0x1,tp_dst=80" -ovs-ofctl -O Openflow13 del-flows br-int "table=11,tcp,reg0=0x1,tp_dst=22" - -ovs-ofctl -O Openflow13 add-flow br-int "table=11,tcp,reg0=0x1,tp_dst=80 \ -actions=move:NXM_NX_TUN_ID[0..31]->NXM_NX_NSH_C2[],push_nsh,\ -load:0x1->NXM_NX_NSH_MDTYPE[],load:0x3->NXM_NX_NSH_NP[],\ -load:$ip->NXM_NX_NSH_C1[],load:$nsp->NXM_NX_NSP[0..23],\ -load:0xff->NXM_NX_NSI[],load:$ip->NXM_NX_TUN_IPV4_DST[],\ -load:$nsp->NXM_NX_TUN_ID[0..31],resubmit($output_port,0)" - -ovs-ofctl -O Openflow13 add-flow br-int "table=11,tcp,reg0=0x1,tp_dst=22\ - actions=move:NXM_NX_TUN_ID[0..31]->NXM_NX_NSH_C2[],push_nsh,\ -load:0x1->NXM_NX_NSH_MDTYPE[],load:0x3->NXM_NX_NSH_NP[],\ -load:$ip->NXM_NX_NSH_C1[],load:$nsp->NXM_NX_NSP[0..23],\ -load:0xff->NXM_NX_NSI[],load:$ip->NXM_NX_TUN_IPV4_DST[],\ -load:$nsp->NXM_NX_TUN_ID[0..31],resubmit($output_port,0)" diff --git a/testcases/features/sfc/delete.sh b/testcases/features/sfc/delete.sh deleted file mode 100755 index c04ae637..00000000 --- a/testcases/features/sfc/delete.sh +++ /dev/null @@ -1,15 +0,0 @@ -tacker sfc-classifier-delete red_http -tacker sfc-classifier-delete blue_ssh -tacker sfc-classifier-delete red_ssh -tacker sfc-classifier-delete blue_http -tacker sfc-delete red -tacker sfc-delete blue -tacker vnf-delete testVNF1 -tacker vnf-delete testVNF2 -tacker vnfd-delete test-vnfd1 -tacker vnfd-delete test-vnfd2 -openstack stack delete sfc --y -openstack stack delete sfc_test1 --y -openstack stack delete sfc_test2 --y -nova delete client -nova delete server diff --git a/testcases/features/sfc/ovs_utils.py b/testcases/features/sfc/ovs_utils.py deleted file mode 100644 index 20ab5a7e..00000000 --- a/testcases/features/sfc/ovs_utils.py +++ /dev/null @@ -1,117 +0,0 @@ -############################################################################## -# Copyright (c) 2015 Ericsson AB and others. -# Author: George Paraskevopoulos (geopar@intracom-telecom.com) -# All rights reserved. This program and the accompanying materials -# are made available under the terms of the Apache License, Version 2.0 -# which accompanies this distribution, and is available at -# http://www.apache.org/licenses/LICENSE-2.0 -############################################################################## - -import functest.utils.functest_logger as rl -import os -import time -import shutil - -logger = rl.Logger('ovs_utils').getLogger() - - -class OVSLogger(object): - def __init__(self, basedir, ft_resdir): - self.ovs_dir = basedir - self.ft_resdir = ft_resdir - self.__mkdir_p(self.ovs_dir) - - def __mkdir_p(self, dirpath): - if not os.path.exists(dirpath): - os.makedirs(dirpath) - - def __ssh_host(self, ssh_conn, host_prefix='10.20.0'): - try: - _, stdout, _ = ssh_conn.exec_command('hostname -I') - hosts = stdout.readline().strip().split(' ') - found_host = [h for h in hosts if h.startswith(host_prefix)][0] - return found_host - except Exception, e: - logger.error(e) - - def __dump_to_file(self, operation, host, text, timestamp=None): - ts = (timestamp if timestamp is not None - else time.strftime("%Y%m%d-%H%M%S")) - dumpdir = os.path.join(self.ovs_dir, ts) - self.__mkdir_p(dumpdir) - fname = '{0}_{1}'.format(operation, host) - with open(os.path.join(dumpdir, fname), 'w') as f: - f.write(text) - - def __remote_cmd(self, ssh_conn, cmd): - try: - _, stdout, stderr = ssh_conn.exec_command(cmd) - errors = stderr.readlines() - if len(errors) > 0: - host = self.__ssh_host(ssh_conn) - logger.error(''.join(errors)) - raise Exception('Could not execute {0} in {1}' - .format(cmd, host)) - output = ''.join(stdout.readlines()) - return output - except Exception, e: - logger.error('[__remote_command(ssh_client, {0})]: {1}' - .format(cmd, e)) - return None - - def create_artifact_archive(self): - shutil.make_archive(self.ovs_dir, - 'zip', - root_dir=os.path.dirname(self.ovs_dir), - base_dir=self.ovs_dir) - shutil.copy2('{0}.zip'.format(self.ovs_dir), self.ft_resdir) - - def ofctl_dump_flows(self, ssh_conn, br='br-int', - choose_table=None, timestamp=None): - try: - cmd = 'ovs-ofctl -OOpenFlow13 dump-flows {0}'.format(br) - if choose_table is not None: - cmd = '{0} table={1}'.format(cmd, choose_table) - output = self.__remote_cmd(ssh_conn, cmd) - operation = 'ofctl_dump_flows' - host = self.__ssh_host(ssh_conn) - self.__dump_to_file(operation, host, output, timestamp=timestamp) - return output - except Exception, e: - logger.error('[ofctl_dump_flows(ssh_client, {0}, {1})]: {2}' - .format(br, choose_table, e)) - return None - - def vsctl_show(self, ssh_conn, timestamp=None): - try: - cmd = 'ovs-vsctl show' - output = self.__remote_cmd(ssh_conn, cmd) - operation = 'vsctl_show' - host = self.__ssh_host(ssh_conn) - self.__dump_to_file(operation, host, output, timestamp=timestamp) - return output - except Exception, e: - logger.error('[vsctl_show(ssh_client)]: {0}'.format(e)) - return None - - def dump_ovs_logs(self, controller_clients, compute_clients, - related_error=None, timestamp=None): - if timestamp is None: - timestamp = time.strftime("%Y%m%d-%H%M%S") - - for controller_client in controller_clients: - self.ofctl_dump_flows(controller_client, - timestamp=timestamp) - self.vsctl_show(controller_client, - timestamp=timestamp) - - for compute_client in compute_clients: - self.ofctl_dump_flows(compute_client, - timestamp=timestamp) - self.vsctl_show(compute_client, - timestamp=timestamp) - - if related_error is not None: - dumpdir = os.path.join(self.ovs_dir, timestamp) - with open(os.path.join(dumpdir, 'error'), 'w') as f: - f.write(related_error) diff --git a/testcases/features/sfc/prepare_odl_sfc.bash b/testcases/features/sfc/prepare_odl_sfc.bash deleted file mode 100755 index 80ed9bd9..00000000 --- a/testcases/features/sfc/prepare_odl_sfc.bash +++ /dev/null @@ -1,38 +0,0 @@ -#!/bin/bash - -# -# Author: George Paraskevopoulos (geopar@intracom-telecom.com) -# Manuel Buil (manuel.buil@ericsson.com) -# Prepares the controller and the compute nodes for the odl-sfc testcase -# -# -# All rights reserved. This program and the accompanying materials -# are made available under the terms of the Apache License, Version 2.0 -# which accompanies this distribution, and is available at -# http://www.apache.org/licenses/LICENSE-2.0 -# - -ODL_SFC_LOG=/home/opnfv/functest/results/odl-sfc.log -ODL_SFC_DIR=${FUNCTEST_REPO_DIR}/testcases/features/sfc - -# Split the output to the log file and redirect STDOUT and STDERR to /dev/null -bash ${ODL_SFC_DIR}/server_presetup_CI.bash |& \ - tee -a ${ODL_SFC_LOG} 1>/dev/null 2>&1 - -# Get return value from PIPESTATUS array (bash specific feature) -ret_val=${PIPESTATUS[0]} -if [ $ret_val != 0 ]; then - echo "The tacker server deployment failed" - exit $ret_val -fi -echo "The tacker server was deployed successfully" - -bash ${ODL_SFC_DIR}/compute_presetup_CI.bash |& \ - tee -a ${ODL_SFC_LOG} 1>/dev/null 2>&1 - -ret_val=${PIPESTATUS[0]} -if [ $ret_val != 0 ]; then - exit $ret_val -fi - -exit 0 diff --git a/testcases/features/sfc/prepare_odl_sfc.py b/testcases/features/sfc/prepare_odl_sfc.py deleted file mode 100755 index 78f4d764..00000000 --- a/testcases/features/sfc/prepare_odl_sfc.py +++ /dev/null @@ -1,96 +0,0 @@ -# -# Author: George Paraskevopoulos (geopar@intracom-telecom.com) -# Manuel Buil (manuel.buil@ericsson.com) -# Prepares the controller and the compute nodes for the odl-sfc testcase -# -# -# All rights reserved. This program and the accompanying materials -# are made available under the terms of the Apache License, Version 2.0 -# which accompanies this distribution, and is available at -# http://www.apache.org/licenses/LICENSE-2.0 -# - -import os -import sys -import subprocess -import paramiko -import functest.utils.functest_logger as ft_logger - -logger = ft_logger.Logger("ODL_SFC").getLogger() - -try: - FUNCTEST_REPO_DIR = os.environ['FUNCTEST_REPO_DIR'] -except: - logger.debug("FUNCTEST_REPO_DIR does not exist!!!!!") - -FUNCTEST_REPO_DIR = "/home/opnfv/repos/functest" - -try: - INSTALLER_IP = os.environ['INSTALLER_IP'] - -except: - logger.debug("INSTALLER_IP does not exist. We create 10.20.0.2") - INSTALLER_IP = "10.20.0.2" - -os.environ['ODL_SFC_LOG'] = "/home/opnfv/functest/results/odl-sfc.log" -os.environ['ODL_SFC_DIR'] = FUNCTEST_REPO_DIR + "/testcases/features/sfc" - -command = os.environ['ODL_SFC_DIR'] + ("/server_presetup_CI.bash | " - "tee -a ${ODL_SFC_LOG} " - "1>/dev/null 2>&1") - -output = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE) - -# This code is for debugging purposes -# for line in iter(output.stdout.readline, ''): -# i = line.rstrip() -# print(i) - -# Make sure the process is finished before checking the returncode -if not output.poll(): - output.wait() - -# Get return value -if output.returncode: - print("The presetup of the server did not work") - sys.exit(output.returncode) - -logger.info("The presetup of the server worked ") - -ssh_options = "-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no" -ssh = paramiko.SSHClient() -ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy()) - -try: - ssh.connect(INSTALLER_IP, username="root", - password="r00tme", timeout=2) - command = "fuel node | grep compute | awk '{print $10}'" - logger.info("Executing ssh to collect the compute IPs") - (stdin, stdout, stderr) = ssh.exec_command(command) -except: - logger.debug("Something went wrong in the ssh to collect the computes IP") - -output = stdout.readlines() -for ip in output: - command = os.environ['ODL_SFC_DIR'] + ("/compute_presetup_CI.bash " - "" + ip.rstrip() + "| tee -a " - "${ODL_SFC_LOG} 1>/dev/null 2>&1") - - output = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE) - -# This code is for debugging purposes -# for line in iter(output.stdout.readline, ''): -# print(line) -# sys.stdout.flush() - - output.stdout.close() - - if not (output.poll()): - output.wait() - - # Get return value - if output.returncode: - print("The compute config did not work on compute %s" % ip) - sys.exit(output.returncode) - -sys.exit(0) diff --git a/testcases/features/sfc/server_presetup_CI.bash b/testcases/features/sfc/server_presetup_CI.bash deleted file mode 100755 index 240353f5..00000000 --- a/testcases/features/sfc/server_presetup_CI.bash +++ /dev/null @@ -1,13 +0,0 @@ -#!/bin/bash -set -e -ssh_options='-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no' -BASEDIR=`dirname $0` -INSTALLER_IP=${INSTALLER_IP:-10.20.0.2} - -pushd $BASEDIR -ip=$(sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'fuel node'|grep controller|awk '{print $10}' | head -1) -echo $ip - -sshpass -p r00tme scp $ssh_options delete.sh ${INSTALLER_IP}:/root -sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'scp '"$ip"':/root/tackerc .' -sshpass -p r00tme scp $ssh_options ${INSTALLER_IP}:/root/tackerc $BASEDIR diff --git a/testcases/features/sfc/sfc.py b/testcases/features/sfc/sfc.py deleted file mode 100755 index 2cc7ac9c..00000000 --- a/testcases/features/sfc/sfc.py +++ /dev/null @@ -1,545 +0,0 @@ -import argparse -import os -import subprocess -import sys -import time -import functest.utils.functest_logger as ft_logger -import functest.utils.functest_utils as ft_utils -import functest.utils.openstack_utils as os_utils -import re -import json -import SSHUtils as ssh_utils -import ovs_utils - -parser = argparse.ArgumentParser() - -parser.add_argument("-r", "--report", - help="Create json result file", - action="store_true") - -args = parser.parse_args() - -""" logging configuration """ -logger = ft_logger.Logger("ODL_SFC").getLogger() - -FUNCTEST_RESULTS_DIR = '/home/opnfv/functest/results/odl-sfc' -FUNCTEST_REPO = ft_utils.FUNCTEST_REPO -REPO_PATH = os.environ['repos_dir'] + '/functest/' -HOME = os.environ['HOME'] + "/" -CLIENT = "client" -SERVER = "server" -FLAVOR = "custom" -IMAGE_NAME = "sf_nsh_colorado" -IMAGE_FILENAME = "sf_nsh_colorado.qcow2" -IMAGE_FORMAT = "qcow2" -IMAGE_DIR = "/home/opnfv/functest/data" -IMAGE_PATH = IMAGE_DIR + "/" + IMAGE_FILENAME -IMAGE_URL = "http://artifacts.opnfv.org/sfc/demo/" + IMAGE_FILENAME - -# NEUTRON Private Network parameters -NET_NAME = "example-net" -SUBNET_NAME = "example-subnet" -SUBNET_CIDR = "11.0.0.0/24" -ROUTER_NAME = "example-router" -SECGROUP_NAME = "example-sg" -SECGROUP_DESCR = "Example Security group" -SFC_TEST_DIR = REPO_PATH + "/testcases/features/sfc/" -TACKER_SCRIPT = SFC_TEST_DIR + "sfc_tacker.bash" -TACKER_CHANGECLASSI = SFC_TEST_DIR + "sfc_change_classi.bash" -ssh_options = '-q -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no' -json_results = {"tests": 4, "failures": 0} - -PROXY = { - 'ip': '10.20.0.2', - 'username': 'root', - 'password': 'r00tme' -} - -# run given command locally and return commands output if success - - -def run_cmd(cmd, wdir=None, ignore_stderr=False, ignore_no_output=True): - pipe = subprocess.Popen(cmd, shell=True, - stdin=subprocess.PIPE, - stdout=subprocess.PIPE, - stderr=subprocess.PIPE, cwd=wdir) - - (output, errors) = pipe.communicate() - if output: - output = output.strip() - if pipe.returncode < 0: - logger.error(errors) - return False - if errors: - logger.error(errors) - if ignore_stderr: - return True - else: - return False - - if ignore_no_output: - if not output: - return True - - return output - -# run given command on OpenStack controller - - -def run_cmd_on_cntlr(cmd): - ip_cntlrs = get_openstack_node_ips("controller") - if not ip_cntlrs: - return None - - ssh_cmd = "ssh %s %s %s" % (ssh_options, ip_cntlrs[0], cmd) - return run_cmd_on_fm(ssh_cmd) - -# run given command on OpenStack Compute node - - -def run_cmd_on_compute(cmd): - ip_computes = get_openstack_node_ips("compute") - if not ip_computes: - return None - - ssh_cmd = "ssh %s %s %s" % (ssh_options, ip_computes[0], cmd) - return run_cmd_on_fm(ssh_cmd) - -# run given command on Fuel Master - - -def run_cmd_on_fm(cmd, username="root", passwd="r00tme"): - ip = os.environ.get("INSTALLER_IP") - ssh_cmd = "sshpass -p %s ssh %s %s@%s %s" % ( - passwd, ssh_options, username, ip, cmd) - return run_cmd(ssh_cmd) - -# run given command on Remote Machine, Can be VM - - -def run_cmd_remote(ip, cmd, username="root", passwd="opnfv"): - ssh_opt_append = "%s -o ConnectTimeout=50 " % ssh_options - ssh_cmd = "sshpass -p %s ssh %s %s@%s %s" % ( - passwd, ssh_opt_append, username, ip, cmd) - return run_cmd(ssh_cmd) - -# Get OpenStack Nodes IP Address - - -def get_openstack_node_ips(role): - fuel_env = os.environ.get("FUEL_ENV") - if fuel_env is not None: - cmd = "fuel2 node list -f json -e %s" % fuel_env - else: - cmd = "fuel2 node list -f json" - - nodes = run_cmd_on_fm(cmd) - ips = [] - nodes = json.loads(nodes) - for node in nodes: - if role in node["roles"]: - ips.append(node["ip"]) - - return ips - -# Configures IPTABLES on OpenStack Controller - - -def configure_iptables(): - iptable_cmds = ["iptables -P INPUT ACCEPT", - "iptables -t nat -P INPUT ACCEPT", - "iptables -A INPUT -m state \ - --state NEW,ESTABLISHED,RELATED -j ACCEPT"] - - for cmd in iptable_cmds: - logger.info("Configuring %s on contoller" % cmd) - run_cmd_on_cntlr(cmd) - - return - - -def download_image(): - if not os.path.isfile(IMAGE_PATH): - logger.info("Downloading image") - ft_utils.download_url(IMAGE_URL, IMAGE_DIR) - - logger.info("Using old image") - return - - -def setup_glance(glance_client): - image_id = os_utils.create_glance_image(glance_client, - IMAGE_NAME, - IMAGE_PATH, - disk=IMAGE_FORMAT, - container="bare", - public=True) - - return image_id - - -def setup_neutron(neutron_client): - n_dict = os_utils.create_network_full(neutron_client, - NET_NAME, - SUBNET_NAME, - ROUTER_NAME, - SUBNET_CIDR) - if not n_dict: - logger.error("failed to create neutron network") - sys.exit(-1) - - network_id = n_dict["net_id"] - return network_id - - -def setup_ingress_egress_secgroup(neutron_client, protocol, - min_port=None, max_port=None): - secgroups = os_utils.get_security_groups(neutron_client) - for sg in secgroups: - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'ingress', protocol, - port_range_min=min_port, - port_range_max=max_port) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'egress', protocol, - port_range_min=min_port, - port_range_max=max_port) - return - - -def setup_security_groups(neutron_client): - sg_id = os_utils.create_security_group_full(neutron_client, - SECGROUP_NAME, SECGROUP_DESCR) - setup_ingress_egress_secgroup(neutron_client, "icmp") - setup_ingress_egress_secgroup(neutron_client, "udp", 67, 68) - setup_ingress_egress_secgroup(neutron_client, "tcp", 22, 22) - setup_ingress_egress_secgroup(neutron_client, "tcp", 80, 80) - return sg_id - - -def boot_instance(nova_client, name, flavor, image_id, network_id, sg_id): - logger.info("Creating instance '%s'..." % name) - logger.debug( - "Configuration:\n name=%s \n flavor=%s \n image=%s \n " - "network=%s \n" % (name, flavor, image_id, network_id)) - - instance = os_utils.create_instance_and_wait_for_active(flavor, - image_id, - network_id, - name) - - if instance is None: - logger.error("Error while booting instance.") - sys.exit(-1) - - instance_ip = instance.networks.get(NET_NAME)[0] - logger.debug("Instance '%s' got private ip '%s'." % - (name, instance_ip)) - - logger.info("Adding '%s' to security group %s" % (name, SECGROUP_NAME)) - os_utils.add_secgroup_to_instance(nova_client, instance.id, sg_id) - - return instance_ip - - -def ping(remote, pkt_cnt=1, iface=None, retries=100, timeout=None): - ping_cmd = 'ping' - - if timeout: - ping_cmd = ping_cmd + ' -w %s' % timeout - - grep_cmd = "grep -e 'packet loss' -e rtt" - - if iface is not None: - ping_cmd = ping_cmd + ' -I %s' % iface - - ping_cmd = ping_cmd + ' -i 0 -c %d %s' % (pkt_cnt, remote) - cmd = ping_cmd + '|' + grep_cmd - - while retries > 0: - output = run_cmd(cmd) - if not output: - return False - - match = re.search('(\d*)% packet loss', output) - if not match: - return False - - packet_loss = int(match.group(1)) - if packet_loss == 0: - return True - - retries = retries - 1 - - return False - - -def get_floating_ips(nova_client, neutron_client): - ips = [] - instances = nova_client.servers.list(search_opts={'all_tenants': 1}) - for instance in instances: - floatip_dic = os_utils.create_floating_ip(neutron_client) - floatip = floatip_dic['fip_addr'] - instance.add_floating_ip(floatip) - logger.info("Instance name and ip %s:%s " % (instance.name, floatip)) - logger.info("Waiting for instance %s:%s to come up" % - (instance.name, floatip)) - if not ping(floatip): - logger.info("Instance %s:%s didn't come up" % - (instance.name, floatip)) - sys.exit(1) - - if instance.name == "server": - logger.info("Server:%s is reachable" % floatip) - server_ip = floatip - elif instance.name == "client": - logger.info("Client:%s is reachable" % floatip) - client_ip = floatip - else: - logger.info("SF:%s is reachable" % floatip) - ips.append(floatip) - - return server_ip, client_ip, ips[1], ips[0] - -# Start http server on a give machine, Can be VM - - -def start_http_server(ip): - cmd = "\'python -m SimpleHTTPServer 80" - cmd = cmd + " > /dev/null 2>&1 &\'" - return run_cmd_remote(ip, cmd) - -# Set firewall using vxlan_tool.py on a give machine, Can be VM - - -def vxlan_firewall(sf, iface="eth0", port="22", block=True): - cmd = "python vxlan_tool.py" - cmd = cmd + " -i " + iface + " -d forward -v off" - if block: - cmd = "python vxlan_tool.py -i eth0 -d forward -v off -b " + port - - cmd = "sh -c 'cd /root;nohup " + cmd + " > /dev/null 2>&1 &'" - run_cmd_remote(sf, cmd) - -# Run netcat on a give machine, Can be VM - - -def netcat(s_ip, c_ip, port="80", timeout=5): - cmd = "nc -zv " - cmd = cmd + " -w %s %s %s" % (timeout, s_ip, port) - cmd = cmd + " 2>&1" - output = run_cmd_remote(c_ip, cmd) - logger.info("%s" % output) - return output - - -def is_ssh_blocked(srv_prv_ip, client_ip): - res = netcat(srv_prv_ip, client_ip, port="22") - match = re.search("nc:.*timed out:.*", res, re.M) - if match: - return True - - return False - - -def is_http_blocked(srv_prv_ip, client_ip): - res = netcat(srv_prv_ip, client_ip, port="80") - match = re.search(".* 80 port.* succeeded!", res, re.M) - if match: - return False - - return True - - -def capture_err_logs(controller_clients, compute_clients, error): - ovs_logger = ovs_utils.OVSLogger( - os.path.join(os.getcwd(), 'ovs-logs'), - FUNCTEST_RESULTS_DIR) - - timestamp = time.strftime("%Y%m%d-%H%M%S") - ovs_logger.dump_ovs_logs(controller_clients, - compute_clients, - related_error=error, - timestamp=timestamp) - return - - -def update_json_results(name, result): - json_results.update({name: result}) - if result is not "Passed": - json_results["failures"] += 1 - - return - - -def get_ssh_clients(role): - clients = [] - for ip in get_openstack_node_ips(role): - s_client = ssh_utils.get_ssh_client(ip, - 'root', - proxy=PROXY) - clients.append(s_client) - - return clients - -# Check SSH connectivity to VNFs - - -def check_ssh(ips, retries=100): - check = [False, False] - logger.info("Checking SSH connectivity to the SFs with ips %s" % str(ips)) - while retries and not all(check): - for index, ip in enumerate(ips): - check[index] = run_cmd_remote(ip, "exit") - - if all(check): - logger.info("SSH connectivity to the SFs established") - return True - - time.sleep(3) - retries -= 1 - - return False - - -def main(): - installer_type = os.environ.get("INSTALLER_TYPE") - if installer_type != "fuel": - logger.error( - '\033[91mCurrently supported only Fuel Installer type\033[0m') - sys.exit(1) - - installer_ip = os.environ.get("INSTALLER_IP") - if not installer_ip: - logger.error( - '\033[91minstaller ip is not set\033[0m') - logger.error( - '\033[91mexport INSTALLER_IP=<ip>\033[0m') - sys.exit(1) - - env_list = run_cmd_on_fm("fuel2 env list -f json") - fuel_env = os.environ.get("FUEL_ENV") - if len(eval(env_list)) > 1 and fuel_env is None: - out = run_cmd_on_fm("fuel env") - logger.error( - '\033[91mMore than one fuel env found\033[0m\n %s' % out) - logger.error( - '\033[91mexport FUEL_ENV=<env-id> to set ENV\033[0m') - sys.exit(1) - - start_time = time.time() - status = "PASS" - configure_iptables() - download_image() - _, custom_flv_id = os_utils.get_or_create_flavor( - FLAVOR, 1500, 10, 1, public=True) - if not custom_flv_id: - logger.error("Failed to create custom flavor") - sys.exit(1) - - glance_client = os_utils.get_glance_client() - neutron_client = os_utils.get_neutron_client() - nova_client = os_utils.get_nova_client() - - controller_clients = get_ssh_clients("controller") - compute_clients = get_ssh_clients("compute") - - image_id = setup_glance(glance_client) - network_id = setup_neutron(neutron_client) - sg_id = setup_security_groups(neutron_client) - - boot_instance( - nova_client, CLIENT, FLAVOR, image_id, network_id, sg_id) - srv_prv_ip = boot_instance( - nova_client, SERVER, FLAVOR, image_id, network_id, sg_id) - - subprocess.call(TACKER_SCRIPT, shell=True) - server_ip, client_ip, sf1, sf2 = get_floating_ips( - nova_client, neutron_client) - - if not check_ssh([sf1, sf2]): - logger.error("Cannot establish SSH connection to the SFs") - sys.exit(1) - - logger.info("Starting HTTP server on %s" % server_ip) - if not start_http_server(server_ip): - logger.error( - '\033[91mFailed to start HTTP server on %s\033[0m' % server_ip) - sys.exit(1) - - logger.info("Starting HTTP firewall on %s" % sf2) - vxlan_firewall(sf2, port="80") - logger.info("Starting SSH firewall on %s" % sf1) - vxlan_firewall(sf1, port="22") - - logger.info("Wait for ODL to update the classification rules in OVS") - time.sleep(120) - - logger.info("Test SSH") - if is_ssh_blocked(srv_prv_ip, client_ip): - logger.info('\033[92mTEST 1 [PASSED] ==> SSH BLOCKED\033[0m') - update_json_results("Test 1: SSH Blocked", "Passed") - else: - error = ('\033[91mTEST 1 [FAILED] ==> SSH NOT BLOCKED\033[0m') - logger.error(error) - capture_err_logs(controller_clients, compute_clients, error) - update_json_results("Test 1: SSH Blocked", "Failed") - - logger.info("Test HTTP") - if not is_http_blocked(srv_prv_ip, client_ip): - logger.info('\033[92mTEST 2 [PASSED] ==> HTTP WORKS\033[0m') - update_json_results("Test 2: HTTP works", "Passed") - else: - error = ('\033[91mTEST 2 [FAILED] ==> HTTP BLOCKED\033[0m') - logger.error(error) - capture_err_logs(controller_clients, compute_clients, error) - update_json_results("Test 2: HTTP works", "Failed") - - logger.info("Changing the classification") - subprocess.call(TACKER_CHANGECLASSI, shell=True) - logger.info("Wait for ODL to update the classification rules in OVS") - time.sleep(100) - - logger.info("Test HTTP") - if is_http_blocked(srv_prv_ip, client_ip): - logger.info('\033[92mTEST 3 [PASSED] ==> HTTP Blocked\033[0m') - update_json_results("Test 3: HTTP Blocked", "Passed") - else: - error = ('\033[91mTEST 3 [FAILED] ==> HTTP WORKS\033[0m') - logger.error(error) - capture_err_logs(controller_clients, compute_clients, error) - update_json_results("Test 3: HTTP Blocked", "Failed") - - logger.info("Test SSH") - if not is_ssh_blocked(srv_prv_ip, client_ip): - logger.info('\033[92mTEST 4 [PASSED] ==> SSH Works\033[0m') - update_json_results("Test 4: SSH Works", "Passed") - else: - error = ('\033[91mTEST 4 [FAILED] ==> SSH BLOCKED\033[0m') - logger.error(error) - capture_err_logs(controller_clients, compute_clients, error) - update_json_results("Test 4: SSH Works", "Failed") - - if json_results["failures"]: - status = "FAIL" - logger.error('\033[91mSFC TESTS: %s :( FOUND %s FAIL \033[0m' % ( - status, json_results["failures"])) - - if args.report: - stop_time = time.time() - logger.debug("Promise Results json: " + str(json_results)) - ft_utils.push_results_to_db("sfc", - "functest-odl-sfc", - start_time, - stop_time, - status, - json_results) - - if status == "PASS": - logger.info('\033[92mSFC ALL TESTS: %s :)\033[0m' % status) - sys.exit(0) - - sys.exit(1) - -if __name__ == '__main__': - main() diff --git a/testcases/features/sfc/sfc_change_classi.bash b/testcases/features/sfc/sfc_change_classi.bash deleted file mode 100755 index 70375ab3..00000000 --- a/testcases/features/sfc/sfc_change_classi.bash +++ /dev/null @@ -1,7 +0,0 @@ -tacker sfc-classifier-delete red_http -tacker sfc-classifier-delete red_ssh - -tacker sfc-classifier-create --name blue_http --chain blue --match source_port=0,dest_port=80,protocol=6 -tacker sfc-classifier-create --name blue_ssh --chain blue --match source_port=0,dest_port=22,protocol=6 - -tacker sfc-classifier-list diff --git a/testcases/features/sfc/sfc_colorado1.py b/testcases/features/sfc/sfc_colorado1.py deleted file mode 100755 index d31541d1..00000000 --- a/testcases/features/sfc/sfc_colorado1.py +++ /dev/null @@ -1,596 +0,0 @@ -import os -import subprocess -import sys -import time -import argparse -import paramiko - -import functest.utils.functest_logger as ft_logger -import functest.utils.functest_utils as ft_utils -import functest.utils.openstack_utils as os_utils -import SSHUtils as ssh_utils -import ovs_utils - -parser = argparse.ArgumentParser() - -parser.add_argument("-r", "--report", - help="Create json result file", - action="store_true") - -args = parser.parse_args() - -""" logging configuration """ -logger = ft_logger.Logger("ODL_SFC").getLogger() - -FUNCTEST_RESULTS_DIR = '/home/opnfv/functest/results/odl-sfc' -FUNCTEST_REPO = ft_utils.FUNCTEST_REPO - -HOME = os.environ['HOME'] + "/" - -VM_BOOT_TIMEOUT = 180 -INSTANCE_NAME = "client" -FLAVOR = "custom" -IMAGE_NAME = "sf_nsh_colorado" -IMAGE_FILENAME = "sf_nsh_colorado.qcow2" -IMAGE_FORMAT = "qcow2" -IMAGE_PATH = "/home/opnfv/functest/data" + "/" + IMAGE_FILENAME - -# NEUTRON Private Network parameters - -NET_NAME = "example-net" -SUBNET_NAME = "example-subnet" -SUBNET_CIDR = "11.0.0.0/24" -ROUTER_NAME = "example-router" - -SECGROUP_NAME = "example-sg" -SECGROUP_DESCR = "Example Security group" - -INSTANCE_NAME_2 = "server" - -# TEST_DB = ft_utils.get_parameter_from_yaml("results.test_db_url") - -PRE_SETUP_SCRIPT = 'sfc_pre_setup.bash' -TACKER_SCRIPT = 'sfc_tacker.bash' -TEARDOWN_SCRIPT = "sfc_teardown.bash" -TACKER_CHANGECLASSI = "sfc_change_classi.bash" - -ssh_options = '-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no' - -PROXY = { - 'ip': '10.20.0.2', - 'username': 'root', - 'password': 'r00tme' -} - - -def check_ssh(ip): - cmd = "sshpass -p opnfv ssh " + ssh_options + " -q " + ip + " exit" - success = subprocess.call(cmd, shell=True) == 0 - if not success: - logger.debug("Wating for SSH connectivity in SF with IP: %s" % ip) - return success - - -def main(): - - # Allow any port so that tacker commands reaches the server. - # This will be deleted when tacker is included in OPNFV installation - - status = "PASS" - failures = 0 - start_time = time.time() - json_results = {} - - contr_cmd = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " 'fuel node'|grep controller|awk '{print $10}'") - logger.info("Executing script to get ip_server: '%s'" % contr_cmd) - process = subprocess.Popen(contr_cmd, - shell=True, - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) - ip_server = process.stdout.readline().rstrip() - - comp_cmd = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " 'fuel node'|grep compute|awk '{print $10}'") - logger.info("Executing script to get compute IPs: '%s'" % comp_cmd) - process = subprocess.Popen(comp_cmd, - shell=True, - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) - ip_computes = [ip.strip() for ip in process.stdout.readlines()] - - iptable_cmd1 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " ssh " + ip_server + " iptables -P INPUT ACCEPT ") - iptable_cmd2 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " ssh " + ip_server + " iptables -t nat -P INPUT ACCEPT ") - iptable_cmd3 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " ssh " + ssh_options + " " + ip_server + - " iptables -A INPUT -m state" - " --state NEW,ESTABLISHED,RELATED -j ACCEPT") - - logger.info("Changing firewall policy in controller: '%s'" % iptable_cmd1) - subprocess.call(iptable_cmd1, shell=True, stderr=subprocess.PIPE) - - logger.info("Changing firewall policy in controller: '%s'" % iptable_cmd2) - subprocess.call(iptable_cmd2, shell=True, stderr=subprocess.PIPE) - - logger.info("Changing firewall policy in controller: '%s'" % iptable_cmd3) - subprocess.call(iptable_cmd2, shell=True, stderr=subprocess.PIPE) - -# Getting the different clients - - nova_client = os_utils.get_nova_client() - neutron_client = os_utils.get_neutron_client() - glance_client = os_utils.get_glance_client() - - ovs_logger = ovs_utils.OVSLogger( - os.path.join(os.getcwd(), 'ovs-logs'), - FUNCTEST_RESULTS_DIR) - - controller_clients = [ssh_utils.get_ssh_client(ip_server, - 'root', - proxy=PROXY)] - compute_clients = [] - for c_ip in ip_computes: - c_client = ssh_utils.get_ssh_client(c_ip, - 'root', - proxy=PROXY) - compute_clients.append(c_client) - -# Download the image - - if not os.path.isfile(IMAGE_PATH): - logger.info("Downloading image") - ft_utils.download_url( - "http://artifacts.opnfv.org/sfc/demo/sf_nsh_colorado.qcow2", - "/home/opnfv/functest/data/") - else: - logger.info("Using old image") - -# Create glance image and the neutron network - - image_id = os_utils.create_glance_image(glance_client, - IMAGE_NAME, - IMAGE_PATH, - disk=IMAGE_FORMAT, - container="bare", - public=True) - - network_dic = os_utils.create_network_full(neutron_client, - NET_NAME, - SUBNET_NAME, - ROUTER_NAME, - SUBNET_CIDR) - if not network_dic: - logger.error( - "There has been a problem when creating the neutron network") - sys.exit(-1) - - network_id = network_dic["net_id"] - - sg_id = os_utils.create_security_group_full(neutron_client, - SECGROUP_NAME, SECGROUP_DESCR) - - secgroups = os_utils.get_security_groups(neutron_client) - - for sg in secgroups: - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'ingress', 'udp', - port_range_min=67, - port_range_max=68) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'egress', 'udp', - port_range_min=67, - port_range_max=68) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'ingress', 'tcp', - port_range_min=22, - port_range_max=22) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'egress', 'tcp', - port_range_min=22, - port_range_max=22) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'ingress', 'tcp', - port_range_min=80, - port_range_max=80) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'egress', 'tcp', - port_range_min=80, - port_range_max=80) - - _, custom_flv_id = os_utils.get_or_create_flavor( - 'custom', 1500, 10, 1, public=True) - if not custom_flv_id: - logger.error("Failed to create custom flavor") - sys.exit(1) - - iterator = 0 - while(iterator < 6): - # boot INSTANCE - logger.info("Creating instance '%s'..." % INSTANCE_NAME) - logger.debug( - "Configuration:\n name=%s \n flavor=%s \n image=%s \n " - "network=%s \n" % (INSTANCE_NAME, FLAVOR, image_id, network_id)) - instance = os_utils.create_instance_and_wait_for_active( - FLAVOR, - image_id, - network_id, - INSTANCE_NAME, - av_zone='nova') - - if instance is None: - logger.error("Error while booting instance.") - iterator += 1 - continue - # Retrieve IP of INSTANCE - instance_ip = instance.networks.get(NET_NAME)[0] - logger.debug("Instance '%s' got private ip '%s'." % - (INSTANCE_NAME, instance_ip)) - - logger.info("Adding '%s' to security group '%s'..." - % (INSTANCE_NAME, SECGROUP_NAME)) - os_utils.add_secgroup_to_instance(nova_client, instance.id, sg_id) - - logger.info("Creating floating IP for VM '%s'..." % INSTANCE_NAME) - floatip_dic = os_utils.create_floating_ip(neutron_client) - floatip_client = floatip_dic['fip_addr'] - # floatip_id = floatip_dic['fip_id'] - - if floatip_client is None: - logger.error("Cannot create floating IP.") - iterator += 1 - continue - logger.info("Floating IP created: '%s'" % floatip_client) - - logger.info("Associating floating ip: '%s' to VM '%s' " - % (floatip_client, INSTANCE_NAME)) - if not os_utils.add_floating_ip(nova_client, - instance.id, - floatip_client): - logger.error("Cannot associate floating IP to VM.") - iterator += 1 - continue - - # STARTING SECOND VM (server) ### - - # boot INTANCE - logger.info("Creating instance '%s'..." % INSTANCE_NAME_2) - logger.debug( - "Configuration:\n name=%s \n flavor=%s \n image=%s \n " - "network=%s \n" % (INSTANCE_NAME_2, FLAVOR, image_id, network_id)) - instance_2 = os_utils.create_instance_and_wait_for_active( - FLAVOR, - image_id, - network_id, - INSTANCE_NAME_2, - av_zone='nova') - - if instance_2 is None: - logger.error("Error while booting instance.") - iterator += 1 - continue - # Retrieve IP of INSTANCE - instance_ip_2 = instance_2.networks.get(NET_NAME)[0] - logger.debug("Instance '%s' got private ip '%s'." % - (INSTANCE_NAME_2, instance_ip_2)) - - logger.info("Adding '%s' to security group '%s'..." - % (INSTANCE_NAME_2, SECGROUP_NAME)) - os_utils.add_secgroup_to_instance(nova_client, instance_2.id, sg_id) - - logger.info("Creating floating IP for VM '%s'..." % INSTANCE_NAME_2) - floatip_dic = os_utils.create_floating_ip(neutron_client) - floatip_server = floatip_dic['fip_addr'] - # floatip_id = floatip_dic['fip_id'] - - if floatip_server is None: - logger.error("Cannot create floating IP.") - iterator += 1 - continue - logger.info("Floating IP created: '%s'" % floatip_server) - - logger.info("Associating floating ip: '%s' to VM '%s' " - % (floatip_server, INSTANCE_NAME_2)) - - if not os_utils.add_floating_ip(nova_client, - instance_2.id, - floatip_server): - logger.error("Cannot associate floating IP to VM.") - iterator += 1 - continue - - # CREATION OF THE 2 SF #### - - tacker_script = "%s/testcases/features/sfc/%s" % \ - (FUNCTEST_REPO, TACKER_SCRIPT) - logger.info("Executing tacker script: '%s'" % tacker_script) - subprocess.call(tacker_script, shell=True) - - # SSH CALL TO START HTTP SERVER - ssh = paramiko.SSHClient() - ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy()) - - try: - ssh.connect(floatip_server, username="root", - password="opnfv", timeout=2) - command = "python -m SimpleHTTPServer 80 > /dev/null 2>&1 &" - logger.info("Starting HTTP server") - (stdin, stdout, stderr) = ssh.exec_command(command) - except: - logger.debug("Waiting for %s..." % floatip_server) - time.sleep(6) - # timeout -= 1 - - instances = nova_client.servers.list(search_opts={'all_tenants': 1}) - ips = [] - try: - for instance in instances: - if "server" not in instance.name: - if "client" not in instance.name: - logger.debug( - "This is the instance name: %s " % instance.name) - floatip_dic = os_utils.create_floating_ip( - neutron_client) - floatip = floatip_dic['fip_addr'] - ips.append(floatip) - instance.add_floating_ip(floatip) - except: - logger.debug("Problems assigning floating IP to SFs") - - # If no IPs were obtained, then we cant continue - if not ips: - logger.error('Failed to obtain IPs, cant continue, exiting') - return - - logger.debug("Floating IPs for SFs: %s..." % ips) - - # Check SSH connectivity to VNFs - r = 0 - retries = 100 - check = [False, False] - - logger.info("Checking SSH connectivity to the SFs with ips {0}" - .format(str(ips))) - while r < retries and not all(check): - try: - check = [check_ssh(ips[0]), check_ssh(ips[1])] - except Exception: - logger.exception("SSH check failed") - check = [False, False] - time.sleep(3) - r += 1 - - if not all(check): - logger.error("Cannot establish SSH connection to the SFs") - iterator += 1 - continue - - logger.info("SSH connectivity to the SFs established") - - # SSH TO START THE VXLAN_TOOL ON SF1 - logger.info("Configuring the SFs") - try: - ssh.connect(ips[0], username="root", - password="opnfv", timeout=2) - command = ("nohup python vxlan_tool.py -i eth0 " - "-d forward -v off -b 80 > /dev/null 2>&1 &") - (stdin, stdout, stderr) = ssh.exec_command(command) - except: - logger.debug("Waiting for %s..." % ips[0]) - time.sleep(6) - # timeout -= 1 - - try: - n = 0 - while 1: - (stdin, stdout, stderr) = ssh.exec_command( - "ps aux | grep \"vxlan_tool.py\" | grep -v grep") - if len(stdout.readlines()) > 0: - logger.debug("HTTP firewall started") - break - else: - n += 1 - if (n > 7): - break - logger.debug("HTTP firewall not started") - time.sleep(3) - except Exception: - logger.exception("vxlan_tool not started in SF1") - - # SSH TO START THE VXLAN_TOOL ON SF2 - try: - ssh.connect(ips[1], username="root", - password="opnfv", timeout=2) - command = ("nohup python vxlan_tool.py -i eth0 " - "-d forward -v off -b 22 > /dev/null 2>&1 &") - (stdin, stdout, stderr) = ssh.exec_command(command) - except: - logger.debug("Waiting for %s..." % ips[1]) - time.sleep(6) - # timeout -= 1 - - try: - n = 0 - while 1: - (stdin, stdout, stderr) = ssh.exec_command( - "ps aux | grep \"vxlan_tool.py\" | grep -v grep") - if len(stdout.readlines()) > 0: - logger.debug("SSH firewall started") - break - else: - n += 1 - if (n > 7): - break - logger.debug("SSH firewall not started") - time.sleep(3) - except Exception: - logger.exception("vxlan_tool not started in SF2") - - i = 0 - - # SSH TO EXECUTE cmd_client - logger.info("TEST STARTED") - time.sleep(70) - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 22 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - # WRITE THE CORRECT WAY TO DO LOGGING - if "timed out" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 1 [PASSED] " - "==> SSH BLOCKED" + '\033[0m') - i = i + 1 - json_results.update({"Test 1: SSH Blocked": "Passed"}) - else: - timestamp = time.strftime("%Y%m%d-%H%M%S") - error = ('\033[91m' + "TEST 1 [FAILED] " - "==> SSH NOT BLOCKED" + '\033[0m') - logger.error(error) - ovs_logger.dump_ovs_logs(controller_clients, - compute_clients, - related_error=error, - timestamp=timestamp) - status = "FAIL" - json_results.update({"Test 1: SSH Blocked": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - # SSH TO EXECUTE cmd_client - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 80 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - if "succeeded" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 2 [PASSED] " - "==> HTTP WORKS" + '\033[0m') - i = i + 1 - json_results.update({"Test 2: HTTP works": "Passed"}) - else: - timestamp = time.strftime("%Y%m%d-%H%M%S") - error = ('\033[91m' + "TEST 2 [FAILED] " - "==> HTTP BLOCKED" + '\033[0m') - logger.error(error) - ovs_logger.dump_ovs_logs(controller_clients, - compute_clients, - related_error=error, - timestamp=timestamp) - status = "FAIL" - json_results.update({"Test 2: HTTP works": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - # CHANGE OF CLASSIFICATION # - logger.info("Changing the classification") - tacker_classi = "%s/testcases/features/sfc/%s" % \ - (FUNCTEST_REPO, TACKER_CHANGECLASSI) - subprocess.call(tacker_classi, shell=True) - - logger.info("Wait for ODL to update the classification rules in OVS") - time.sleep(100) - - # SSH TO EXECUTE cmd_client - - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 80 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - if "timed out" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 3 [PASSED] " - "==> HTTP BLOCKED" + '\033[0m') - i = i + 1 - json_results.update({"Test 3: HTTP Blocked": "Passed"}) - else: - timestamp = time.strftime("%Y%m%d-%H%M%S") - error = ('\033[91m' + "TEST 3 [FAILED] " - "==> HTTP NOT BLOCKED" + '\033[0m') - logger.error(error) - ovs_logger.dump_ovs_logs(controller_clients, - compute_clients, - related_error=error, - timestamp=timestamp) - status = "FAIL" - json_results.update({"Test 3: HTTP Blocked": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - # SSH TO EXECUTE cmd_client - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 22 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - if "succeeded" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 4 [PASSED] " - "==> SSH WORKS" + '\033[0m') - i = i + 1 - json_results.update({"Test 4: SSH works": "Passed"}) - else: - timestamp = time.strftime("%Y%m%d-%H%M%S") - error = ('\033[91m' + "TEST 4 [FAILED] " - "==> SSH BLOCKED" + '\033[0m') - logger.error(error) - ovs_logger.dump_ovs_logs(controller_clients, - compute_clients, - related_error=error, - timestamp=timestamp) - status = "FAIL" - json_results.update({"Test 4: SSH works": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - ovs_logger.create_artifact_archive() - - iterator += 1 - if i == 4: - for x in range(0, 5): - logger.info('\033[92m' + "SFC TEST WORKED" - " :) \n" + '\033[0m') - break - else: - logger.info("Iterating again!") - delete = "bash %s/testcases/features/sfc/delete.sh" % \ - (FUNCTEST_REPO) - try: - subprocess.call(delete, shell=True, stderr=subprocess.PIPE) - time.sleep(10) - except Exception, e: - logger.error("Problem when executing the delete.sh") - logger.error("Problem %s" % e) - - if args.report: - stop_time = time.time() - json_results.update({"tests": "4", "failures": int(failures)}) - logger.debug("Promise Results json: " + str(json_results)) - ft_utils.push_results_to_db("sfc", - "functest-odl-sfc", - start_time, - stop_time, - status, - json_results) - if status == "PASS": - sys.exit(0) - else: - sys.exit(1) - -if __name__ == '__main__': - main() diff --git a/testcases/features/sfc/sfc_tacker.bash b/testcases/features/sfc/sfc_tacker.bash deleted file mode 100755 index 690d5f52..00000000 --- a/testcases/features/sfc/sfc_tacker.bash +++ /dev/null @@ -1,31 +0,0 @@ -#!/bin/bash -BASEDIR=`dirname $0` - -#import VNF descriptor -tacker vnfd-create --vnfd-file ${BASEDIR}/test-vnfd1.yaml -tacker vnfd-create --vnfd-file ${BASEDIR}/test-vnfd2.yaml - -#create instances of the imported VNF -tacker vnf-create --name testVNF1 --vnfd-name test-vnfd1 -tacker vnf-create --name testVNF2 --vnfd-name test-vnfd2 - -key=true -while $key;do - sleep 3 - active=`tacker vnf-list | grep -E 'PENDING|ERROR'` - echo -e "checking if SFs are up: $active" - if [ -z "$active" ]; then - key=false - fi -done - -#create service chain -tacker sfc-create --name red --chain testVNF1 -tacker sfc-create --name blue --chain testVNF2 - -#create classifier -tacker sfc-classifier-create --name red_http --chain red --match source_port=0,dest_port=80,protocol=6 -tacker sfc-classifier-create --name red_ssh --chain red --match source_port=0,dest_port=22,protocol=6 - -tacker sfc-list -tacker sfc-classifier-list diff --git a/testcases/features/sfc/tacker_client_install.sh b/testcases/features/sfc/tacker_client_install.sh deleted file mode 100755 index adb9a44b..00000000 --- a/testcases/features/sfc/tacker_client_install.sh +++ /dev/null @@ -1,43 +0,0 @@ -MYDIR=$(dirname $(readlink -f "$0")) -CLIENT=$(echo python-python-tackerclient_*_all.deb) -CLIREPO="tacker-client" - -# Function checks whether a python egg is available, if not, installs -function chkPPkg() { - PKG="$1" - IPPACK=$(python - <<'____EOF' -import pip -from os.path import join -for package in pip.get_installed_distributions(): - print(package.location) - print(join(package.location, *package._get_metadata("top_level.txt"))) -____EOF -) - echo "$IPPACK" | grep -q "$PKG" - if [ $? -ne 0 ];then - pip install "$PKG" - fi -} - -function envSetup() { - apt-get install -y python-all debhelper fakeroot - #pip install --upgrade python-keystoneclient==1.7.4 - chkPPkg stdeb -} - -# Function installs python-tackerclient from github -function deployTackerClient() { - cd $MYDIR - git clone -b 'SFC_refactor' https://github.com/trozet/python-tackerclient.git $CLIREPO - cd $CLIREPO - python setup.py --command-packages=stdeb.command bdist_deb - cd "deb_dist" - CLIENT=$(echo python-python-tackerclient_*_all.deb) - cp $CLIENT $MYDIR - dpkg -i "${MYDIR}/${CLIENT}" - apt-get -f -y install - dpkg -i "${MYDIR}/${CLIENT}" -} - -envSetup -deployTackerClient diff --git a/testcases/features/sfc/test-vnfd1.yaml b/testcases/features/sfc/test-vnfd1.yaml deleted file mode 100644 index 5c672e38..00000000 --- a/testcases/features/sfc/test-vnfd1.yaml +++ /dev/null @@ -1,31 +0,0 @@ -template_name: test-vnfd1 -description: firewall1-example - -service_properties: - Id: firewall1-vnfd - vendor: tacker - version: 1 - type: - - firewall1 -vdus: - vdu1: - id: vdu1 - vm_image: sf_nsh_colorado - instance_type: custom - service_type: firewall1 - - network_interfaces: - management: - network: example-net - management: true - - placement_policy: - availability_zone: nova - - auto-scaling: noop - monitoring_policy: noop - failure_policy: respawn - - config: - param0: key0 - param1: key1 diff --git a/testcases/features/sfc/test-vnfd2.yaml b/testcases/features/sfc/test-vnfd2.yaml deleted file mode 100644 index 8a570ab9..00000000 --- a/testcases/features/sfc/test-vnfd2.yaml +++ /dev/null @@ -1,31 +0,0 @@ -template_name: test-vnfd2 -description: firewall2-example - -service_properties: - Id: firewall2-vnfd - vendor: tacker - version: 1 - type: - - firewall2 -vdus: - vdu1: - id: vdu1 - vm_image: sf_nsh_colorado - instance_type: custom - service_type: firewall2 - - network_interfaces: - management: - network: example-net - management: true - - placement_policy: - availability_zone: nova - - auto-scaling: noop - monitoring_policy: noop - failure_policy: respawn - - config: - param0: key0 - param1: key1 |