diff options
-rwxr-xr-x | ci/exec_test.sh | 4 | ||||
-rw-r--r-- | testcases/features/sfc/SSHUtils.py | 68 | ||||
-rwxr-xr-x | testcases/features/sfc/compute_presetup_CI.bash | 4 | ||||
-rwxr-xr-x | testcases/features/sfc/delete.sh | 9 | ||||
-rw-r--r-- | testcases/features/sfc/ovs_utils.py | 65 | ||||
-rwxr-xr-x | testcases/features/sfc/sfc.py | 844 | ||||
-rwxr-xr-x | testcases/features/sfc/sfc_colorado1.py | 609 |
7 files changed, 553 insertions, 1050 deletions
diff --git a/ci/exec_test.sh b/ci/exec_test.sh index 6fdc42666..b0006f87c 100755 --- a/ci/exec_test.sh +++ b/ci/exec_test.sh @@ -159,11 +159,13 @@ function run_test(){ python ${FUNCTEST_REPO_DIR}/testcases/features/domino.py $report ;; "odl-sfc") + pip install --upgrade python-keystoneclient==1.7.4 ODL_SFC_DIR=${FUNCTEST_REPO_DIR}/testcases/features/sfc # pass FUNCTEST_REPO_DIR inside prepare_odl_sfc.bash FUNCTEST_REPO_DIR=${FUNCTEST_REPO_DIR} python ${ODL_SFC_DIR}/prepare_odl_sfc.py || exit $? source ${ODL_SFC_DIR}/tackerc - python ${ODL_SFC_DIR}/sfc_colorado1.py $report + python ${ODL_SFC_DIR}/sfc.py $report + pip install --upgrade python-keystoneclient==3.5.0 ;; "parser") python ${FUNCTEST_REPO_DIR}/testcases/vnf/vRNC/parser.py $report diff --git a/testcases/features/sfc/SSHUtils.py b/testcases/features/sfc/SSHUtils.py index 55cc5108b..9c8c2c727 100644 --- a/testcases/features/sfc/SSHUtils.py +++ b/testcases/features/sfc/SSHUtils.py @@ -16,16 +16,16 @@ import os logger = rl.Logger('SSHUtils').getLogger() -def get_ssh_client(hostname, username, password=None, jumphost=None): +def get_ssh_client(hostname, username, password=None, proxy=None): client = None try: - if jumphost is None: + if proxy is None: client = paramiko.SSHClient() else: - client = JumpHostHopClient() - client.configure_jump_host(jumphost['ip'], - jumphost['username'], - jumphost['password']) + client = ProxyHopClient() + client.configure_jump_host(proxy['ip'], + proxy['username'], + proxy['password']) if client is None: raise Exception('Could not connect to client') @@ -62,30 +62,30 @@ def put_file(ssh_conn, src, dest): return None -class JumpHostHopClient(paramiko.SSHClient): +class ProxyHopClient(paramiko.SSHClient): ''' - Connect to a remote server using a jumphost hop + Connect to a remote server using a proxy hop ''' def __init__(self, *args, **kwargs): - self.logger = rl.Logger("JumpHostHopClient").getLogger() - self.jumphost_ssh = None - self.jumphost_transport = None - self.jumphost_channel = None - self.jumphost_ip = None - self.jumphost_ssh_key = None + self.logger = rl.Logger("ProxyHopClient").getLogger() + self.proxy_ssh = None + self.proxy_transport = None + self.proxy_channel = None + self.proxy_ip = None + self.proxy_ssh_key = None self.local_ssh_key = os.path.join(os.getcwd(), 'id_rsa') - super(JumpHostHopClient, self).__init__(*args, **kwargs) + super(ProxyHopClient, self).__init__(*args, **kwargs) def configure_jump_host(self, jh_ip, jh_user, jh_pass, jh_ssh_key='/root/.ssh/id_rsa'): - self.jumphost_ip = jh_ip - self.jumphost_ssh_key = jh_ssh_key - self.jumphost_ssh = paramiko.SSHClient() - self.jumphost_ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy()) - self.jumphost_ssh.connect(jh_ip, - username=jh_user, - password=jh_pass) - self.jumphost_transport = self.jumphost_ssh.get_transport() + self.proxy_ip = jh_ip + self.proxy_ssh_key = jh_ssh_key + self.proxy_ssh = paramiko.SSHClient() + self.proxy_ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy()) + self.proxy_ssh.connect(jh_ip, + username=jh_user, + password=jh_pass) + self.proxy_transport = self.proxy_ssh.get_transport() def connect(self, hostname, port=22, username='root', password=None, pkey=None, key_filename=None, timeout=None, allow_agent=True, @@ -93,28 +93,28 @@ class JumpHostHopClient(paramiko.SSHClient): gss_kex=False, gss_deleg_creds=True, gss_host=None, banner_timeout=None): try: - if self.jumphost_ssh is None: + if self.proxy_ssh is None: raise Exception('You must configure the jump ' 'host before calling connect') - get_file_res = get_file(self.jumphost_ssh, - self.jumphost_ssh_key, + get_file_res = get_file(self.proxy_ssh, + self.proxy_ssh_key, self.local_ssh_key) if get_file_res is None: raise Exception('Could\'t fetch SSH key from jump host') - jumphost_key = (paramiko.RSAKey - .from_private_key_file(self.local_ssh_key)) + proxy_key = (paramiko.RSAKey + .from_private_key_file(self.local_ssh_key)) - self.jumphost_channel = self.jumphost_transport.open_channel( + self.proxy_channel = self.proxy_transport.open_channel( "direct-tcpip", (hostname, 22), - (self.jumphost_ip, 22)) + (self.proxy_ip, 22)) self.set_missing_host_key_policy(paramiko.AutoAddPolicy()) - super(JumpHostHopClient, self).connect(hostname, - username=username, - pkey=jumphost_key, - sock=self.jumphost_channel) + super(ProxyHopClient, self).connect(hostname, + username=username, + pkey=proxy_key, + sock=self.proxy_channel) os.remove(self.local_ssh_key) except Exception, e: self.logger.error(e) diff --git a/testcases/features/sfc/compute_presetup_CI.bash b/testcases/features/sfc/compute_presetup_CI.bash index 39276b43a..36148aa15 100755 --- a/testcases/features/sfc/compute_presetup_CI.bash +++ b/testcases/features/sfc/compute_presetup_CI.bash @@ -14,8 +14,8 @@ pushd $BASEDIR ip=$1 echo $ip -sshpass -p r00tme scp $ssh_options correct_classifier.bash ${INSTALLER_IP}:/root -sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'scp correct_classifier.bash '"$ip"':/root' +#sshpass -p r00tme scp $ssh_options correct_classifier.bash ${INSTALLER_IP}:/root +#sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'scp correct_classifier.bash '"$ip"':/root' sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'ssh root@'"$ip"' ifconfig br-int up' output=$(sshpass -p r00tme ssh $ssh_options root@${INSTALLER_IP} 'ssh root@'"$ip"' ip route | \ diff --git a/testcases/features/sfc/delete.sh b/testcases/features/sfc/delete.sh index b8d1fde1c..c04ae6375 100755 --- a/testcases/features/sfc/delete.sh +++ b/testcases/features/sfc/delete.sh @@ -8,11 +8,8 @@ tacker vnf-delete testVNF1 tacker vnf-delete testVNF2 tacker vnfd-delete test-vnfd1 tacker vnfd-delete test-vnfd2 -#openstack stack delete sfc --y -heat stack-delete sfc --y -#openstack stack delete sfc_test1 --y -heat stack-delete sfc_test1 --y -#openstack stack delete sfc_test2 --y -heat stack-delete sfc_test2 --y +openstack stack delete sfc --y +openstack stack delete sfc_test1 --y +openstack stack delete sfc_test2 --y nova delete client nova delete server diff --git a/testcases/features/sfc/ovs_utils.py b/testcases/features/sfc/ovs_utils.py index 83a36286d..af1f232c8 100644 --- a/testcases/features/sfc/ovs_utils.py +++ b/testcases/features/sfc/ovs_utils.py @@ -10,21 +10,30 @@ import functest.utils.functest_logger as rl import os import time +import shutil +import re logger = rl.Logger('ovs_utils').getLogger() class OVSLogger(object): - def __init__(self, basedir): - self.ovs_dir = os.path.join(basedir, 'odl-sfc/ovs') + def __init__(self, basedir, ft_resdir): + self.ovs_dir = basedir + self.ft_resdir = ft_resdir self.__mkdir_p(self.ovs_dir) def __mkdir_p(self, dirpath): if not os.path.exists(dirpath): os.makedirs(dirpath) - def __ssh_host(self, ssh_conn): - return ssh_conn.get_transport().getpeername()[0] + def __ssh_host(self, ssh_conn, host_prefix='10.20.0'): + try: + _, stdout, _ = ssh_conn.exec_command('hostname -I') + hosts = stdout.readline().strip().split(' ') + found_host = [h for h in hosts if h.startswith(host_prefix)][0] + return found_host + except Exception, e: + logger.error(e) def __dump_to_file(self, operation, host, text, timestamp=None): ts = (timestamp if timestamp is not None @@ -51,6 +60,13 @@ class OVSLogger(object): .format(cmd, e)) return None + def create_artifact_archive(self): + shutil.make_archive(self.ovs_dir, + 'zip', + root_dir=os.path.dirname(self.ovs_dir), + base_dir=self.ovs_dir) + shutil.copy2('{0}.zip'.format(self.ovs_dir), self.ft_resdir) + def ofctl_dump_flows(self, ssh_conn, br='br-int', choose_table=None, timestamp=None): try: @@ -78,3 +94,44 @@ class OVSLogger(object): except Exception, e: logger.error('[vsctl_show(ssh_client)]: {0}'.format(e)) return None + + def dump_ovs_logs(self, controller_clients, compute_clients, + related_error=None, timestamp=None): + if timestamp is None: + timestamp = time.strftime("%Y%m%d-%H%M%S") + + for controller_client in controller_clients: + self.ofctl_dump_flows(controller_client, + timestamp=timestamp) + self.vsctl_show(controller_client, + timestamp=timestamp) + + for compute_client in compute_clients: + self.ofctl_dump_flows(compute_client, + timestamp=timestamp) + self.vsctl_show(compute_client, + timestamp=timestamp) + + if related_error is not None: + dumpdir = os.path.join(self.ovs_dir, timestamp) + with open(os.path.join(dumpdir, 'error'), 'w') as f: + f.write(related_error) + + def ofctl_time_counter(self, ssh_conn): + try: + # We get the flows from table 11 + table = 11 + br = "br-int" + output = self.ofctl_dump_flows(ssh_conn, br, table) + pattern = "NXM_NX_NSP" + rsps = [] + lines = output.split(",") + for line in lines: + is_there = re.findall(pattern, line) + if is_there: + value = line.split(":")[1].split("-")[0] + rsps.append(value) + return rsps + except Exception, e: + logger.error('Error when countering %s' % e) + return None diff --git a/testcases/features/sfc/sfc.py b/testcases/features/sfc/sfc.py index 789bb714e..376083555 100755 --- a/testcases/features/sfc/sfc.py +++ b/testcases/features/sfc/sfc.py @@ -1,14 +1,16 @@ +import argparse import os import subprocess import sys import time - -import argparse -import paramiko - import functest.utils.functest_logger as ft_logger import functest.utils.functest_utils as ft_utils import functest.utils.openstack_utils as os_utils +import re +import json +import SSHUtils as ssh_utils +import ovs_utils +import thread parser = argparse.ArgumentParser() @@ -21,113 +23,152 @@ args = parser.parse_args() """ logging configuration """ logger = ft_logger.Logger("ODL_SFC").getLogger() -FUNCTEST_RESULTS_DIR = '/home/opnfv/functest/results/' +FUNCTEST_RESULTS_DIR = '/home/opnfv/functest/results/odl-sfc' FUNCTEST_REPO = ft_utils.FUNCTEST_REPO - +REPO_PATH = os.environ['repos_dir'] + '/functest/' HOME = os.environ['HOME'] + "/" - -VM_BOOT_TIMEOUT = 180 -INSTANCE_NAME = "client" +CLIENT = "client" +SERVER = "server" FLAVOR = "custom" IMAGE_NAME = "sf_nsh_colorado" IMAGE_FILENAME = "sf_nsh_colorado.qcow2" IMAGE_FORMAT = "qcow2" -IMAGE_PATH = "/home/opnfv/functest/data" + "/" + IMAGE_FILENAME +IMAGE_DIR = "/home/opnfv/functest/data" +IMAGE_PATH = IMAGE_DIR + "/" + IMAGE_FILENAME +IMAGE_URL = "http://artifacts.opnfv.org/sfc/demo/" + IMAGE_FILENAME # NEUTRON Private Network parameters - NET_NAME = "example-net" SUBNET_NAME = "example-subnet" SUBNET_CIDR = "11.0.0.0/24" ROUTER_NAME = "example-router" - SECGROUP_NAME = "example-sg" SECGROUP_DESCR = "Example Security group" +SFC_TEST_DIR = REPO_PATH + "/testcases/features/sfc/" +TACKER_SCRIPT = SFC_TEST_DIR + "sfc_tacker.bash" +TACKER_CHANGECLASSI = SFC_TEST_DIR + "sfc_change_classi.bash" +ssh_options = '-q -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no' +json_results = {"tests": 4, "failures": 0} + +PROXY = { + 'ip': '10.20.0.2', + 'username': 'root', + 'password': 'r00tme' +} + +# run given command locally and return commands output if success + + +def run_cmd(cmd, wdir=None, ignore_stderr=False, ignore_no_output=True): + pipe = subprocess.Popen(cmd, shell=True, + stdin=subprocess.PIPE, + stdout=subprocess.PIPE, + stderr=subprocess.PIPE, cwd=wdir) + + (output, errors) = pipe.communicate() + if output: + output = output.strip() + if pipe.returncode < 0: + logger.error(errors) + return False + if errors: + logger.error(errors) + if ignore_stderr: + return True + else: + return False -INSTANCE_NAME_2 = "server" + if ignore_no_output: + if not output: + return True -# TEST_DB = ft_utils.get_parameter_from_yaml("results.test_db_url") + return output -PRE_SETUP_SCRIPT = 'sfc_pre_setup.bash' -TACKER_SCRIPT = 'sfc_tacker.bash' -TEARDOWN_SCRIPT = "sfc_teardown.bash" -TACKER_CHANGECLASSI = "sfc_change_classi.bash" +# run given command on OpenStack controller -ssh_options = '-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no' +def run_cmd_on_cntlr(cmd): + ip_cntlrs = get_openstack_node_ips("controller") + if not ip_cntlrs: + return None -def check_ssh(ip): - cmd = "sshpass -p opnfv ssh " + ssh_options + " -q " + ip + " exit" - success = subprocess.call(cmd, shell=True) == 0 - if not success: - logger.debug("Wating for SSH connectivity in SF with IP: %s" % ip) - return success + ssh_cmd = "ssh %s %s %s" % (ssh_options, ip_cntlrs[0], cmd) + return run_cmd_on_fm(ssh_cmd) +# run given command on OpenStack Compute node -def main(): - # Allow any port so that tacker commands reaches the server. - # This will be deleted when tacker is included in OPNFV installation +def run_cmd_on_compute(cmd): + ip_computes = get_openstack_node_ips("compute") + if not ip_computes: + return None - status = "PASS" - failures = 0 - start_time = time.time() - json_results = {} - - contr_cmd = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " 'fuel node'|grep controller|awk '{print $10}'") - logger.info("Executing script to get ip_server: '%s'" % contr_cmd) - process = subprocess.Popen(contr_cmd, - shell=True, - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) - ip_server = process.stdout.readline().rstrip() - - contr_cmd2 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " 'fuel node'|grep compute|awk '{print $10}'") - logger.info("Executing script to get ip_compute: '%s'" % contr_cmd2) - process = subprocess.Popen(contr_cmd2, - shell=True, - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) - ip_compute = process.stdout.readline().rstrip() - - iptable_cmd1 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " ssh " + ip_server + " iptables -P INPUT ACCEPT ") - iptable_cmd2 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " ssh " + ip_server + " iptables -t nat -P INPUT ACCEPT ") - iptable_cmd3 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " ssh " + ssh_options + " " + ip_server + - " iptables -A INPUT -m state" - " --state NEW,ESTABLISHED,RELATED -j ACCEPT") - - logger.info("Changing firewall policy in controller: '%s'" % iptable_cmd1) - subprocess.call(iptable_cmd1, shell=True, stderr=subprocess.PIPE) - - logger.info("Changing firewall policy in controller: '%s'" % iptable_cmd2) - subprocess.call(iptable_cmd2, shell=True, stderr=subprocess.PIPE) - - logger.info("Changing firewall policy in controller: '%s'" % iptable_cmd3) - subprocess.call(iptable_cmd3, shell=True, stderr=subprocess.PIPE) - -# Getting the different clients + ssh_cmd = "ssh %s %s %s" % (ssh_options, ip_computes[0], cmd) + return run_cmd_on_fm(ssh_cmd) - nova_client = os_utils.get_nova_client() - neutron_client = os_utils.get_neutron_client() - glance_client = os_utils.get_glance_client() +# run given command on Fuel Master + + +def run_cmd_on_fm(cmd, username="root", passwd="r00tme"): + ip = os.environ.get("INSTALLER_IP") + ssh_cmd = "sshpass -p %s ssh %s %s@%s %s" % ( + passwd, ssh_options, username, ip, cmd) + return run_cmd(ssh_cmd) + +# run given command on Remote Machine, Can be VM + + +def run_cmd_remote(ip, cmd, username="root", passwd="opnfv"): + ssh_opt_append = "%s -o ConnectTimeout=50 " % ssh_options + ssh_cmd = "sshpass -p %s ssh %s %s@%s %s" % ( + passwd, ssh_opt_append, username, ip, cmd) + return run_cmd(ssh_cmd) -# Download the image +# Get OpenStack Nodes IP Address + +def get_openstack_node_ips(role): + fuel_env = os.environ.get("FUEL_ENV") + if fuel_env is not None: + cmd = "fuel2 node list -f json -e %s" % fuel_env + else: + cmd = "fuel2 node list -f json" + + nodes = run_cmd_on_fm(cmd) + ips = [] + nodes = json.loads(nodes) + for node in nodes: + if role in node["roles"]: + ips.append(node["ip"]) + + return ips + +# Configures IPTABLES on OpenStack Controller + + +def configure_iptables(): + iptable_cmds = ["iptables -P INPUT ACCEPT", + "iptables -t nat -P INPUT ACCEPT", + "iptables -A INPUT -m state \ + --state NEW,ESTABLISHED,RELATED -j ACCEPT"] + + for cmd in iptable_cmds: + logger.info("Configuring %s on contoller" % cmd) + run_cmd_on_cntlr(cmd) + + return + + +def download_image(): if not os.path.isfile(IMAGE_PATH): logger.info("Downloading image") - ft_utils.download_url( - "http://artifacts.opnfv.org/sfc/demo/sf_nsh_colorado.qcow2", - "/home/opnfv/functest/data/") - else: - logger.info("Using old image") + ft_utils.download_url(IMAGE_URL, IMAGE_DIR) -# Create glance image and the neutron network + logger.info("Using old image") + return + +def setup_glance(glance_client): image_id = os_utils.create_glance_image(glance_client, IMAGE_NAME, IMAGE_PATH, @@ -135,379 +176,392 @@ def main(): container="bare", public=True) - network_dic = os_utils.create_network_full(neutron_client, - NET_NAME, - SUBNET_NAME, - ROUTER_NAME, - SUBNET_CIDR) - if not network_dic: - logger.error( - "There has been a problem when creating the neutron network") + return image_id + + +def setup_neutron(neutron_client): + n_dict = os_utils.create_network_full(neutron_client, + NET_NAME, + SUBNET_NAME, + ROUTER_NAME, + SUBNET_CIDR) + if not n_dict: + logger.error("failed to create neutron network") sys.exit(-1) - network_id = network_dic["net_id"] + network_id = n_dict["net_id"] + return network_id - sg_id = os_utils.create_security_group_full(neutron_client, - SECGROUP_NAME, SECGROUP_DESCR) +def setup_ingress_egress_secgroup(neutron_client, protocol, + min_port=None, max_port=None): secgroups = os_utils.get_security_groups(neutron_client) - for sg in secgroups: os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'ingress', 'udp', - port_range_min=67, - port_range_max=68) + 'ingress', protocol, + port_range_min=min_port, + port_range_max=max_port) os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'egress', 'udp', - port_range_min=67, - port_range_max=68) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'ingress', 'tcp', - port_range_min=22, - port_range_max=22) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'egress', 'tcp', - port_range_min=22, - port_range_max=22) + 'egress', protocol, + port_range_min=min_port, + port_range_max=max_port) + return + + +def setup_security_groups(neutron_client): + sg_id = os_utils.create_security_group_full(neutron_client, + SECGROUP_NAME, SECGROUP_DESCR) + setup_ingress_egress_secgroup(neutron_client, "icmp") + setup_ingress_egress_secgroup(neutron_client, "udp", 67, 68) + setup_ingress_egress_secgroup(neutron_client, "tcp", 22, 22) + setup_ingress_egress_secgroup(neutron_client, "tcp", 80, 80) + return sg_id - _, custom_flv_id = os_utils.get_or_create_flavor( - 'custom', 1500, 10, 1, public=True) - if not custom_flv_id: - logger.error("Failed to create custom flavor") - sys.exit(1) - # boot INSTANCE - logger.info("Creating instance '%s'..." % INSTANCE_NAME) +def boot_instance(nova_client, name, flavor, image_id, network_id, sg_id): + logger.info("Creating instance '%s'..." % name) logger.debug( "Configuration:\n name=%s \n flavor=%s \n image=%s \n " - "network=%s \n" % (INSTANCE_NAME, FLAVOR, image_id, network_id)) - instance = os_utils.create_instance_and_wait_for_active(FLAVOR, + "network=%s \n" % (name, flavor, image_id, network_id)) + + instance = os_utils.create_instance_and_wait_for_active(flavor, image_id, network_id, - INSTANCE_NAME, - av_zone='nova') + name) if instance is None: logger.error("Error while booting instance.") sys.exit(-1) - # Retrieve IP of INSTANCE + instance_ip = instance.networks.get(NET_NAME)[0] logger.debug("Instance '%s' got private ip '%s'." % - (INSTANCE_NAME, instance_ip)) + (name, instance_ip)) - logger.info("Adding '%s' to security group '%s'..." - % (INSTANCE_NAME, SECGROUP_NAME)) + logger.info("Adding '%s' to security group %s" % (name, SECGROUP_NAME)) os_utils.add_secgroup_to_instance(nova_client, instance.id, sg_id) - logger.info("Creating floating IP for VM '%s'..." % INSTANCE_NAME) - floatip_dic = os_utils.create_floating_ip(neutron_client) - floatip_client = floatip_dic['fip_addr'] - # floatip_id = floatip_dic['fip_id'] + return instance_ip - if floatip_client is None: - logger.error("Cannot create floating IP.") - sys.exit(-1) - logger.info("Floating IP created: '%s'" % floatip_client) - logger.info("Associating floating ip: '%s' to VM '%s' " - % (floatip_client, INSTANCE_NAME)) - if not os_utils.add_floating_ip(nova_client, instance.id, floatip_client): - logger.error("Cannot associate floating IP to VM.") - sys.exit(-1) +def ping(remote, pkt_cnt=1, iface=None, retries=100, timeout=None): + ping_cmd = 'ping' -# STARTING SECOND VM (server) ### + if timeout: + ping_cmd = ping_cmd + ' -w %s' % timeout - # boot INTANCE - logger.info("Creating instance '%s'..." % INSTANCE_NAME_2) - logger.debug( - "Configuration:\n name=%s \n flavor=%s \n image=%s \n " - "network=%s \n" % (INSTANCE_NAME, FLAVOR, image_id, network_id)) - instance_2 = os_utils.create_instance_and_wait_for_active(FLAVOR, - image_id, - network_id, - INSTANCE_NAME_2, - av_zone='nova') - - if instance_2 is None: - logger.error("Error while booting instance.") - sys.exit(-1) - # Retrieve IP of INSTANCE - instance_ip_2 = instance_2.networks.get(NET_NAME)[0] - logger.debug("Instance '%s' got private ip '%s'." % - (INSTANCE_NAME_2, instance_ip_2)) + grep_cmd = "grep -e 'packet loss' -e rtt" - logger.info("Adding '%s' to security group '%s'..." - % (INSTANCE_NAME_2, SECGROUP_NAME)) - os_utils.add_secgroup_to_instance(nova_client, instance_2.id, sg_id) + if iface is not None: + ping_cmd = ping_cmd + ' -I %s' % iface - logger.info("Creating floating IP for VM '%s'..." % INSTANCE_NAME_2) - floatip_dic = os_utils.create_floating_ip(neutron_client) - floatip_server = floatip_dic['fip_addr'] - # floatip_id = floatip_dic['fip_id'] + ping_cmd = ping_cmd + ' -i 0 -c %d %s' % (pkt_cnt, remote) + cmd = ping_cmd + '|' + grep_cmd - if floatip_server is None: - logger.error("Cannot create floating IP.") - sys.exit(-1) - logger.info("Floating IP created: '%s'" % floatip_server) + while retries > 0: + output = run_cmd(cmd) + if not output: + return False - logger.info("Associating floating ip: '%s' to VM '%s' " - % (floatip_server, INSTANCE_NAME_2)) + match = re.search('(\d*)% packet loss', output) + if not match: + return False - if not os_utils.add_floating_ip(nova_client, - instance_2.id, - floatip_server): - logger.error("Cannot associate floating IP to VM.") - sys.exit(-1) - - # CREATION OF THE 2 SF #### + packet_loss = int(match.group(1)) + if packet_loss == 0: + return True - tacker_script = "%s/testcases/features/sfc/%s" % \ - (FUNCTEST_REPO, TACKER_SCRIPT) - logger.info("Executing tacker script: '%s'" % tacker_script) - subprocess.call(tacker_script, shell=True) + retries = retries - 1 - # SSH CALL TO START HTTP SERVER - ssh = paramiko.SSHClient() - ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy()) + return False - try: - ssh.connect(floatip_server, username="root", - password="opnfv", timeout=2) - command = "python -m SimpleHTTPServer 80 > /dev/null 2>&1 &" - logger.info("Starting HTTP server") - (stdin, stdout, stderr) = ssh.exec_command(command) - except: - logger.debug("Waiting for %s..." % floatip_server) - time.sleep(6) - # timeout -= 1 - instances = nova_client.servers.list(search_opts={'all_tenants': 1}) +def get_floating_ips(nova_client, neutron_client): ips = [] - try: - for instance in instances: - if "server" not in instance.name: - if "client" not in instance.name: - logger.debug( - "This is the instance name: %s " % instance.name) - floatip_dic = os_utils.create_floating_ip(neutron_client) - floatip = floatip_dic['fip_addr'] - ips.append(floatip) - instance.add_floating_ip(floatip) - except: - logger.debug("Problems assigning floating IP to SFs") - - # If no IPs were obtained, then we cant continue - if not ips: - logger.error('Failed to obtain IPs, cant continue, exiting') - return - - logger.debug("Floating IPs for SFs: %s..." % ips) - - # Check SSH connectivity to VNFs - r = 0 - retries = 100 - check = [False, False] + instances = nova_client.servers.list(search_opts={'all_tenants': 1}) + for instance in instances: + floatip_dic = os_utils.create_floating_ip(neutron_client) + floatip = floatip_dic['fip_addr'] + instance.add_floating_ip(floatip) + logger.info("Instance name and ip %s:%s " % (instance.name, floatip)) + logger.info("Waiting for instance %s:%s to come up" % + (instance.name, floatip)) + if not ping(floatip): + logger.info("Instance %s:%s didn't come up" % + (instance.name, floatip)) + sys.exit(1) + + if instance.name == "server": + logger.info("Server:%s is reachable" % floatip) + server_ip = floatip + elif instance.name == "client": + logger.info("Client:%s is reachable" % floatip) + client_ip = floatip + else: + logger.info("SF:%s is reachable" % floatip) + ips.append(floatip) + + return server_ip, client_ip, ips[1], ips[0] + +# Start http server on a give machine, Can be VM + + +def start_http_server(ip): + cmd = "\'python -m SimpleHTTPServer 80" + cmd = cmd + " > /dev/null 2>&1 &\'" + return run_cmd_remote(ip, cmd) + +# Set firewall using vxlan_tool.py on a give machine, Can be VM + + +def vxlan_firewall(sf, iface="eth0", port="22", block=True): + cmd = "python vxlan_tool.py" + cmd = cmd + " -i " + iface + " -d forward -v off" + if block: + cmd = "python vxlan_tool.py -i eth0 -d forward -v off -b " + port + + cmd = "sh -c 'cd /root;nohup " + cmd + " > /dev/null 2>&1 &'" + run_cmd_remote(sf, cmd) + +# Run netcat on a give machine, Can be VM + + +def netcat(s_ip, c_ip, port="80", timeout=5): + cmd = "nc -zv " + cmd = cmd + " -w %s %s %s" % (timeout, s_ip, port) + cmd = cmd + " 2>&1" + output = run_cmd_remote(c_ip, cmd) + logger.info("%s" % output) + return output + +def is_ssh_blocked(srv_prv_ip, client_ip): + res = netcat(srv_prv_ip, client_ip, port="22") + match = re.search("nc:.*timed out:.*", res, re.M) + if match: + return True + + return False + + +def is_http_blocked(srv_prv_ip, client_ip): + res = netcat(srv_prv_ip, client_ip, port="80") + match = re.search(".* 80 port.* succeeded!", res, re.M) + if match: + return False + + return True + + +def capture_err_logs(controller_clients, compute_clients, error): + ovs_logger = ovs_utils.OVSLogger( + os.path.join(os.getcwd(), 'ovs-logs'), + FUNCTEST_RESULTS_DIR) + + timestamp = time.strftime("%Y%m%d-%H%M%S") + ovs_logger.dump_ovs_logs(controller_clients, + compute_clients, + related_error=error, + timestamp=timestamp) + return + + +def update_json_results(name, result): + json_results.update({name: result}) + if result is not "Passed": + json_results["failures"] += 1 + + return + + +def get_ssh_clients(role): + clients = [] + for ip in get_openstack_node_ips(role): + s_client = ssh_utils.get_ssh_client(ip, + 'root', + proxy=PROXY) + clients.append(s_client) + + return clients + +# Check SSH connectivity to VNFs + + +def check_ssh(ips, retries=100): + check = [False, False] logger.info("Checking SSH connectivity to the SFs with ips %s" % str(ips)) - while r < retries and not all(check): - try: - check = [check_ssh(ips[0]), check_ssh(ips[1])] - except Exception: - logger.exception("SSH check failed") - check = [False, False] + while retries and not all(check): + for index, ip in enumerate(ips): + check[index] = run_cmd_remote(ip, "exit") + + if all(check): + logger.info("SSH connectivity to the SFs established") + return True + time.sleep(3) - r += 1 + retries -= 1 - if not all(check): - logger.error("Cannot establish SSH connection to the SFs") - sys.exit(1) + return False - logger.info("SSH connectivity to the SFs established") +# Measure the time it takes to update the classification rules - # SSH TO START THE VXLAN_TOOL ON SF1 - logger.info("Configuring the SFs") - try: - ssh.connect(ips[0], username="root", - password="opnfv", timeout=2) - command = ("nohup python vxlan_tool.py -i eth0 " - "-d forward -v off -b 80 > /dev/null 2>&1 &") - (stdin, stdout, stderr) = ssh.exec_command(command) - except: - logger.debug("Waiting for %s..." % ips[0]) - time.sleep(6) - # timeout -= 1 - try: - while 1: - (stdin, stdout, stderr) = ssh.exec_command( - "ps aux | grep \"vxlan_tool.py\" | grep -v grep") - if len(stdout.readlines()) > 0: - logger.debug("HTTP firewall started") - break +def capture_time_log(compute_clients): + ovs_logger = ovs_utils.OVSLogger( + os.path.join(os.getcwd(), 'ovs-logs'), + "test") + i = 0 + first_RSP = "" + start_time = time.time() + while True: + rsps = ovs_logger.ofctl_time_counter(compute_clients[0]) + if not i: + if len(rsps) > 0: + first_RSP = rsps[0] + i = i + 1 else: - logger.debug("HTTP firewall not started") - time.sleep(3) - except Exception: - logger.exception("vxlan_tool not started in SF1") + first_RSP = 0 + i = i + 1 + if (len(rsps) > 1): + if(first_RSP != rsps[0]): + if (rsps[0] == rsps[1]): + stop_time = time.time() + logger.info("classification rules updated") + difference = stop_time - start_time + logger.info("It took %s seconds" % difference) + break + time.sleep(1) + return - # SSH TO START THE VXLAN_TOOL ON SF2 - try: - ssh.connect(ips[1], username="root", - password="opnfv", timeout=2) - command = ("nohup python vxlan_tool.py -i eth0 " - "-d forward -v off -b 22 > /dev/null 2>&1 &") - (stdin, stdout, stderr) = ssh.exec_command(command) - except: - logger.debug("Waiting for %s..." % ips[1]) - time.sleep(6) - # timeout -= 1 - try: - while 1: - (stdin, stdout, stderr) = ssh.exec_command( - "ps aux | grep \"vxlan_tool.py\" | grep -v grep") - if len(stdout.readlines()) > 0: - logger.debug("SSH firewall started") - break - else: - logger.debug("SSH firewall not started") - time.sleep(3) - except Exception: - logger.exception("vxlan_tool not started in SF2") +def main(): + installer_type = os.environ.get("INSTALLER_TYPE") + if installer_type != "fuel": + logger.error( + '\033[91mCurrently supported only Fuel Installer type\033[0m') + sys.exit(1) + + installer_ip = os.environ.get("INSTALLER_IP") + if not installer_ip: + logger.error( + '\033[91minstaller ip is not set\033[0m') + logger.error( + '\033[91mexport INSTALLER_IP=<ip>\033[0m') + sys.exit(1) + + start_time = time.time() + status = "PASS" + configure_iptables() + download_image() + _, custom_flv_id = os_utils.get_or_create_flavor( + FLAVOR, 1500, 10, 1, public=True) + if not custom_flv_id: + logger.error("Failed to create custom flavor") + sys.exit(1) - # SSH to modify the classification flows in compute + glance_client = os_utils.get_glance_client() + neutron_client = os_utils.get_neutron_client() + nova_client = os_utils.get_nova_client() - contr_cmd3 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " 'ssh " + ip_compute + " 'bash correct_classifier.bash''") - logger.info("Executing script to modify the classi: '%s'" % contr_cmd3) - process = subprocess.Popen(contr_cmd3, - shell=True, - stdout=subprocess.PIPE) + controller_clients = get_ssh_clients("controller") + compute_clients = get_ssh_clients("compute") - i = 0 + image_id = setup_glance(glance_client) + network_id = setup_neutron(neutron_client) + sg_id = setup_security_groups(neutron_client) - # SSH TO EXECUTE cmd_client - logger.info("TEST STARTED") - time.sleep(70) - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 22 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - # WRITE THE CORRECT WAY TO DO LOGGING - if "timed out" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 1 [PASSED] " - "==> SSH BLOCKED" + '\033[0m') - i = i + 1 - json_results.update({"Test 1: SSH Blocked": "Passed"}) - else: - logger.error('\033[91m' + "TEST 1 [FAILED] " - "==> SSH NOT BLOCKED" + '\033[0m') - status = "FAIL" - json_results.update({"Test 1: SSH Blocked": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - # SSH TO EXECUTE cmd_client + boot_instance( + nova_client, CLIENT, FLAVOR, image_id, network_id, sg_id) + srv_prv_ip = boot_instance( + nova_client, SERVER, FLAVOR, image_id, network_id, sg_id) + + subprocess.call(TACKER_SCRIPT, shell=True) + + # Start measuring the time it takes to implement the classification rules try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 80 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - if "succeeded" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 2 [PASSED] " - "==> HTTP WORKS" + '\033[0m') - i = i + 1 - json_results.update({"Test 2: HTTP works": "Passed"}) - else: - logger.error('\033[91m' + "TEST 2 [FAILED] " - "==> HTTP BLOCKED" + '\033[0m') - status = "FAIL" - json_results.update({"Test 2: HTTP works": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - # CHANGE OF CLASSIFICATION # - logger.info("Changing the classification") - tacker_classi = "%s/testcases/features/sfc/%s" % \ - (FUNCTEST_REPO, TACKER_CHANGECLASSI) - subprocess.call(tacker_classi, shell=True) + thread.start_new_thread(capture_time_log, (compute_clients,)) + except Exception, e: + logger.error("Unable to start the thread that counts time %s" % e) - logger.info("Wait for ODL to update the classification rules in OVS") - time.sleep(10) + server_ip, client_ip, sf1, sf2 = get_floating_ips( + nova_client, neutron_client) - # SSH to modify the classification flows in compute + if not check_ssh([sf1, sf2]): + logger.error("Cannot establish SSH connection to the SFs") + sys.exit(1) - contr_cmd4 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " 'ssh " + ip_compute + " 'bash correct_classifier.bash''") - logger.info("Executing script to modify the classi: '%s'" % contr_cmd4) - process = subprocess.Popen(contr_cmd4, - shell=True, - stdout=subprocess.PIPE) + logger.info("Starting HTTP server on %s" % server_ip) + if not start_http_server(server_ip): + logger.error( + '\033[91mFailed to start HTTP server on %s\033[0m' % server_ip) + sys.exit(1) - # SSH TO EXECUTE cmd_client + logger.info("Starting HTTP firewall on %s" % sf2) + vxlan_firewall(sf2, port="80") + logger.info("Starting SSH firewall on %s" % sf1) + vxlan_firewall(sf1, port="22") + logger.info("Wait for ODL to update the classification rules in OVS") + time.sleep(120) + + logger.info("Test SSH") + if is_ssh_blocked(srv_prv_ip, client_ip): + logger.info('\033[92mTEST 1 [PASSED] ==> SSH BLOCKED\033[0m') + update_json_results("Test 1: SSH Blocked", "Passed") + else: + error = ('\033[91mTEST 1 [FAILED] ==> SSH NOT BLOCKED\033[0m') + logger.error(error) + capture_err_logs(controller_clients, compute_clients, error) + update_json_results("Test 1: SSH Blocked", "Failed") + + logger.info("Test HTTP") + if not is_http_blocked(srv_prv_ip, client_ip): + logger.info('\033[92mTEST 2 [PASSED] ==> HTTP WORKS\033[0m') + update_json_results("Test 2: HTTP works", "Passed") + else: + error = ('\033[91mTEST 2 [FAILED] ==> HTTP BLOCKED\033[0m') + logger.error(error) + capture_err_logs(controller_clients, compute_clients, error) + update_json_results("Test 2: HTTP works", "Failed") + + logger.info("Changing the classification") + subprocess.call(TACKER_CHANGECLASSI, shell=True) + + # Start measuring the time it takes to implement the classification rules try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 80 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - if "timed out" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 3 [PASSED] " - "==> HTTP BLOCKED" + '\033[0m') - i = i + 1 - json_results.update({"Test 3: HTTP Blocked": "Passed"}) - else: - logger.error('\033[91m' + "TEST 3 [FAILED] " - "==> HTTP NOT BLOCKED" + '\033[0m') - status = "FAIL" - json_results.update({"Test 3: HTTP Blocked": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - # SSH TO EXECUTE cmd_client - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 22 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - if "succeeded" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 4 [PASSED] " - "==> SSH WORKS" + '\033[0m') - i = i + 1 - json_results.update({"Test 4: SSH works": "Passed"}) - else: - logger.error('\033[91m' + "TEST 4 [FAILED] " - "==> SSH BLOCKED" + '\033[0m') - status = "FAIL" - json_results.update({"Test 4: SSH works": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - if i == 4: - for x in range(0, 5): - logger.info('\033[92m' + "SFC TEST WORKED" - " :) \n" + '\033[0m') + thread.start_new_thread(capture_time_log, (compute_clients,)) + except Exception, e: + logger.error("Unable to start the thread that counts time %s" % e) + + logger.info("Wait for ODL to update the classification rules in OVS") + time.sleep(100) + + logger.info("Test HTTP") + if is_http_blocked(srv_prv_ip, client_ip): + logger.info('\033[92mTEST 3 [PASSED] ==> HTTP Blocked\033[0m') + update_json_results("Test 3: HTTP Blocked", "Passed") + else: + error = ('\033[91mTEST 3 [FAILED] ==> HTTP WORKS\033[0m') + logger.error(error) + capture_err_logs(controller_clients, compute_clients, error) + update_json_results("Test 3: HTTP Blocked", "Failed") + + logger.info("Test SSH") + if not is_ssh_blocked(srv_prv_ip, client_ip): + logger.info('\033[92mTEST 4 [PASSED] ==> SSH Works\033[0m') + update_json_results("Test 4: SSH Works", "Passed") + else: + error = ('\033[91mTEST 4 [FAILED] ==> SSH BLOCKED\033[0m') + logger.error(error) + capture_err_logs(controller_clients, compute_clients, error) + update_json_results("Test 4: SSH Works", "Failed") + + if json_results["failures"]: + status = "FAIL" + logger.error('\033[91mSFC TESTS: %s :( FOUND %s FAIL \033[0m' % ( + status, json_results["failures"])) if args.report: stop_time = time.time() - json_results.update({"tests": "4", "failures": int(failures)}) logger.debug("Promise Results json: " + str(json_results)) ft_utils.push_results_to_db("sfc", "functest-odl-sfc", @@ -515,10 +569,12 @@ def main(): stop_time, status, json_results) + if status == "PASS": + logger.info('\033[92mSFC ALL TESTS: %s :)\033[0m' % status) sys.exit(0) - else: - sys.exit(1) + + sys.exit(1) if __name__ == '__main__': main() diff --git a/testcases/features/sfc/sfc_colorado1.py b/testcases/features/sfc/sfc_colorado1.py deleted file mode 100755 index 734421fc3..000000000 --- a/testcases/features/sfc/sfc_colorado1.py +++ /dev/null @@ -1,609 +0,0 @@ -import os -import subprocess -import sys -import time -import argparse -import paramiko - -import functest.utils.functest_logger as ft_logger -import functest.utils.functest_utils as ft_utils -import functest.utils.openstack_utils as os_utils -import SSHUtils as ssh_utils -import ovs_utils - -parser = argparse.ArgumentParser() - -parser.add_argument("-r", "--report", - help="Create json result file", - action="store_true") - -args = parser.parse_args() - -""" logging configuration """ -logger = ft_logger.Logger("ODL_SFC").getLogger() - -FUNCTEST_RESULTS_DIR = '/home/opnfv/functest/results/' -FUNCTEST_REPO = ft_utils.FUNCTEST_REPO - -HOME = os.environ['HOME'] + "/" - -VM_BOOT_TIMEOUT = 180 -INSTANCE_NAME = "client" -FLAVOR = "custom" -IMAGE_NAME = "sf_nsh_colorado" -IMAGE_FILENAME = "sf_nsh_colorado.qcow2" -IMAGE_FORMAT = "qcow2" -IMAGE_PATH = "/home/opnfv/functest/data" + "/" + IMAGE_FILENAME - -# NEUTRON Private Network parameters - -NET_NAME = "example-net" -SUBNET_NAME = "example-subnet" -SUBNET_CIDR = "11.0.0.0/24" -ROUTER_NAME = "example-router" - -SECGROUP_NAME = "example-sg" -SECGROUP_DESCR = "Example Security group" - -INSTANCE_NAME_2 = "server" - -# TEST_DB = ft_utils.get_parameter_from_yaml("results.test_db_url") - -PRE_SETUP_SCRIPT = 'sfc_pre_setup.bash' -TACKER_SCRIPT = 'sfc_tacker.bash' -TEARDOWN_SCRIPT = "sfc_teardown.bash" -TACKER_CHANGECLASSI = "sfc_change_classi.bash" - -ssh_options = '-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no' - - -def check_ssh(ip): - cmd = "sshpass -p opnfv ssh " + ssh_options + " -q " + ip + " exit" - success = subprocess.call(cmd, shell=True) == 0 - if not success: - logger.debug("Wating for SSH connectivity in SF with IP: %s" % ip) - return success - - -def main(): - - # Allow any port so that tacker commands reaches the server. - # This will be deleted when tacker is included in OPNFV installation - - status = "PASS" - failures = 0 - start_time = time.time() - json_results = {} - - contr_cmd = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " 'fuel node'|grep controller|awk '{print $10}'") - logger.info("Executing script to get ip_server: '%s'" % contr_cmd) - process = subprocess.Popen(contr_cmd, - shell=True, - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) - ip_server = process.stdout.readline().rstrip() - - contr_cmd2 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " 'fuel node'|grep compute|awk '{print $10}'") - logger.info("Executing script to get ip_compute: '%s'" % contr_cmd2) - process = subprocess.Popen(contr_cmd2, - shell=True, - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) - ip_compute = process.stdout.readline().rstrip() - - iptable_cmd1 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " ssh " + ip_server + " iptables -P INPUT ACCEPT ") - iptable_cmd2 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " ssh " + ip_server + " iptables -t nat -P INPUT ACCEPT ") - iptable_cmd3 = ("sshpass -p r00tme ssh " + ssh_options + " root@10.20.0.2" - " ssh " + ssh_options + " " + ip_server + - " iptables -A INPUT -m state" - " --state NEW,ESTABLISHED,RELATED -j ACCEPT") - - logger.info("Changing firewall policy in controller: '%s'" % iptable_cmd1) - subprocess.call(iptable_cmd1, shell=True, stderr=subprocess.PIPE) - - logger.info("Changing firewall policy in controller: '%s'" % iptable_cmd2) - subprocess.call(iptable_cmd2, shell=True, stderr=subprocess.PIPE) - - logger.info("Changing firewall policy in controller: '%s'" % iptable_cmd3) - subprocess.call(iptable_cmd2, shell=True, stderr=subprocess.PIPE) - -# Getting the different clients - - nova_client = os_utils.get_nova_client() - neutron_client = os_utils.get_neutron_client() - glance_client = os_utils.get_glance_client() - - ovs_logger = ovs_utils.OVSLogger(FUNCTEST_RESULTS_DIR) - jumphost = { - 'ip': '10.20.0.2', - 'username': 'root', - 'password': 'r00tme' - } - controller_client = ssh_utils.get_ssh_client(ip_server, - 'root', - jumphost=jumphost) - -# Download the image - - if not os.path.isfile(IMAGE_PATH): - logger.info("Downloading image") - ft_utils.download_url( - "http://artifacts.opnfv.org/sfc/demo/sf_nsh_colorado.qcow2", - "/home/opnfv/functest/data/") - else: - logger.info("Using old image") - -# Create glance image and the neutron network - - image_id = os_utils.create_glance_image(glance_client, - IMAGE_NAME, - IMAGE_PATH, - disk=IMAGE_FORMAT, - container="bare", - public=True) - - network_dic = os_utils.create_network_full(neutron_client, - NET_NAME, - SUBNET_NAME, - ROUTER_NAME, - SUBNET_CIDR) - if not network_dic: - logger.error( - "There has been a problem when creating the neutron network") - sys.exit(-1) - - network_id = network_dic["net_id"] - - sg_id = os_utils.create_security_group_full(neutron_client, - SECGROUP_NAME, SECGROUP_DESCR) - - secgroups = os_utils.get_security_groups(neutron_client) - - for sg in secgroups: - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'ingress', 'udp', - port_range_min=67, - port_range_max=68) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'egress', 'udp', - port_range_min=67, - port_range_max=68) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'ingress', 'tcp', - port_range_min=22, - port_range_max=22) - os_utils.create_secgroup_rule(neutron_client, sg['id'], - 'egress', 'tcp', - port_range_min=22, - port_range_max=22) - - _, custom_flv_id = os_utils.get_or_create_flavor( - 'custom', 1500, 10, 1, public=True) - if not custom_flv_id: - logger.error("Failed to create custom flavor") - sys.exit(1) - - iterator = 0 - while(iterator < 6): - # boot INSTANCE - logger.info("Creating instance '%s'..." % INSTANCE_NAME) - logger.debug( - "Configuration:\n name=%s \n flavor=%s \n image=%s \n " - "network=%s \n" % (INSTANCE_NAME, FLAVOR, image_id, network_id)) - instance = os_utils.create_instance_and_wait_for_active( - FLAVOR, - image_id, - network_id, - INSTANCE_NAME, - av_zone='nova') - - if instance is None: - logger.error("Error while booting instance.") - iterator += 1 - continue - # Retrieve IP of INSTANCE - instance_ip = instance.networks.get(NET_NAME)[0] - logger.debug("Instance '%s' got private ip '%s'." % - (INSTANCE_NAME, instance_ip)) - - logger.info("Adding '%s' to security group '%s'..." - % (INSTANCE_NAME, SECGROUP_NAME)) - os_utils.add_secgroup_to_instance(nova_client, instance.id, sg_id) - - logger.info("Creating floating IP for VM '%s'..." % INSTANCE_NAME) - floatip_dic = os_utils.create_floating_ip(neutron_client) - floatip_client = floatip_dic['fip_addr'] - # floatip_id = floatip_dic['fip_id'] - - if floatip_client is None: - logger.error("Cannot create floating IP.") - iterator += 1 - continue - logger.info("Floating IP created: '%s'" % floatip_client) - - logger.info("Associating floating ip: '%s' to VM '%s' " - % (floatip_client, INSTANCE_NAME)) - if not os_utils.add_floating_ip(nova_client, - instance.id, - floatip_client): - logger.error("Cannot associate floating IP to VM.") - iterator += 1 - continue - - # STARTING SECOND VM (server) ### - - # boot INTANCE - logger.info("Creating instance '%s'..." % INSTANCE_NAME_2) - logger.debug( - "Configuration:\n name=%s \n flavor=%s \n image=%s \n " - "network=%s \n" % (INSTANCE_NAME, FLAVOR, image_id, network_id)) - instance_2 = os_utils.create_instance_and_wait_for_active( - FLAVOR, - image_id, - network_id, - INSTANCE_NAME_2, - av_zone='nova') - - if instance_2 is None: - logger.error("Error while booting instance.") - iterator += 1 - continue - # Retrieve IP of INSTANCE - instance_ip_2 = instance_2.networks.get(NET_NAME)[0] - logger.debug("Instance '%s' got private ip '%s'." % - (INSTANCE_NAME_2, instance_ip_2)) - - logger.info("Adding '%s' to security group '%s'..." - % (INSTANCE_NAME_2, SECGROUP_NAME)) - os_utils.add_secgroup_to_instance(nova_client, instance_2.id, sg_id) - - logger.info("Creating floating IP for VM '%s'..." % INSTANCE_NAME_2) - floatip_dic = os_utils.create_floating_ip(neutron_client) - floatip_server = floatip_dic['fip_addr'] - # floatip_id = floatip_dic['fip_id'] - - if floatip_server is None: - logger.error("Cannot create floating IP.") - iterator += 1 - continue - logger.info("Floating IP created: '%s'" % floatip_server) - - logger.info("Associating floating ip: '%s' to VM '%s' " - % (floatip_server, INSTANCE_NAME_2)) - - if not os_utils.add_floating_ip(nova_client, - instance_2.id, - floatip_server): - logger.error("Cannot associate floating IP to VM.") - iterator += 1 - continue - - # CREATION OF THE 2 SF #### - - tacker_script = "%s/testcases/features/sfc/%s" % \ - (FUNCTEST_REPO, TACKER_SCRIPT) - logger.info("Executing tacker script: '%s'" % tacker_script) - subprocess.call(tacker_script, shell=True) - - # SSH CALL TO START HTTP SERVER - ssh = paramiko.SSHClient() - ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy()) - - try: - ssh.connect(floatip_server, username="root", - password="opnfv", timeout=2) - command = "python -m SimpleHTTPServer 80 > /dev/null 2>&1 &" - logger.info("Starting HTTP server") - (stdin, stdout, stderr) = ssh.exec_command(command) - except: - logger.debug("Waiting for %s..." % floatip_server) - time.sleep(6) - # timeout -= 1 - - instances = nova_client.servers.list(search_opts={'all_tenants': 1}) - ips = [] - try: - for instance in instances: - if "server" not in instance.name: - if "client" not in instance.name: - logger.debug( - "This is the instance name: %s " % instance.name) - floatip_dic = os_utils.create_floating_ip( - neutron_client) - floatip = floatip_dic['fip_addr'] - ips.append(floatip) - instance.add_floating_ip(floatip) - except: - logger.debug("Problems assigning floating IP to SFs") - - # If no IPs were obtained, then we cant continue - if not ips: - logger.error('Failed to obtain IPs, cant continue, exiting') - return - - logger.debug("Floating IPs for SFs: %s..." % ips) - - # Check SSH connectivity to VNFs - r = 0 - retries = 100 - check = [False, False] - - logger.info("Checking SSH connectivity to the SFs with ips {0}" - .format(str(ips))) - while r < retries and not all(check): - try: - check = [check_ssh(ips[0]), check_ssh(ips[1])] - except Exception: - logger.exception("SSH check failed") - check = [False, False] - time.sleep(3) - r += 1 - - if not all(check): - logger.error("Cannot establish SSH connection to the SFs") - iterator += 1 - continue - - logger.info("SSH connectivity to the SFs established") - - # SSH TO START THE VXLAN_TOOL ON SF1 - logger.info("Configuring the SFs") - try: - ssh.connect(ips[0], username="root", - password="opnfv", timeout=2) - command = ("nohup python vxlan_tool.py -i eth0 " - "-d forward -v off -b 80 > /dev/null 2>&1 &") - (stdin, stdout, stderr) = ssh.exec_command(command) - except: - logger.debug("Waiting for %s..." % ips[0]) - time.sleep(6) - # timeout -= 1 - - try: - n = 0 - while 1: - (stdin, stdout, stderr) = ssh.exec_command( - "ps aux | grep \"vxlan_tool.py\" | grep -v grep") - if len(stdout.readlines()) > 0: - logger.debug("HTTP firewall started") - break - else: - n += 1 - if (n > 7): - break - logger.debug("HTTP firewall not started") - time.sleep(3) - except Exception: - logger.exception("vxlan_tool not started in SF1") - - # SSH TO START THE VXLAN_TOOL ON SF2 - try: - ssh.connect(ips[1], username="root", - password="opnfv", timeout=2) - command = ("nohup python vxlan_tool.py -i eth0 " - "-d forward -v off -b 22 > /dev/null 2>&1 &") - (stdin, stdout, stderr) = ssh.exec_command(command) - except: - logger.debug("Waiting for %s..." % ips[1]) - time.sleep(6) - # timeout -= 1 - - try: - n = 0 - while 1: - (stdin, stdout, stderr) = ssh.exec_command( - "ps aux | grep \"vxlan_tool.py\" | grep -v grep") - if len(stdout.readlines()) > 0: - logger.debug("SSH firewall started") - break - else: - n += 1 - if (n > 7): - break - logger.debug("SSH firewall not started") - time.sleep(3) - except Exception: - logger.exception("vxlan_tool not started in SF2") - - # SSH to modify the classification flows in compute - - contr_cmd3 = ("sshpass -p r00tme ssh " + ssh_options + " " - "root@10.20.0.2 'ssh " + ip_compute + " " - "'bash correct_classifier.bash''") - logger.info("Executing script to modify the classi: '%s'" % contr_cmd3) - process = subprocess.Popen(contr_cmd3, - shell=True, - stdout=subprocess.PIPE) - - i = 0 - - # SSH TO EXECUTE cmd_client - logger.info("TEST STARTED") - time.sleep(70) - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 22 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - # WRITE THE CORRECT WAY TO DO LOGGING - if "timed out" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 1 [PASSED] " - "==> SSH BLOCKED" + '\033[0m') - i = i + 1 - json_results.update({"Test 1: SSH Blocked": "Passed"}) - else: - timestamp = time.strftime("%Y%m%d-%H%M%S") - error = ('\033[91m' + "TEST 1 [FAILED] " - "==> SSH NOT BLOCKED" + '\033[0m') - logger.error(error) - ovs_logger.ofctl_dump_flows(controller_client, - timestamp=timestamp) - ovs_logger.vsctl_show(controller_client, - timestamp=timestamp) - - dumpdir = os.path.join(ovs_logger.ovs_dir, timestamp) - with open(os.path.join(dumpdir, 'error'), 'w') as f: - f.write(error) - status = "FAIL" - json_results.update({"Test 1: SSH Blocked": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - # SSH TO EXECUTE cmd_client - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 80 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - if "succeeded" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 2 [PASSED] " - "==> HTTP WORKS" + '\033[0m') - i = i + 1 - json_results.update({"Test 2: HTTP works": "Passed"}) - else: - error = ('\033[91m' + "TEST 2 [FAILED] " - "==> HTTP BLOCKED" + '\033[0m') - logger.error(error) - ovs_logger.ofctl_dump_flows(controller_client, - timestamp=timestamp) - ovs_logger.vsctl_show(controller_client, - timestamp=timestamp) - - dumpdir = os.path.join(ovs_logger.ovs_dir, timestamp) - with open(os.path.join(dumpdir, 'error'), 'w') as f: - f.write(error) - status = "FAIL" - json_results.update({"Test 2: HTTP works": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - # CHANGE OF CLASSIFICATION # - logger.info("Changing the classification") - tacker_classi = "%s/testcases/features/sfc/%s" % \ - (FUNCTEST_REPO, TACKER_CHANGECLASSI) - subprocess.call(tacker_classi, shell=True) - - logger.info("Wait for ODL to update the classification rules in OVS") - time.sleep(10) - - # SSH to modify the classification flows in compute - - contr_cmd4 = ("sshpass -p r00tme ssh " + ssh_options + " " - "root@10.20.0.2 'ssh " + ip_compute + " " - "'bash correct_classifier.bash''") - logger.info("Executing script to modify the classi: '%s'" % contr_cmd4) - process = subprocess.Popen(contr_cmd4, - shell=True, - stdout=subprocess.PIPE) - - # SSH TO EXECUTE cmd_client - - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 80 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - if "timed out" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 3 [PASSED] " - "==> HTTP BLOCKED" + '\033[0m') - i = i + 1 - json_results.update({"Test 3: HTTP Blocked": "Passed"}) - else: - error = ('\033[91m' + "TEST 3 [FAILED] " - "==> HTTP NOT BLOCKED" + '\033[0m') - logger.error(error) - ovs_logger.ofctl_dump_flows(controller_client, - timestamp=timestamp) - ovs_logger.vsctl_show(controller_client, - timestamp=timestamp) - - dumpdir = os.path.join(ovs_logger.ovs_dir, timestamp) - with open(os.path.join(dumpdir, 'error'), 'w') as f: - f.write(error) - status = "FAIL" - json_results.update({"Test 3: HTTP Blocked": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - # SSH TO EXECUTE cmd_client - try: - ssh.connect(floatip_client, username="root", - password="opnfv", timeout=2) - command = "nc -w 5 -zv " + instance_ip_2 + " 22 2>&1" - (stdin, stdout, stderr) = ssh.exec_command(command) - - if "succeeded" in stdout.readlines()[0]: - logger.info('\033[92m' + "TEST 4 [PASSED] " - "==> SSH WORKS" + '\033[0m') - i = i + 1 - json_results.update({"Test 4: SSH works": "Passed"}) - else: - error = ('\033[91m' + "TEST 4 [FAILED] " - "==> SSH BLOCKED" + '\033[0m') - logger.error(error) - ovs_logger.ofctl_dump_flows(controller_client, - timestamp=timestamp) - ovs_logger.vsctl_show(controller_client, - timestamp=timestamp) - - dumpdir = os.path.join(ovs_logger.ovs_dir, timestamp) - with open(os.path.join(dumpdir, 'error'), 'w') as f: - f.write(error) - status = "FAIL" - json_results.update({"Test 4: SSH works": "Failed"}) - failures += 1 - except: - logger.debug("Waiting for %s..." % floatip_client) - time.sleep(6) - # timeout -= 1 - - iterator += 1 - if i == 4: - for x in range(0, 5): - logger.info('\033[92m' + "SFC TEST WORKED" - " :) \n" + '\033[0m') - break - else: - logger.info("Iterating again!") - delete = "bash %s/testcases/features/sfc/delete.sh" % \ - (FUNCTEST_REPO) - try: - subprocess.call(delete, shell=True, stderr=subprocess.PIPE) - time.sleep(10) - except Exception, e: - logger.error("Problem when executing the delete.sh") - logger.error("Problem %s" % e) - - if args.report: - stop_time = time.time() - json_results.update({"tests": "4", "failures": int(failures)}) - logger.debug("Promise Results json: " + str(json_results)) - ft_utils.push_results_to_db("sfc", - "functest-odl-sfc", - start_time, - stop_time, - status, - json_results) - if status == "PASS": - sys.exit(0) - else: - sys.exit(1) - -if __name__ == '__main__': - main() |