diff options
Diffstat (limited to 'tests')
-rw-r--r-- | tests/VES_Reference.sh | 411 | ||||
-rwxr-xr-x | tests/blueprints/tosca-vnfd-hello-ves/blueprint.yaml | 107 | ||||
-rwxr-xr-x | tests/blueprints/tosca-vnfd-hello-ves/start.sh | 84 | ||||
-rw-r--r-- | tests/utils/setenv.sh | 82 | ||||
-rw-r--r-- | tests/vHello_VES.sh | 322 | ||||
-rw-r--r-- | tests/vLamp_Ansible_VES.sh | 46 |
6 files changed, 1052 insertions, 0 deletions
diff --git a/tests/VES_Reference.sh b/tests/VES_Reference.sh new file mode 100644 index 0000000..fa5bde1 --- /dev/null +++ b/tests/VES_Reference.sh @@ -0,0 +1,411 @@ +#!/bin/bash +# Copyright 2016 AT&T Intellectual Property, Inc +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +# What this is: Deployment script for the VNF Event Stream (VES) Reference VNF +# and Test Collector. Runs the VES Collector in a docker container on the +# OPNFV jumphost, and the VES Reference VNF as an OpenStack VM. +# +# Status: this is a work in progress, under test. +# +# How to use: +# $ git clone https://gerrit.opnfv.org/gerrit/ves +# $ cd ves/tests +# $ bash VES_Reference.sh [setup|start|run|stop|clean] +# setup: setup test environment +# start: install blueprint and run test +# run: setup test environment and run test +# stop: stop test and uninstall blueprint +# clean: cleanup after test + +trap 'fail' ERR + +pass() { + echo "$0: Hooray!" + set +x #echo off + exit 0 +} + +fail() { + echo "$0: Test Failed!" + set +x + exit 1 +} + +function setenv () { + echo "$0: Setup OpenStack environment variables" + source utils/setenv.sh /tmp/VES +} + +get_floating_net () { + network_ids=($(neutron net-list|grep -v "+"|grep -v name|awk '{print $2}')) + for id in ${network_ids[@]}; do + [[ $(neutron net-show ${id}|grep 'router:external'|grep -i "true") != "" ]] && FLOATING_NETWORK_ID=${id} + done + if [[ $FLOATING_NETWORK_ID ]]; then + FLOATING_NETWORK_NAME=$(openstack network show $FLOATING_NETWORK_ID | awk "/ name / { print \$4 }") + else + echo "$0: Floating network not found" + exit 1 + fi +} + +try () { + count=$1 + $3 + while [[ $? -eq 1 && $count -gt 0 ]] + do + sleep $2 + let count=$count-1 + $3 + done + if [[ $count -eq 0 ]]; then echo "$0: Command \"$3\" was not successful after $1 tries"; fi +} + +function create_container () { + echo "$0: Creating docker container" + echo "$0: Copy this script to /tmp/VES" + mkdir /tmp/VES + cp $0 /tmp/VES/. + chmod 755 /tmp/VES/*.sh + + echo "$0: reset blueprints folder" + if [[ -d /tmp/VES/blueprints/ ]]; then rm -rf /tmp/VES/blueprints/; fi + mkdir -p /tmp/VES/blueprints/ + + echo "$0: Setup admin-openrc.sh" + setenv + + echo "$0: Setup container" + if [ "$dist" == "Ubuntu" ]; then + # xenial is needed for python 3.5 + sudo docker pull ubuntu:xenial + sudo service docker start + # Port 30000 is the default for the VES Collector + sudo docker run -it -d -p 30000:30000 -v /tmp/VES/:/tmp/VES \ + --name VES ubuntu:xenial /bin/bash + else + # Centos + echo "Centos-based install" + sudo tee /etc/yum.repos.d/docker.repo <<-'EOF' +[dockerrepo] +name=Docker Repository--parents +baseurl=https://yum.dockerproject.org/repo/main/centos/7/ +enabled=1 +gpgcheck=1 +gpgkey=https://yum.dockerproject.org/gpg +EOF + sudo yum install -y docker-engine + # xenial is needed for python 3.5 + sudo service docker start + sudo docker pull ubuntu:xenial + # Port 30000 is the default for the VES Collector + sudo docker run -i -t -d -p 30000:30000 -v /tmp/VES/:/tmp/VES \ + --name VES ubuntu:xenial /bin/bash + fi +} + +setup_Openstack () { + echo "$0: install OpenStack clients" + pip install --upgrade python-openstackclient + pip install --upgrade python-glanceclient + pip install --upgrade python-neutronclient + pip install --upgrade python-heatclient +# pip install --upgrade keystonemiddleware + + echo "$0: setup OpenStack environment" + source /tmp/VES/admin-openrc.sh + + echo "$0: determine external (public) network as the floating ip network" echo "$0: setup OpenStack environment" + get_floating_net + + echo "$0: Setup centos7-server glance image if needed" + if [[ -z $(openstack image list | awk "/ centos7-server / { print \$2 }") ]]; \ + then glance --os-image-api-version 1 image-create \ + --name centos7-server \ + --disk-format qcow2 \ + --location http://cloud.centos.org/centos/7/images/CentOS-7-x86_64-GenericCloud-1607.qcow2 \ + --container-format bare; fi + + if [[ -z $(neutron net-list | awk "/ internal / { print \$2 }") ]]; then + echo "$0: Create internal network" + neutron net-create internal + + echo "$0: Create internal subnet" + neutron subnet-create internal 10.0.0.0/24 --name internal \ + --gateway 10.0.0.1 --enable-dhcp \ + --allocation-pool start=10.0.0.2,end=10.0.0.254 \ + --dns-nameserver 8.8.8.8 + fi + + if [[ -z $(neutron router-list | awk "/ public_router / { print \$2 }") ]]; then + echo "$0: Create router" + neutron router-create public_router + + echo "$0: Create router gateway" + neutron router-gateway-set public_router $FLOATING_NETWORK_NAME + + echo "$0: Add router interface for internal network" + neutron router-interface-add public_router subnet=internal + fi +} + +setup_Collector () { + echo "$0: Install dependencies - OS specific" + if [ "$dist" == "Ubuntu" ]; then + apt-get update + apt-get install -y python + apt-get install -y python-pip + apt-get install -y git + else + yum install -y python + yum install -y python-pip + yum install -y git + fi + pip install --upgrade pip + + echo "$0: clone VES Collector repo" + cd /tmp/VES/blueprints/ + git clone https://github.com/att/evel-test-collector.git + echo "$0: update collector.conf" + cd /tmp/VES/blueprints/evel-test-collector + sed -i -- 's~/var/log/att/~/tmp/VES/~g' config/collector.conf +} + +start_Collector () { + echo "$0: start the VES Collector" + cd /tmp/VES/blueprints/evel-test-collector + python code/collector/collector.py \ + --config config/collector.conf \ + --section default \ + --verbose +} + +setup_Reference_VNF_VM () { + echo "$0: Create Nova key pair" + nova keypair-add VES > /tmp/VES/VES-key + chmod 600 /tmp/VES/VES-key + + echo "$0: Add ssh key" + eval $(ssh-agent -s) + ssh-add /tmp/VES/VES-key + + echo "$0: clone VES Reference VNF repo" + cd /tmp/VES/blueprints/ + git clone https://github.com/att/evel-reporting-reference-vnf.git + + echo "$0: customize VES Reference VNF Heat template" + cd evel-reporting-reference-vnf/hot + ID=$(openstack image list | awk "/ centos7-server / { print \$2 }") + sed -i -- "s/40299aa3-2921-43b0-86b9-56c28a2b5232/$ID/g" event_reporting_vnf.env.yaml + ID=$(neutron net-list | awk "/ internal / { print \$2 }") + sed -i -- "s/84985f60-fbba-4a78-ba83-2815ff620dbc/$ID/g" event_reporting_vnf.env.yaml + sed -i -- "s/127.0.0.1/$JUMPHOST/g" event_reporting_vnf.env.yaml + sed -i -- "s/my-keyname/VES/g" event_reporting_vnf.env.yaml + + echo "$0: Create VES Reference VNF via Heat" + heat stack-create -e event_reporting_vnf.env.yaml \ + -f event_reporting_vnf.template.yaml VES + + echo "$0: Wait for VES Reference VNF to go Active" + COUNTER=0 + until [[ $(heat stack-list | awk "/ VES / { print \$6 }") == "CREATE_COMPLETE" ]]; do + sleep 5 + let COUNTER+=1 + if [[ $COUNTER > "20" ]]; then fail; fi + done + + echo "$0: Get Server ID" + SID=$(heat resource-list VES | awk "/ OS::Nova::Server / { print \$4 }") + + echo "$0: associate SSH security group" + # TODO: Update Heat template to include security group + if [[ $(openstack security group list | awk "/ vHello / { print \$2 }") ]]; then neutron security-group-delete vHello; fi + openstack security group create VES_Reference + openstack security group rule create --ingress --protocol TCP --dst-port 22:22 VES_Reference + openstack security group rule create --ingress --protocol TCP --dst-port 80:80 VES_Reference + openstack server add security group $SID VES_Reference + + echo "$0: associate floating IP" + # TODO: Update Heat template to include floating IP (if supported) + FIP=$(openstack floating ip create $FLOATING_NETWORK_NAME | awk "/floating_ip_address/ { print \$4 }") + nova floating-ip-associate $SID $FIP + +# scp -i /tmp/VES/VES-key -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no /tmp/VES/VES_Reference.sh centos@$FIP:/home/centos + scp -i /tmp/VES/VES-key -o UserKnownHostsFile=/dev/null \ + -o StrictHostKeyChecking=no \ + $0 centos@$FIP:/home/centos +# run thru setup_Reference_VNF manually to verify +# ssh -i /tmp/VES/VES-key -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no centos@$FIP +# ssh -i /tmp/VES/VES-key -x -o UserKnownHostsFile=/dev/null +# -o StrictHostKeyChecking=no +# centos@$FIP \ +# "nohup source $0 setup_VNF &" +} + +setup_Reference_VNF () { + echo "$0: Install dependencies" + sudo yum update -y + sudo yum install -y wget + sudo yum install -y gcc + sudo yum install -y openssl-devel + sudo yum install -y epel-release + sudo yum install -y python-pip + sudo pip install --upgrade pip + sudo yum install -y git + + echo "$0: Install Django" + sudo pip install django + + echo "$0: Install Apache" + sudo yum install -y httpd httpd-devel + + echo "$0: Install mod_python" + sudo yum install -y python-devel + mkdir ~/mod_python-3.4.1 + cd ~/mod_python-3.4.1 + wget http://dist.modpython.org/dist/mod_python-3.4.1.tgz + tar xvf mod_python-3.4.1.tgz + cd mod_python-3.4.1 + + # Edit .../dist/version.sh to remove the dependency on Git as described at + # http://stackoverflow.com/questions/20022952/fatal-not-a-git-repository-when-installing-mod-python + sed \ + -e 's/(git describe --always)/(git describe --always 2>\/dev\/null)/g' \ + -e 's/`git describe --always`/`git describe --always 2>\/dev\/null`/g' \ + -i $( find . -type f -name Makefile\* -o -name version.sh ) + + ./configure + make + sudo make install + make test + + echo "$0: Install mod_wsgi" + sudo yum install -y mod_wsgi + + echo "$0: clone VES Reference VNF repo" + cd ~ + git clone https://github.com/att/evel-reporting-reference-vnf.git + + echo "$0: Setup collector" + + sudo mkdir -p /opt/att/collector + sudo install -m=644 -t /opt/att/collector ~/evel-reporting-reference-vnf/code/collector/* + + echo "$0: Setup Reference VNF website" + sudo mkdir -p /opt/att/website/ + sudo cp -r ~/evel-reporting-reference-vnf/code/webserver/django/* /opt/att/website/ + sudo chown -R root:root /opt/att/website/ + sudo mkdir -p /var/log/att/ + echo "eh?" | sudo tee /var/log/att/django.log + + echo "$0: Create database" + + cd /opt/att/website + sudo python manage.py migrate + sudo python manage.py createsuperuser + sudo rm -f /var/log/att/django.log + + sudo systemctl daemon-reload + sudo systemctl enable httpd + sudo systemctl restart httpd + + echo "$0: Setup website backend" + sudo mkdir -p /opt/att/backend/ + sudo install -m=644 -t /opt/att/backend ~/evel-reporting-reference-vnf/code/backend/* + sudo install -m=644 ~/evel-reporting-reference-vnf/config/backend.service /etc/systemd/system + sudo systemctl daemon-reload + sudo systemctl enable backend + sudo systemctl restart backend + + + echo "$0: Change security context for database" + chcon -t httpd_sys_content_t db.sqlite3 + chcon -t httpd_sys_content_t . + setsebool -P httpd_unified 1 + setsebool -P httpd_can_network_connect=1 + + echo "$0: Gather static files" + sudo python manage.py collectstatic + + echo "$0: Install jsonschema" + sudo pip install jsonschema + + echo "$0: Put backend.service into /etc/systemd/system" + sudo systemctl daemon-reload + sudo systemctl start backend + sudo systemctl status backend + sudo systemctl enable backend + + # from initialize-event-database.sh + cd /opt/att/website + sudo python manage.py migrate + sudo python manage.py createsuperuser + + # from go-webserver.sh + sudo python /opt/att/website/manage.py runserver & + + # from go-backend.sh + sudo python /opt/att/backend/backend.py --config ~/evel-reporting-reference-vnf/config/backend.conf --section default --verbose & +} + +clean () { + echo "$0: delete container" + CONTAINER=$(sudo docker ps -a | awk "/VES/ { print \$1 }") + sudo docker stop $CONTAINER + sudo docker rm -v $CONTAINER +} + +forward_to_container () { + echo "$0: pass $1 command to VES_Reference.sh in container" + CONTAINER=$(sudo docker ps -a | awk "/VES/ { print \$1 }") + sudo docker exec $CONTAINER /bin/bash /tmp/VES/VES_Reference.sh $1 $1 + if [ $? -eq 1 ]; then fail; fi +} + +dist=`grep DISTRIB_ID /etc/*-release | awk -F '=' '{print $2}'` +case "$1" in + setup) + if [[ $# -eq 1 ]]; then + create_container + echo "$0: Execute VES_Reference.sh in the container" + CONTAINER=$(sudo docker ps -l | awk "/VES/ { print \$1 }") + if [ "$dist" == "Ubuntu" ]; then + sudo docker exec -it $CONTAINER /bin/bash /tmp/VES/VES_Reference.sh setup setup + else + sudo docker exec -i -t $CONTAINER /bin/bash /tmp/VES/VES_Reference.sh setup setup + fi + else + # Running in the container, continue VES setup + setup_Collector + setup_Openstack + setup_Reference_VNF_VM + start_Collector + fi + pass + ;; + setup_VNF) + setup_Reference_VNF + ;; + clean) + echo "$0: Uninstall" + clean + pass + ;; + *) + echo "usage: bash VES_Reference.sh [setup|clean]" + echo "setup: setup test environment" + echo "clean: cleanup after test" + fail +esac diff --git a/tests/blueprints/tosca-vnfd-hello-ves/blueprint.yaml b/tests/blueprints/tosca-vnfd-hello-ves/blueprint.yaml new file mode 100755 index 0000000..fc9e1b8 --- /dev/null +++ b/tests/blueprints/tosca-vnfd-hello-ves/blueprint.yaml @@ -0,0 +1,107 @@ +tosca_definitions_version: tosca_simple_profile_for_nfv_1_0_0 + +description: Hello VES + +metadata: + template_name: tosca-vnfd-hello-ves + +topology_template: + node_templates: + VDU1: + type: tosca.nodes.nfv.VDU.Tacker + capabilities: + nfv_compute: + properties: + num_cpus: 1 + mem_size: 1024 MB + disk_size: 4 GB + properties: + image: models-xenial-server + availability_zone: nova + mgmt_driver: noop + config: | + param0: key1 + param1: key2 + + CP1: + type: tosca.nodes.nfv.CP.Tacker + properties: + management: true + anti_spoofing_protection: false + requirements: + - virtualLink: + node: VL1 + - virtualBinding: + node: VDU1 + + CP2: + type: tosca.nodes.nfv.CP.Tacker + properties: + anti_spoofing_protection: false + requirements: + - virtualLink: + node: VL2 + - virtualBinding: + node: VDU1 + + VL1: + type: tosca.nodes.nfv.VL + properties: + network_name: vnf_mgmt + vendor: Tacker + + VL2: + type: tosca.nodes.nfv.VL + properties: + network_name: vnf_private + vendor: Tacker + + VDU2: + type: tosca.nodes.nfv.VDU.Tacker + capabilities: + nfv_compute: + properties: + num_cpus: 1 + mem_size: 1024 MB + disk_size: 4 GB + properties: + image: models-xenial-server + availability_zone: nova + mgmt_driver: noop + config: | + param0: key1 + param1: key2 + + CP3: + type: tosca.nodes.nfv.CP.Tacker + properties: + management: true + anti_spoofing_protection: false + requirements: + - virtualLink: + node: VL3 + - virtualBinding: + node: VDU2 + + CP4: + type: tosca.nodes.nfv.CP.Tacker + properties: + anti_spoofing_protection: false + requirements: + - virtualLink: + node: VL4 + - virtualBinding: + node: VDU2 + + VL3: + type: tosca.nodes.nfv.VL + properties: + network_name: vnf_mgmt + vendor: Tacker + + VL4: + type: tosca.nodes.nfv.VL + properties: + network_name: vnf_private + vendor: Tacker + diff --git a/tests/blueprints/tosca-vnfd-hello-ves/start.sh b/tests/blueprints/tosca-vnfd-hello-ves/start.sh new file mode 100755 index 0000000..6c8fbeb --- /dev/null +++ b/tests/blueprints/tosca-vnfd-hello-ves/start.sh @@ -0,0 +1,84 @@ +#!/bin/bash +# Copyright 2016 AT&T Intellectual Property, Inc +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +# What this is: Startup script for a simple web server as part of the +# vHello_VES test of the OPNFV VES project. +# +# Status: this is a work in progress, under test. +# +# How to use: +# $ bash start.sh IP ID +# IP: IP address of the collector +# ID: username:password to use in REST +# + +echo "$0: Setup website and dockerfile" +mkdir ~/www +mkdir ~/www/html + +# ref: https://hub.docker.com/_/nginx/ +cat > ~/www/Dockerfile <<EOM +FROM nginx +COPY html /usr/share/nginx/html +EOM + +cat > ~/www/html/index.html <<EOM +<!DOCTYPE html> +<html> +<head> +<title>Hello World!</title> +<meta name="viewport" content="width=device-width, minimum-scale=1.0, initial-scale=1"/> +<style> +body { width: 100%; background-color: white; color: black; padding: 0px; margin: 0px; font-family: sans-serif; font-size:100%; } +</style> +</head> +<body> +Hello World!<br> +<a href="http://wiki.opnfv.org"><img src="https://www.opnfv.org/sites/all/themes/opnfv/logo.png"></a> +</body></html> +EOM + +echo "$0: Install docker" +# Per https://docs.docker.com/engine/installation/linux/ubuntulinux/ +# Per https://www.digitalocean.com/community/tutorials/how-to-install-and-use-docker-on-ubuntu-16-04 +sudo apt-get install apt-transport-https ca-certificates +sudo apt-key adv --keyserver hkp://p80.pool.sks-keyservers.net:80 --recv-keys 58118E89F3A912897C070ADBF76221572C52609D +echo "deb https://apt.dockerproject.org/repo ubuntu-xenial main" | sudo tee /etc/apt/sources.list.d/docker.list +sudo apt-get update +sudo apt-get purge lxc-docker +sudo apt-get install -y linux-image-extra-$(uname -r) linux-image-extra-virtual +sudo apt-get install -y docker-engine + +echo "$0: Get nginx container and start website in docker" +# Per https://hub.docker.com/_/nginx/ +sudo docker pull nginx +cd ~/www +sudo docker build -t vhello . +sudo docker run --name vHello -d -p 80:80 vhello + +echo "$0: setup VES event delivery for the nginx server" + +# id=$(sudo ls /var/lib/docker/containers) +# sudo tail -f /var/lib/docker/containers/$id/$id-json.log + +export COL_IP=$1 +export COL_ID=$2 + +while true +do + sleep 30 + curl --user $COL_ID -H "Content-Type: application/json" -X POST -d '{ "event": { "commonEventHeader": { "domain": "fault", "eventType": "Fault_MobileCallRecording_PilotNumberPoolExhaustion", "eventId": "ab305d54-85b4-a31b-7db2-fb6b9e546015", "sequence": "0", "priority": "High", "sourceId": "de305d54-75b4-431b-adb2-eb6b9e546014", "sourceName": "EricssonECE", "functionalRole": "SCF", "startEpochMicrosec": "1413378172000000", "lastEpochMicrosec": "1413378172000000", "reportingEntityId": "de305d54-75b4-431b-adb2-eb6b9e546014", "reportingEntityName": "EricssonECE" }, "faultFields": { "alarmCondition": "PilotNumberPoolExhaustion", "eventSourceType": "other(0)", "specificProblem": "Calls cannot complete because pilot numbers are unavailable", "eventSeverity": "CRITICAL", "vfStatus": "Active" } } }' http://$COL_IP:30000/eventListener/v1 +done + diff --git a/tests/utils/setenv.sh b/tests/utils/setenv.sh new file mode 100644 index 0000000..94c0b0b --- /dev/null +++ b/tests/utils/setenv.sh @@ -0,0 +1,82 @@ +#!/bin/bash +# Copyright 2016 AT&T Intellectual Property, Inc +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +# What this is: OpenStack environment file setup for OPNFV deployments. Sets up +# the environment parameters allowing use of OpenStack CLI commands, and as needed +# for OPNFV test scripts. +# +# Status: this is a work in progress, under test. +# +# How to use: +# $ wget https://git.opnfv.org/cgit/ves/plain/tests/utils/setenv.sh -O [folder] +# folder: folder to place the script in +# $ source /tmp/setenv.sh [target] +# folder: folder in which to put the created admin-openrc.sh file + +# TODO: Find a more precise way to determine the OPNFV install... currently +# this assumes that the script is running on the OPNFV jumphost, and +# Ubuntu=JOID, Centos=Apex + +dist=`grep DISTRIB_ID /etc/*-release | awk -F '=' '{print $2}'` + +if [ "$dist" == "Ubuntu" ]; then + # Ubuntu: assumes JOID-based install, and that this script is being run on the jumphost. + echo "$0: Ubuntu-based install" + echo "$0: Create the environment file" + KEYSTONE_HOST=$(juju status --format=short | awk "/keystone\/0/ { print \$3 }") + cat <<EOF >$1/admin-openrc.sh +export CONGRESS_HOST=$(juju status --format=short | awk "/openstack-dashboard/ { print \$3 }") +export HORIZON_HOST=$(juju status --format=short | awk "/openstack-dashboard/ { print \$3 }") +export KEYSTONE_HOST=$KEYSTONE_HOST +export CEILOMETER_HOST=$(juju status --format=short | awk "/ceilometer\/0/ { print \$3 }") +export CINDER_HOST=$(juju status --format=short | awk "/cinder\/0/ { print \$3 }") +export GLANCE_HOST=$(juju status --format=short | awk "/glance\/0/ { print \$3 }") +export NEUTRON_HOST=$(juju status --format=short | awk "/neutron-api\/0/ { print \$3 }") +export NOVA_HOST=$(juju status --format=short | awk "/nova-cloud-controller\/0/ { print \$3 }") +export JUMPHOST=$(ifconfig brAdm | awk "/inet addr/ { print \$2 }" | sed 's/addr://g') +export OS_USERNAME=admin +export OS_PASSWORD=openstack +export OS_TENANT_NAME=admin +export OS_AUTH_URL=http://$KEYSTONE_HOST:5000/v2.0 +export OS_REGION_NAME=RegionOne +EOF +else + # Centos: assumes Apex-based install, and that this script is being run on the Undercloud controller VM. + echo "$0: Centos-based install" + echo "$0: Setup undercloud environment so we can get overcloud Controller server address" + source ~/stackrc + echo "$0: Get address of Controller node" + export CONTROLLER_HOST1=$(openstack server list | awk "/overcloud-controller-0/ { print \$8 }" | sed 's/ctlplane=//g') + echo "$0: Create the environment file" + cat <<EOF >$1/admin-openrc.sh +export HORIZON_HOST=$CONTROLLER_HOST1 +export CONGRESS_HOST=$CONTROLLER_HOST1 +export KEYSTONE_HOST=$CONTROLLER_HOST1 +export CEILOMETER_HOST=$CONTROLLER_HOST1 +export CINDER_HOST=$CONTROLLER_HOST1 +export GLANCE_HOST=$CONTROLLER_HOST1 +export NEUTRON_HOST=$CONTROLLER_HOST1 +export NOVA_HOST=$CONTROLLER_HOST1 +export JUMPHOST=$(ip addr show eth0 | grep "inet\b" | awk '{print $2}' | cut -d/ -f1) +EOF + cat ~/overcloudrc >>$1/admin-openrc.sh + source ~/overcloudrc + export OS_REGION_NAME=$(openstack endpoint list | awk "/ nova / { print \$4 }") + # sed command below is a workaound for a bug - region shows up twice for some reason + cat <<EOF | sed '$d' $1/admin-openrc.sh +export OS_REGION_NAME=$OS_REGION_NAME +EOF +fi +source $1/admin-openrc.sh diff --git a/tests/vHello_VES.sh b/tests/vHello_VES.sh new file mode 100644 index 0000000..85120d2 --- /dev/null +++ b/tests/vHello_VES.sh @@ -0,0 +1,322 @@ +#!/bin/bash +# Copyright 2016 AT&T Intellectual Property, Inc +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# +# What this is: Deployment test for the VES agent and collector based +# upon the Tacker Hello World blueprint +# +# Status: this is a work in progress, under test. +# +# How to use: +# $ git clone https://gerrit.opnfv.org/gerrit/ves +# $ cd ves/tests +# $ bash vHello_VES.sh [setup|start|run|stop|clean|collector] +# setup: setup test environment +# start: install blueprint and run test +# run: setup test environment and run test +# stop: stop test and uninstall blueprint +# clean: cleanup after test +# collector: attach to the collector VM and run the collector + +set -x + +trap 'fail' ERR + +pass() { + echo "$0: Hooray!" + set +x #echo off + exit 0 +} + +fail() { + echo "$0: Test Failed!" + set +x + exit 1 +} + +get_floating_net () { + network_ids=($(neutron net-list|grep -v "+"|grep -v name|awk '{print $2}')) + for id in ${network_ids[@]}; do + [[ $(neutron net-show ${id}|grep 'router:external'|grep -i "true") != "" ]] && FLOATING_NETWORK_ID=${id} + done + if [[ $FLOATING_NETWORK_ID ]]; then + FLOATING_NETWORK_NAME=$(openstack network show $FLOATING_NETWORK_ID | awk "/ name / { print \$4 }") + else + echo "$0: Floating network not found" + exit 1 + fi +} + +try () { + count=$1 + $3 + while [[ $? -eq 1 && $count -gt 0 ]] + do + sleep $2 + let count=$count-1 + $3 + done + if [[ $count -eq 0 ]]; then echo "$0: Command \"$3\" was not successful after $1 tries"; fi +} + +setup () { + echo "$0: Setup temp test folder /tmp/tacker and copy this script there" + mkdir -p /tmp/tacker + chmod 777 /tmp/tacker/ + cp $0 /tmp/tacker/. + chmod 755 /tmp/tacker/*.sh + + echo "$0: tacker-setup part 1" + wget https://git.opnfv.org/cgit/models/plain/tests/utils/tacker-setup.sh -O /tmp/tacker/tacker-setup.sh + bash /tmp/tacker/tacker-setup.sh tacker-cli init + + echo "$0: tacker-setup part 2" + CONTAINER=$(sudo docker ps -l | awk "/tacker/ { print \$1 }") + dist=`grep DISTRIB_ID /etc/*-release | awk -F '=' '{print $2}'` + if [ "$dist" == "Ubuntu" ]; then + echo "$0: JOID workaround for Colorado - enable ML2 port security" + juju set neutron-api enable-ml2-port-security=true + + echo "$0: Execute tacker-setup.sh in the container" + sudo docker exec -it $CONTAINER /bin/bash /tmp/tacker/tacker-setup.sh tacker-cli setup + else + echo "$0: Execute tacker-setup.sh in the container" + sudo docker exec -i -t $CONTAINER /bin/bash /tmp/tacker/tacker-setup.sh tacker-cli setup + fi + + echo "$0: reset blueprints folder" + if [[ -d /tmp/tacker/blueprints/tosca-vnfd-hello-ves ]]; then rm -rf /tmp/tacker/blueprints/tosca-vnfd-hello-ves; fi + mkdir -p /tmp/tacker/blueprints/tosca-vnfd-hello-ves + + echo "$0: copy tosca-vnfd-hello-ves to blueprints folder" + cp -r blueprints/tosca-vnfd-hello-ves /tmp/tacker/blueprints + + # Following two steps are in testing still. The guestfish step needs work. + + # echo "$0: Create Nova key pair" + # mkdir -p ~/.ssh + # nova keypair-delete vHello + # nova keypair-add vHello > /tmp/tacker/vHello.pem + # chmod 600 /tmp/tacker/vHello.pem + # pubkey=$(nova keypair-show vHello | grep "Public key:" | sed -- 's/Public key: //g') + # nova keypair-show vHello | grep "Public key:" | sed -- 's/Public key: //g' >/tmp/tacker/vHello.pub + + echo "$0: Inject key into xenial server image" + # wget http://cloud-images.ubuntu.com/xenial/current/xenial-server-cloudimg-amd64-disk1.img + # sudo yum install -y libguestfs-tools + # guestfish <<EOF +#add xenial-server-cloudimg-amd64-disk1.img +#run +#mount /dev/sda1 / +#mkdir /home/ubuntu +#mkdir /home/ubuntu/.ssh +#cat <<EOM >/home/ubuntu/.ssh/authorized_keys +#$pubkey +#EOM +#exit +#chown -R ubuntu /home/ubuntu +#EOF + + # Using pre-key-injected image for now, vHello.pem as provided in the blueprint + if [ ! -f /tmp/xenial-server-cloudimg-amd64-disk1.img ]; then + wget -O /tmp/xenial-server-cloudimg-amd64-disk1.img http://artifacts.opnfv.org/models/images/xenial-server-cloudimg-amd64-disk1.img + fi + cp blueprints/tosca-vnfd-hello-ves/vHello.pem /tmp/tacker + chmod 600 /tmp/tacker/vHello.pem + + echo "$0: setup OpenStack CLI environment" + source /tmp/tacker/admin-openrc.sh + + echo "$0: Setup image_id" + image_id=$(openstack image list | awk "/ models-xenial-server / { print \$2 }") + if [[ -z "$image_id" ]]; then glance --os-image-api-version 1 image-create --name models-xenial-server --disk-format qcow2 --file /tmp/xenial-server-cloudimg-amd64-disk1.img --container-format bare; fi +} + +start() { + echo "$0: setup OpenStack CLI environment" + source /tmp/tacker/admin-openrc.sh + + echo "$0: create VNFD" + cd /tmp/tacker/blueprints/tosca-vnfd-hello-ves + tacker vnfd-create --vnfd-file blueprint.yaml --name hello-ves + if [ $? -eq 1 ]; then fail; fi + + echo "$0: create VNF" + tacker vnf-create --vnfd-name hello-ves --name hello-ves + if [ $? -eq 1 ]; then fail; fi + + echo "$0: wait for hello-ves to go ACTIVE" + active="" + while [[ -z $active ]] + do + active=$(tacker vnf-show hello-ves | grep ACTIVE) + if [ "$(tacker vnf-show hello-ves | grep -c ERROR)" == "1" ]; then + echo "$0: hello-ves VNF creation failed with state ERROR" + fail + fi + sleep 10 + done + + echo "$0: directly set port security on ports (bug/unsupported in Mitaka Tacker?)" + HEAT_ID=$(tacker vnf-show hello-ves | awk "/instance_id/ { print \$4 }") + VDU1_ID=$(openstack stack resource list $HEAT_ID | awk "/VDU1 / { print \$4 }") + id=($(neutron port-list|grep -v "+"|grep -v name|awk '{print $2}')) + for id in ${id[@]}; do + if [[ $(neutron port-show $id|grep $VDU1_ID) ]]; then neutron port-update ${id} --port-security-enabled=True; fi + done + + VDU2_ID=$(openstack stack resource list $HEAT_ID | awk "/VDU2 / { print \$4 }") + id=($(neutron port-list|grep -v "+"|grep -v name|awk '{print $2}')) + for id in ${id[@]}; do + if [[ $(neutron port-show $id|grep $VDU2_ID) ]]; then neutron port-update ${id} --port-security-enabled=True; fi + done + + echo "$0: directly assign security group (unsupported in Mitaka Tacker)" + if [[ $(openstack security group list | awk "/ vHello / { print \$2 }") ]]; then openstack security group delete vHello; fi + openstack security group create vHello + openstack security group rule create --ingress --protocol TCP --dst-port 22:22 vHello + openstack security group rule create --ingress --protocol TCP --dst-port 80:80 vHello + openstack server add security group $VDU1_ID vHello + openstack server add security group $VDU1_ID default + openstack server add security group $VDU2_ID vHello + openstack server add security group $VDU2_ID default + + echo "$0: associate floating IPs" + get_floating_net + FIP=$(openstack floating ip create $FLOATING_NETWORK_NAME | awk "/floating_ip_address/ { print \$4 }") + nova floating-ip-associate $VDU1_ID $FIP + FIP=$(openstack floating ip create $FLOATING_NETWORK_NAME | awk "/floating_ip_address/ { print \$4 }") + nova floating-ip-associate $VDU2_ID $FIP + + echo "$0: get web server addresses" + VDU1_IP=$(openstack server show $VDU1_ID | awk "/ addresses / { print \$6 }") + VDU1_URL="http://$VUD1_IP" + VDU2_IP=$(openstack server show $VDU2_ID | awk "/ addresses / { print \$6 }") + VDU2_URL="http://$VUD2_IP:30000" + + echo "$0: wait 30 seconds for server SSH to be available" + sleep 30 + + echo "$0: Setup the VES Collector in VDU2" + chown root /tmp/tacker/vHello.pem + # Note below: python (2.7) is required due to dependency on module 'ConfigParser' + ssh -i /tmp/tacker/vHello.pem -x -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no ubuntu@$VDU2_IP << EOF +sudo apt-get update +sudo apt-get upgrade -y +sudo apt-get install -y python python-jsonschema +sudo mkdir /var/log/att +sudo chown ubuntu /var/log/att +touch /var/log/att/collector.log +sudo chown ubuntu /home/ubuntu/ +cd /home/ubuntu/ +git clone https://github.com/att/evel-test-collector.git +sed -i -- 's/vel_username = /vel_username = hello/' evel-test-collector/config/collector.conf +sed -i -- 's/vel_password = /vel_password = world/' evel-test-collector/config/collector.conf +nohup python evel-test-collector/code/collector/collector.py \ + --config evel-test-collector/config/collector.conf \ + --section default \ + --verbose > ~/collector.log & +exit +EOF + + echo "$0: start vHello web server in VDU1" + ssh -i /tmp/tacker/vHello.pem -x -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no ubuntu@$VDU1_IP "sudo chown ubuntu /home/ubuntu" + scp -i /tmp/tacker/vHello.pem -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no /tmp/tacker/blueprints/tosca-vnfd-hello-ves/start.sh ubuntu@$VDU1_IP:/home/ubuntu/start.sh + ssh -i /tmp/tacker/vHello.pem -x -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no ubuntu@$VDU1_IP "bash /home/ubuntu/start.sh $VDU2_IP hello:world" + + echo "$0: verify vHello server is running" + apt-get install -y curl + count=10 + while [[ $count -gt 0 ]] + do + sleep 60 + let count=$count-1 + if [[ $(curl http://$VDU1_IP | grep -c "Hello World") == 1 ]]; then pass; fi + done + fail +} + +collector () { + echo "$0: setup OpenStack CLI environment" + source /tmp/tacker/admin-openrc.sh + + echo "$0: find Collector VM IP" + HEAT_ID=$(tacker vnf-show hello-ves | awk "/instance_id/ { print \$4 }") + VDU2_ID=$(openstack stack resource list $HEAT_ID | awk "/VDU2 / { print \$4 }") + VDU2_IP=$(openstack server show $VDU2_ID | awk "/ addresses / { print \$6 }") + VDU2_URL="http://$VUD2_IP:30000" + + echo "$0: Start the VES Collector in VDU2" + ssh -i /tmp/tacker/vHello.pem -x -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no ubuntu@$VDU2_IP << EOF +cd /home/ubuntu/ +python evel-test-collector/code/collector/collector.py \ + --config evel-test-collector/config/collector.conf \ + --section default \ + --verbose +EOF +} + +stop() { + echo "$0: setup OpenStack CLI environment" + source /tmp/tacker/admin-openrc.sh + + echo "$0: uninstall vHello blueprint via CLI" + vid=($(tacker vnf-list|grep hello-ves|awk '{print $2}')); for id in ${vid[@]}; do tacker vnf-delete ${id}; done + vid=($(tacker vnfd-list|grep hello-ves|awk '{print $2}')); for id in ${vid[@]}; do tacker vnfd-delete ${id}; done +# Need to remove the floatingip deletion or make it specific to the vHello VM +# fip=($(neutron floatingip-list|grep -v "+"|grep -v id|awk '{print $2}')); for id in ${fip[@]}; do neutron floatingip-delete ${id}; done + sg=($(openstack security group list|grep vHello|awk '{print $2}')) + for id in ${sg[@]}; do try 5 5 "openstack security group delete ${id}"; done +} + +forward_to_container () { + echo "$0: pass $1 command to this script in the tacker container" + CONTAINER=$(sudo docker ps -a | awk "/tacker/ { print \$1 }") + sudo docker exec $CONTAINER /bin/bash /tmp/tacker/vHello_VES.sh $1 $1 + if [ $? -eq 1 ]; then fail; fi +} + +dist=`grep DISTRIB_ID /etc/*-release | awk -F '=' '{print $2}'` +case "$1" in + setup) + setup + pass + ;; + run) + setup + forward_to_container start + pass + ;; + start|stop|collector) + if [[ $# -eq 1 ]]; then forward_to_container $1 + else + # running inside the tacker container, ready to go + $1 + fi + pass + ;; + clean) + echo "$0: Uninstall Tacker and test environment" + bash /tmp/tacker/tacker-setup.sh $1 clean + pass + ;; + *) + echo "usage: bash vHello_VES.sh [setup|start|run|clean]" + echo "setup: setup test environment" + echo "start: install blueprint and run test" + echo "run: setup test environment and run test" + echo "stop: stop test and uninstall blueprint" + echo "clean: cleanup after test" + fail +esac diff --git a/tests/vLamp_Ansible_VES.sh b/tests/vLamp_Ansible_VES.sh new file mode 100644 index 0000000..1ae6fdc --- /dev/null +++ b/tests/vLamp_Ansible_VES.sh @@ -0,0 +1,46 @@ +#!/bin/bash
+# Copyright 2016 AT&T Intellectual Property, Inc
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+#
+# What this is: Enhancements to the OpenStack Interop Challenge "Lampstack"
+# blueprint to add OPNFV VES event capture.
+#
+# Status: this is a work in progress, under test.
+#
+# How to use:
+# $ bash vLamp_Ansible_VES.sh
+
+echo "$0: Add ssh key"
+eval $(ssh-agent -s)
+ssh-add /tmp/ansible/ansible
+
+echo "$0: setup OpenStack environment"
+source /tmp/ansible/admin-openrc.sh
+
+$BALANCER=$(openstack server show balancer | awk "/ addresses / { print \$6 }")
+sudo cp /tmp/ansible/ansible /tmp/ansible/lampstack
+sudo chown $USER /tmp/ansible/lampstack
+ssh -i /tmp/ansible/lampstack ubuntu@$BALANCER
+
+# scp -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no ~/congress/env.sh $CTLUSER@$CONTROLLER_HOST1:/home/$CTLUSER/congress
+
+echo "$0: Enable haproxy logging"
+# Example /var/log/haproxy.log entries after logging enabled
+# Oct 6 20:03:34 balancer haproxy[2075]: 192.168.37.199:36193 [06/Oct/2016:20:03:34.349] webfarm webfarm/ws10.0.0.9 107/0/1/1/274 304 144 - - ---- 1/1/1/0/0 0/0 "GET /wp-content/themes/iribbon/elements/lib/images/boxes/slidericon.png HTTP/1.1"
+# Oct 6 20:03:34 balancer haproxy[2075]: 192.168.37.199:36194 [06/Oct/2016:20:03:34.365] webfarm webfarm/ws10.0.0.10 95/0/0/1/258 304 144 - - ---- 0/0/0/0/0 0/0 "GET /wp-content/themes/iribbon/elements/lib/images/boxes/blueprint.png HTTP/1.1"
+ssh -x -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no ubuntu@$BALANCER <<EOF
+sudo sed -i -- 's/#$ModLoad imudp/$ModLoad imudp/g' /etc/rsyslog.conf
+sudo sed -i -- 's/#$UDPServerRun 514/$UDPServerRun 514\n$UDPServerAddress 127.0.0.1/g' /etc/rsyslog.conf
+sudo service rsyslog restart
+EOF
|