summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--jjb/apex/apex.yml11
-rw-r--r--jjb/compass4nfv/compass-ci-jobs.yml6
-rw-r--r--jjb/compass4nfv/compass-deploy.sh26
-rw-r--r--jjb/compass4nfv/compass-project-jobs.yml12
-rw-r--r--jjb/doctor/doctor.yml2
-rw-r--r--jjb/fuel/fuel-ci-jobs.yml200
-rwxr-xr-xjjb/fuel/fuel-download-artifact.sh2
-rwxr-xr-xjjb/fuel/fuel-upload-artifact.sh4
-rw-r--r--jjb/functest/functest-ci-jobs.yml3
-rw-r--r--jjb/opnfv/installer-params.yml2
-rw-r--r--jjb/opnfv/slave-params.yml42
-rw-r--r--jjb/qtip/qtip-ci-jobs.yml (renamed from jjb/qtip/qtip.yml)158
-rw-r--r--jjb/qtip/qtip-project-jobs.yml58
-rw-r--r--jjb/releng-macros.yaml9
-rw-r--r--jjb/yardstick/yardstick-ci-jobs.yml9
-rw-r--r--jjb/yardstick/yardstick-project-jobs.yml3
-rwxr-xr-xutils/docs-build.sh2
-rwxr-xr-xutils/fetch_os_creds.sh2
-rw-r--r--utils/push-test-logs.sh4
-rw-r--r--utils/test/result_collection_api/dashboard/dashboard_utils.py2
-rw-r--r--utils/test/result_collection_api/dashboard/functest2Dashboard.py88
-rw-r--r--utils/test/result_collection_api/dashboard/promise2Dashboard.py103
22 files changed, 565 insertions, 183 deletions
diff --git a/jjb/apex/apex.yml b/jjb/apex/apex.yml
index 785056ff0..1ab7f647f 100644
--- a/jjb/apex/apex.yml
+++ b/jjb/apex/apex.yml
@@ -84,9 +84,6 @@
pattern: 'lib/**'
- compare-type: ANT
pattern: 'config/**'
- forbidden-file-paths:
- - compare-type: ANT
- pattern: 'docs/**'
properties:
- build-blocker:
@@ -160,16 +157,16 @@
git-revision: false
block: true
- trigger-builds:
- - project: 'functest-apex-opnfv-jump-1-daily-{stream}'
- predefined-parameters:
- DEPLOY_SCENARIO={scenario}
+ - project: 'yardstick-apex-opnfv-jump-1-daily-{stream}'
block: true
block-thresholds:
build-step-failure-threshold: 'never'
failure-threshold: 'never'
unstable-threshold: 'FAILURE'
- trigger-builds:
- - project: 'yardstick-apex-opnfv-jump-1-daily-{stream}'
+ - project: 'functest-apex-opnfv-jump-1-daily-{stream}'
+ predefined-parameters:
+ DEPLOY_SCENARIO={scenario}
block: true
block-thresholds:
build-step-failure-threshold: 'never'
diff --git a/jjb/compass4nfv/compass-ci-jobs.yml b/jjb/compass4nfv/compass-ci-jobs.yml
index a4be24ec0..5564d6989 100644
--- a/jjb/compass4nfv/compass-ci-jobs.yml
+++ b/jjb/compass4nfv/compass-ci-jobs.yml
@@ -168,11 +168,9 @@
builders:
- shell:
- !include-raw: ./compass-workspace-precleanup.sh
+ !include-raw-escape: ./compass-download-artifact.sh
- shell:
- !include-raw: ./compass-download-artifact.sh
- - shell:
- !include-raw: ./compass-deploy.sh
+ !include-raw-escape: ./compass-deploy.sh
publishers:
- archive:
diff --git a/jjb/compass4nfv/compass-deploy.sh b/jjb/compass4nfv/compass-deploy.sh
index 697a1d573..860f5da1b 100644
--- a/jjb/compass4nfv/compass-deploy.sh
+++ b/jjb/compass4nfv/compass-deploy.sh
@@ -6,24 +6,38 @@ echo "Starting the deployment on baremetal environment using $INSTALLER_TYPE. Th
echo "--------------------------------------------------------"
echo
+# source the properties file so we get OPNFV vars
+source $BUILD_DIRECTORY/latest.properties
+
+# echo the info about artifact that is used during the deployment
+echo "Using ${OPNFV_ARTIFACT_URL/*\/} for deployment"
+
+if [[ ! "$JOB_NAME" =~ (verify|merge) ]]; then
+ # for none-merge deployments
+ # checkout the commit that was used for building the downloaded artifact
+ # to make sure the ISO and deployment mechanism uses same versions
+ echo "Checking out $OPNFV_GIT_SHA1"
+ git checkout $OPNFV_GIT_SHA1 --quiet
+fi
+
echo 1 > /proc/sys/vm/drop_caches
export CONFDIR=$WORKSPACE/deploy/conf
export ISO_URL=file://$BUILD_DIRECTORY/compass.iso
if [[ "$NODE_NAME" =~ "-vm" ]]; then
export NETWORK_CONF=$CONFDIR/vm_environment/$NODE_NAME/network.yml
- export DHA_CONF=$CONFDIR/vm_environment/${{DEPLOY_SCENARIO}}.yml
+ export DHA_CONF=$CONFDIR/vm_environment/${DEPLOY_SCENARIO}.yml
else
export INSTALL_NIC=eth1
export NETWORK_CONF=$CONFDIR/hardware_environment/$NODE_NAME/network.yml
- export DHA_CONF=$CONFDIR/hardware_environment/$NODE_NAME/${{DEPLOY_SCENARIO}}.yml
+ export DHA_CONF=$CONFDIR/hardware_environment/$NODE_NAME/${DEPLOY_SCENARIO}.yml
fi
cd $WORKSPACE
-export OS_VERSION=${{COMPASS_OS_VERSION}}
-export OPENSTACK_VERSION=${{COMPASS_OPENSTACK_VERSION}}
-./deploy.sh --dha ${{DHA_CONF}} --network ${{NETWORK_CONF}}
+export OS_VERSION=${COMPASS_OS_VERSION}
+export OPENSTACK_VERSION=${COMPASS_OPENSTACK_VERSION}
+./deploy.sh --dha ${DHA_CONF} --network ${NETWORK_CONF}
if [ $? -ne 0 ]; then
echo "depolyment failed!"
deploy_ret=1
@@ -34,6 +48,6 @@ echo "--------------------------------------------------------"
echo "Done!"
ssh_options="-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no"
-sshpass -p root scp 2>/dev/null $ssh_options root@${{INSTALLER_IP}}:/var/ansible/run/openstack_${{COMPASS_OPENSTACK_VERSION}}-opnfv2/ansible.log ./ &> /dev/null
+sshpass -p root scp 2>/dev/null $ssh_options root@${INSTALLER_IP}:/var/ansible/run/openstack_${COMPASS_OPENSTACK_VERSION}-opnfv2/ansible.log ./ &> /dev/null
exit $deploy_ret
diff --git a/jjb/compass4nfv/compass-project-jobs.yml b/jjb/compass4nfv/compass-project-jobs.yml
index 8943afb37..11e1d5f8a 100644
--- a/jjb/compass4nfv/compass-project-jobs.yml
+++ b/jjb/compass4nfv/compass-project-jobs.yml
@@ -116,9 +116,9 @@
builders:
- shell:
- !include-raw: ./compass-build.sh
+ !include-raw-escape: ./compass-build.sh
- shell:
- !include-raw: ./compass-deploy.sh
+ !include-raw-escape: ./compass-deploy.sh
publishers:
- archive:
@@ -157,11 +157,11 @@
builders:
- shell:
- !include-raw: ./compass-build.sh
+ !include-raw-escape: ./compass-build.sh
- shell:
- !include-raw: ./compass-upload-artifact.sh
+ !include-raw-escape: ./compass-upload-artifact.sh
- shell:
- !include-raw: ./compass-workspace-cleanup.sh
+ !include-raw-escape: ./compass-workspace-cleanup.sh
- job-template:
name: 'compass-build-ppa-{stream}'
@@ -194,7 +194,7 @@
builders:
- shell:
- !include-raw: ./compass-makeppa.sh
+ !include-raw-escape: ./compass-makeppa.sh
diff --git a/jjb/doctor/doctor.yml b/jjb/doctor/doctor.yml
index 167b696a3..b2cbb0242 100644
--- a/jjb/doctor/doctor.yml
+++ b/jjb/doctor/doctor.yml
@@ -53,4 +53,4 @@
pattern: 'tests/**'
builders:
- - shell: "[ -e tests/run.sh ] && ./tests/run.sh"
+ - shell: "[ -e tests/run.sh ] && bash -n ./tests/run.sh"
diff --git a/jjb/fuel/fuel-ci-jobs.yml b/jjb/fuel/fuel-ci-jobs.yml
index 33c54eced..6b69afe26 100644
--- a/jjb/fuel/fuel-ci-jobs.yml
+++ b/jjb/fuel/fuel-ci-jobs.yml
@@ -35,6 +35,8 @@
#--------------------------------
# master
#--------------------------------
+ - opnfv-jump-2:
+ <<: *master
- ericsson-pod1:
<<: *master
- ericsson-pod2:
@@ -46,23 +48,22 @@
#--------------------------------
scenario:
# HA scenarios
- - 'os-onos-nofeature-ha':
-# auto-trigger-name: 'fuel-{scenario}-{pod}-trigger'
- auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
- 'os-odl_l2-nofeature-ha':
- auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
+ auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
- 'os-odl_l3-nofeature-ha':
- auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
- - 'os-odl_l2-bgpvpn-ha':
- auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
- - 'os-nosdn-nofeature-ha':
- auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
+ auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
+ - 'os-onos-nofeature-ha':
+ auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
- 'os-nosdn-kvm-ha':
+ auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
+ - 'os-nosdn-ovs-ha':
+ auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
+ - 'os-nosdn-nofeature-ha':
+ auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger'
+ - 'os-odl_l2-bgpvpn-ha':
auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
- 'os-nosdn-kvm_ovs-ha':
auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
- - 'os-nosdn-ovs-ha':
- auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
- 'os-nosdn-vlan-ha':
auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
# NOHA scenarios
@@ -226,74 +227,207 @@
########################
# trigger macros
########################
-# os-onos3-nofeature-ha trigger
+# os-nosdn-nofeature-ha trigger
+- trigger:
+ name: 'fuel-os-nosdn-nofeature-ha-opnfv-jump-2-master-trigger'
+ triggers:
+ - timed: '0 14 * * *'
+- trigger:
+ name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod1-master-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod2-master-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-nofeature-ha-virtual-master-trigger'
+ triggers:
+ - timed: ''
- trigger:
- name: 'fuel-os-onos-nofeature-ha-opnfv-jump-2-trigger'
+ name: 'fuel-os-nosdn-nofeature-ha-opnfv-jump-2-brahmaputra-trigger'
triggers:
- - timed: '0 20 * * *'
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod1-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod2-brahmaputra-trigger'
+ triggers:
+ - timed: '0 14 * * *'
+- trigger:
+ name: 'fuel-os-nosdn-nofeature-ha-virtual-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+
+# os-nosdn-kvm-ha trigger
- trigger:
- name: 'fuel-os-onos-nofeature-ha-ericsson-pod1-trigger'
+ name: 'fuel-os-nosdn-kvm-ha-opnfv-jump-2-master-trigger'
+ triggers:
+ - timed: '0 18 * * *'
+- trigger:
+ name: 'fuel-os-nosdn-kvm-ha-ericsson-pod1-master-trigger'
triggers:
- timed: ''
- trigger:
- name: 'fuel-os-onos-nofeature-ha-ericsson-pod2-trigger'
+ name: 'fuel-os-nosdn-kvm-ha-ericsson-pod2-master-trigger'
triggers:
- timed: ''
- trigger:
- name: 'fuel-os-onos-nofeature-ha-virtual-trigger'
+ name: 'fuel-os-nosdn-kvm-ha-virtual-master-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-kvm-ha-opnfv-jump-2-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-kvm-ha-ericsson-pod1-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-kvm-ha-ericsson-pod2-brahmaputra-trigger'
+ triggers:
+ - timed: '0 18 * * *'
+- trigger:
+ name: 'fuel-os-nosdn-kvm-ha-virtual-brahmaputra-trigger'
triggers:
- timed: ''
# os-odl_l2-nofeature-ha trigger
- trigger:
- name: 'fuel-os-odl_l2-nofeature-ha-opnfv-jump-2-trigger'
+ name: 'fuel-os-odl_l2-nofeature-ha-opnfv-jump-2-master-trigger'
+ triggers:
+ - timed: '0 2 * * *'
+- trigger:
+ name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod1-master-trigger'
triggers:
- - timed: '0 0 * * *'
+ - timed: ''
- trigger:
- name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod1-trigger'
+ name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod2-master-trigger'
triggers:
- timed: ''
- trigger:
- name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod2-trigger'
+ name: 'fuel-os-odl_l2-nofeature-ha-virtual-master-trigger'
triggers:
- timed: ''
- trigger:
- name: 'fuel-os-odl_l2-nofeature-ha-virtual-trigger'
+ name: 'fuel-os-odl_l2-nofeature-ha-opnfv-jump-2-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod1-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod2-brahmaputra-trigger'
+ triggers:
+ - timed: '0 2 * * *'
+- trigger:
+ name: 'fuel-os-odl_l2-nofeature-ha-virtual-brahmaputra-trigger'
triggers:
- timed: ''
# os-odl_l3-nofeature-ha trigger
- trigger:
- name: 'fuel-os-odl_l3-nofeature-ha-opnfv-jump-2-trigger'
+ name: 'fuel-os-odl_l3-nofeature-ha-opnfv-jump-2-master-trigger'
triggers:
- - timed: '0 4 * * *'
+ - timed: '0 6 * * *'
- trigger:
- name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod1-trigger'
+ name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod1-master-trigger'
triggers:
- timed: ''
- trigger:
- name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod2-trigger'
+ name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod2-master-trigger'
triggers:
- timed: ''
- trigger:
- name: 'fuel-os-odl_l3-nofeature-ha-virtual-trigger'
+ name: 'fuel-os-odl_l3-nofeature-ha-virtual-master-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-odl_l3-nofeature-ha-opnfv-jump-2-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod1-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod2-brahmaputra-trigger'
+ triggers:
+ - timed: '0 6 * * *'
+- trigger:
+ name: 'fuel-os-odl_l3-nofeature-ha-virtual-brahmaputra-trigger'
triggers:
- timed: ''
-# os-odl_l2-bgpvpn-ha trigger
+# os-onos-nofeature-ha trigger
- trigger:
- name: 'fuel-os-odl_l2-bgpvpn-ha-opnfv-jump-2-trigger'
+ name: 'fuel-os-onos-nofeature-ha-opnfv-jump-2-master-trigger'
triggers:
- - timed: '0 8 * * *'
+ - timed: '0 10 * * *'
+- trigger:
+ name: 'fuel-os-onos-nofeature-ha-ericsson-pod1-master-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-onos-nofeature-ha-ericsson-pod2-master-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-onos-nofeature-ha-virtual-master-trigger'
+ triggers:
+ - timed: ''
- trigger:
- name: 'fuel-os-odl_l2-bgpvpn-ha-ericsson-pod1-trigger'
+ name: 'fuel-os-onos-nofeature-ha-opnfv-jump-2-brahmaputra-trigger'
triggers:
- timed: ''
- trigger:
- name: 'fuel-os-odl_l2-bgpvpn-ha-ericsson-pod2-trigger'
+ name: 'fuel-os-onos-nofeature-ha-ericsson-pod1-brahmaputra-trigger'
triggers:
- timed: ''
- trigger:
- name: 'fuel-os-odl_l2-bgpvpn-ha-virtual-trigger'
+ name: 'fuel-os-onos-nofeature-ha-ericsson-pod2-brahmaputra-trigger'
+ triggers:
+ - timed: '0 10 * * *'
+- trigger:
+ name: 'fuel-os-onos-nofeature-ha-virtual-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+
+
+# os-nosdn-ovs-ha trigger
+- trigger:
+ name: 'fuel-os-nosdn-ovs-ha-opnfv-jump-2-master-trigger'
+ triggers:
+ - timed: '0 22 * * *'
+- trigger:
+ name: 'fuel-os-nosdn-ovs-ha-ericsson-pod1-master-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-ovs-ha-ericsson-pod2-master-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-ovs-ha-virtual-master-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-ovs-ha-opnfv-jump-2-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-ovs-ha-ericsson-pod1-brahmaputra-trigger'
+ triggers:
+ - timed: ''
+- trigger:
+ name: 'fuel-os-nosdn-ovs-ha-ericsson-pod2-brahmaputra-trigger'
+ triggers:
+ - timed: '0 22 * * *'
+- trigger:
+ name: 'fuel-os-nosdn-ovs-ha-virtual-brahmaputra-trigger'
triggers:
- timed: ''
diff --git a/jjb/fuel/fuel-download-artifact.sh b/jjb/fuel/fuel-download-artifact.sh
index ceac859dd..d78ddc5ba 100755
--- a/jjb/fuel/fuel-download-artifact.sh
+++ b/jjb/fuel/fuel-download-artifact.sh
@@ -36,7 +36,7 @@ if [[ ! "$JOB_NAME" =~ (verify|merge) ]]; then
ISOSTORE="/iso_mount/opnfv_ci/${GIT_BRANCH##*/}"
if [[ -f "$ISOSTORE/$OPNFV_ARTIFACT" ]]; then
echo "ISO exists locally. Skipping the download and using the file from ISO store"
- /bin/cp -f $ISOSTORE/$OPNFV_ARTIFACT $WORKSPACE/opnfv.iso
+ ln -s $ISOSTORE/$OPNFV_ARTIFACT $WORKSPACE/opnfv.iso
echo "--------------------------------------------------------"
echo
ls -al $WORKSPACE/opnfv.iso
diff --git a/jjb/fuel/fuel-upload-artifact.sh b/jjb/fuel/fuel-upload-artifact.sh
index 6dd5f34d6..47cf8985c 100755
--- a/jjb/fuel/fuel-upload-artifact.sh
+++ b/jjb/fuel/fuel-upload-artifact.sh
@@ -29,17 +29,17 @@ if [[ ! "$JOB_NAME" =~ (verify|merge) ]]; then
ls -tp | grep -v '/' | tail -n +6 | xargs -d '\n' /bin/rm -f --
# store ISO
- echo "Storing $INSTALLER_TYPE artifact on NFS first"
+ echo "Storing $INSTALLER_TYPE artifact on NFS..."
/bin/cp -f $BUILD_DIRECTORY/opnfv-$OPNFV_ARTIFACT_VERSION.iso \
$ISOSTORE/opnfv-$OPNFV_ARTIFACT_VERSION.iso
fi
fi
# log info to console
-echo
echo "Uploading $INSTALLER_TYPE artifact. This could take some time..."
echo
+cd $WORKSPACE
# upload artifact and additional files to google storage
gsutil cp $BUILD_DIRECTORY/opnfv-$OPNFV_ARTIFACT_VERSION.iso \
gs://$GS_URL/opnfv-$OPNFV_ARTIFACT_VERSION.iso > gsutil.iso.log 2>&1
diff --git a/jjb/functest/functest-ci-jobs.yml b/jjb/functest/functest-ci-jobs.yml
index de0592209..1165b2a2c 100644
--- a/jjb/functest/functest-ci-jobs.yml
+++ b/jjb/functest/functest-ci-jobs.yml
@@ -53,6 +53,9 @@
#--------------------------------
# master
#--------------------------------
+ - opnfv-jump-2:
+ installer: fuel
+ <<: *master
- intel-pod6:
installer: joid
<<: *master
diff --git a/jjb/opnfv/installer-params.yml b/jjb/opnfv/installer-params.yml
index e03b4749e..57ca9864e 100644
--- a/jjb/opnfv/installer-params.yml
+++ b/jjb/opnfv/installer-params.yml
@@ -47,7 +47,7 @@
description: 'Installer used for deploying OPNFV on this POD'
- string:
name: EXTERNAL_NETWORK
- default: 'net04_ext'
+ default: 'admin_floating_net'
description: 'external network for test'
- parameter:
diff --git a/jjb/opnfv/slave-params.yml b/jjb/opnfv/slave-params.yml
index b2c6e7cf5..5e0274d84 100644
--- a/jjb/opnfv/slave-params.yml
+++ b/jjb/opnfv/slave-params.yml
@@ -81,10 +81,6 @@
name: GIT_BASE
default: https://gerrit.opnfv.org/gerrit/$PROJECT
description: 'Git URL to use on this Jenkins Slave'
- - string:
- name: EXTERNAL_NETWORK
- default: 'admin_floating_net'
- description: 'external network'
- parameter:
name: 'intelpod2-jumphost-defaults'
@@ -424,3 +420,41 @@
name: SSH_KEY
default: /root/.ssh/id_rsa
description: 'SSH key to use for Apex'
+
+- parameter:
+ name: 'dell-us-testing-bm-1-defaults'
+ parameters:
+ - node:
+ name: SLAVE_NAME
+ description: 'Slave name on Jenkins'
+ allowed-slaves:
+ - dell-us-testing-bm-1
+ default-slaves:
+ - dell-us-testing-bm-1
+ - string:
+ name: INSTALLER_VERSION
+ default: latest
+ description: 'Version of the installer to deploy'
+ - string:
+ name: GIT_BASE
+ default: https://gerrit.opnfv.org/gerrit/$PROJECT
+ description: 'Git URL to use on this Jenkins Slave'
+
+- parameter:
+ name: 'dell-us-deploying-bm3-defaults'
+ parameters:
+ - node:
+ name: SLAVE_NAME
+ description: 'Slave name on Jenkins'
+ allowed-slaves:
+ - dell-us-deploying-bm3
+ default-slaves:
+ - dell-us-deploying-bm3
+ - string:
+ name: INSTALLER_VERSION
+ default: latest
+ description: 'Version of the installer to deploy'
+ - string:
+ name: GIT_BASE
+ default: https://gerrit.opnfv.org/gerrit/$PROJECT
+ description: 'Git URL to use on this Jenkins Slave'
diff --git a/jjb/qtip/qtip.yml b/jjb/qtip/qtip-ci-jobs.yml
index 450ba938e..d33bdda32 100644
--- a/jjb/qtip/qtip.yml
+++ b/jjb/qtip/qtip-ci-jobs.yml
@@ -1,105 +1,74 @@
+####################################
+# job configuration for qtip
+####################################
- project:
- name: qtip
-
- project: '{name}'
-
- jobs:
- - 'qtip-verify-{stream}'
- - 'qtip-{installer_type}-{pod}-daily-{stream}'
-
+ name: qtip-ci-jobs
+
+ project: 'qtip'
+
+#--------------------------------
+# BRANCH ANCHORS
+#--------------------------------
+ master: &master
+ stream: master
+ branch: '{stream}'
+ gs-pathname: ''
+ brahmaputra: &brahmaputra
+ stream: brahmaputra
+ branch: 'stable/{stream}'
+ gs-pathname: '{stream}'
+#--------------------------------
+# POD, INSTALLER, AND BRANCH MAPPING
+#--------------------------------
+# brahmaputra
+#--------------------------------
pod:
- dell-us-testing-bm-1:
- node: 'dell-us-testing-bm-1'
- installer_type: 'fuel'
- installer_ip: '10.20.0.2'
+ installer: compass
+ <<: *brahmaputra
+ - orange-pod2:
+ installer: joid
+ <<: *brahmaputra
- dell-us-deploying-bm3:
- node: 'dell-us-deploying-bm3'
- installer_type: 'fuel'
- installer_ip: '10.20.0.2'
-
-# only master branch is enabled at the moment to keep no of jobs sane
- stream:
- - master:
- branch: 'master'
- gs-pathname: ''
-# - brahmaputra:
-# branch: 'stable/brahmaputra'
-# gs-pathname: '/brahmaputra'
-
-- job-template:
- name: 'qtip-verify-{stream}'
-
- parameters:
- - project-parameter:
- project: '{project}'
- - gerrit-parameter:
- branch: '{branch}'
- - 'opnfv-build-defaults'
-
- scm:
- - gerrit-trigger-scm:
- credentials-id: '{ssh-credentials}'
- refspec: '$GERRIT_REFSPEC'
- choosing-strategy: 'gerrit'
-
- triggers:
- - gerrit:
- trigger-on:
- - patchset-created-event:
- exclude-drafts: 'false'
- exclude-trivial-rebase: 'false'
- exclude-no-code-change: 'false'
- - draft-published-event
- - comment-added-contains-event:
- comment-contains-value: 'recheck'
- - comment-added-contains-event:
- comment-contains-value: 'reverify'
- projects:
- - project-compare-type: 'ANT'
- project-pattern: '{project}'
- branches:
- - branch-compare-type: 'ANT'
- branch-pattern: '**/{branch}'
- forbidden-file-paths:
- - compare-type: ANT
- pattern: 'docs/**|.gitignore'
-
- builders:
- - shell: |
- echo "Nothing to verify!"
+ installer: fuel
+ <<: *brahmaputra
+
+#--------------------------------
+# master
+#--------------------------------
+ - juniper-pod1:
+ installer: joid
+ <<: *master
+# - zte-build-1: #would be confirmed with the ZTE lab by tomorrow
+# installer: fuel
+# <<: *master
+
+#--------------------------------
+ jobs:
+ - 'qtip-{installer}-{pod}-daily-{stream}'
+################################
+# job templates
+################################
- job-template:
- name: 'qtip-{installer_type}-{pod}-daily-{stream}'
+ name: 'qtip-{installer}-{pod}-daily-{stream}'
disabled: false
- node: '{node}'
-
parameters:
- project-parameter:
project: '{project}'
+ - '{pod}-defaults'
+ - '{installer}-defaults'
- string:
- name: INSTALLER_TYPE
- default: '{installer_type}'
- description: "Installer name that is used for deployment."
- - string:
- name: INSTALLER_IP
- default: '{installer_ip}'
- description: "Installer IP"
- - string:
- name: NODE_NAME
- default: '{node}'
- description: "Name of the POD"
- - string:
- name: GIT_BASE
- default: https://gerrit.opnfv.org/gerrit/$PROJECT
- description: "Used for overriding the GIT URL coming from Global Jenkins configuration in case if the stuff is done on none-LF HW."
+ name: DEPLOY_SCENARIO
+ default: 'os-nosdn-nofeature-ha'
scm:
- git-scm:
credentials-id: '{ssh-credentials}'
refspec: ''
- branch: master
+ branch: '{branch}'
builders:
- 'qtip-cleanup'
@@ -107,9 +76,9 @@
- 'qtip-run-suite'
- 'qtip-pushtoDB'
- triggers:
- - timed: '0 0,13,15,20 * * *'
-
+ publishers:
+ - email:
+ recipients: nauman.ahad@xflowresearch.com, mofassir.arif@xflowresearch.com, vikram@nvirters.com
###########################
#biuilder macros
@@ -123,7 +92,6 @@
envs="INSTALLER_TYPE=${INSTALLER_TYPE} -e INSTALLER_IP=${INSTALLER_IP} -e NODE_NAME=${NODE_NAME}"
suite="TEST_CASE=all"
docker pull opnfv/qtip:latest
-
cmd=" docker run -id -e $envs -e $suite opnfv/qtip:latest /bin/bash"
echo "Qtip: Running docker run command: ${cmd}"
${cmd}
@@ -190,17 +158,3 @@
docker rmi opnfv/qtip:$tag
done
fi
-
-- builder:
- name: qtip-fetch-os-cred
- builders:
- - shell: |
- #!/bin/bash
-
- echo $HOME
- echo $WORKSPACE
- cd $WORKSPACE
- git clone https://gerrit.opnfv.org/gerrit/releng
- cd $WORKSPACE/releng/utils
- #./fetch_os_creds.sh #To test the problem with creating the file in $HOME
- ./fetch_os_creds.sh -d $WORKSPACE/opnfv-stack.sh #To test if the file can be created in $WORKSPACE
diff --git a/jjb/qtip/qtip-project-jobs.yml b/jjb/qtip/qtip-project-jobs.yml
new file mode 100644
index 000000000..c77bac722
--- /dev/null
+++ b/jjb/qtip/qtip-project-jobs.yml
@@ -0,0 +1,58 @@
+- project:
+ name: qtip
+
+ project: '{name}'
+
+ jobs:
+ - 'qtip-verify-{stream}'
+
+# only master branch is enabled at the moment to keep no of jobs sane
+ stream:
+ - master:
+ branch: '{stream}'
+ gs-pathname: ''
+ - brahmaputra:
+ branch: 'stable/{stream}'
+ gs-pathname: '/{stream}'
+
+- job-template:
+ name: 'qtip-verify-{stream}'
+
+ parameters:
+ - project-parameter:
+ project: '{project}'
+ - gerrit-parameter:
+ branch: '{branch}'
+ - 'opnfv-build-defaults'
+
+ scm:
+ - gerrit-trigger-scm:
+ credentials-id: '{ssh-credentials}'
+ refspec: '$GERRIT_REFSPEC'
+ choosing-strategy: 'gerrit'
+
+ triggers:
+ - gerrit:
+ trigger-on:
+ - patchset-created-event:
+ exclude-drafts: 'false'
+ exclude-trivial-rebase: 'false'
+ exclude-no-code-change: 'false'
+ - draft-published-event
+ - comment-added-contains-event:
+ comment-contains-value: 'recheck'
+ - comment-added-contains-event:
+ comment-contains-value: 'reverify'
+ projects:
+ - project-compare-type: 'ANT'
+ project-pattern: '{project}'
+ branches:
+ - branch-compare-type: 'ANT'
+ branch-pattern: '**/{branch}'
+ forbidden-file-paths:
+ - compare-type: ANT
+ pattern: 'docs/**|.gitignore'
+
+ builders:
+ - shell: |
+ echo "Nothing to verify!"
diff --git a/jjb/releng-macros.yaml b/jjb/releng-macros.yaml
index 5d9570705..b43a76d5f 100644
--- a/jjb/releng-macros.yaml
+++ b/jjb/releng-macros.yaml
@@ -195,7 +195,7 @@
echo "###########################"
echo
- gs_base="artifacts.opnfv.org/review"
+ gs_base="artifacts.opnfv.org/$PROJECT/review"
gs_path="$gs_base/$GERRIT_CHANGE_NUMBER"
local_path="upload/$GERRIT_CHANGE_NUMBER"
@@ -283,6 +283,13 @@
[[ $GERRIT_CHANGE_NUMBER =~ .+ ]]
+ gs_path="artifacts.opnfv.org/$PROJECT/review/$GERRIT_CHANGE_NUMBER"
+
+ if gsutil ls "gs://$gs_path" > /dev/null 2>&1 ; then
+ echo
+ echo "Deleting Out-of-dated Documents..."
+ gsutil -m rm -r "gs://$gs_path"
+ fi
gs_path="artifacts.opnfv.org/review/$GERRIT_CHANGE_NUMBER"
if gsutil ls "gs://$gs_path" > /dev/null 2>&1 ; then
diff --git a/jjb/yardstick/yardstick-ci-jobs.yml b/jjb/yardstick/yardstick-ci-jobs.yml
index ec5c547ee..e7221e70c 100644
--- a/jjb/yardstick/yardstick-ci-jobs.yml
+++ b/jjb/yardstick/yardstick-ci-jobs.yml
@@ -71,6 +71,11 @@
#--------------------------------
# master
#--------------------------------
+ - opnfv-jump-2:
+ installer: fuel
+ suite: daily
+ auto-trigger-name: 'brahmaputra-trigger-daily-disabled'
+ <<: *master
- ericsson-pod1:
installer: fuel
suite: daily
@@ -328,7 +333,7 @@
parameters:
- string:
name: YARDSTICK_DB_BACKEND
- default: '-r 213.77.62.197/results'
+ default: '-i 10.118.36.90:8086'
description: 'Arguments to use in order to choose the backend DB'
- parameter:
@@ -336,7 +341,7 @@
parameters:
- string:
name: YARDSTICK_DB_BACKEND
- default: '-r 213.77.62.197/results'
+ default: '-i 10.118.36.90:8086'
description: 'Arguments to use in order to choose the backend DB'
- parameter:
diff --git a/jjb/yardstick/yardstick-project-jobs.yml b/jjb/yardstick/yardstick-project-jobs.yml
index 749ebf574..f8f1abbff 100644
--- a/jjb/yardstick/yardstick-project-jobs.yml
+++ b/jjb/yardstick/yardstick-project-jobs.yml
@@ -120,12 +120,13 @@
# install python packages
easy_install -U setuptools
- pip install Sphinx==1.3.1 doc8 docutils sphinxcontrib-httpdomain
python setup.py develop
# unit tests
./run_tests.sh
+ pip install Sphinx==1.3.1 doc8 docutils sphinxcontrib-httpdomain
+
# generate api doc sources
sphinx-apidoc -o docs/apidocs yardstick
diff --git a/utils/docs-build.sh b/utils/docs-build.sh
index e5d6b4a7f..3c0e8bd80 100755
--- a/utils/docs-build.sh
+++ b/utils/docs-build.sh
@@ -41,7 +41,7 @@ revision="$(git rev-parse --short HEAD)"
rev_full="$(git rev-parse HEAD)"
version="$(git describe --abbrev=0 2> /dev/null || echo draft) ($revision)"
project="$(basename $(git rev-parse --show-toplevel))"
-html_notes="\n Revision: $rev_full\n\n Build date: |today|"
+html_notes=" Revision: $rev_full\n Build date: |today|"
default_conf='releng/docs/etc/conf.py'
opnfv_logo='releng/docs/etc/opnfv-logo.png'
diff --git a/utils/fetch_os_creds.sh b/utils/fetch_os_creds.sh
index d1b192dcb..a5c601b82 100755
--- a/utils/fetch_os_creds.sh
+++ b/utils/fetch_os_creds.sh
@@ -94,7 +94,7 @@ if [ "$installer_type" == "fuel" ]; then
#This file contains the mgmt keystone API, we need the public one for our rc file
admin_ip=$(cat $dest_path | grep "OS_AUTH_URL" | sed 's/^.*\=//' | sed "s/^\([\"']\)\(.*\)\1\$/\2/g" | sed s'/\/$//')
public_ip=$(sshpass -p r00tme ssh $ssh_options root@${installer_ip} \
- "ssh ${controller_ip} 'source openrc; keystone endpoint-list'" \
+ "ssh ${controller_ip} 'source openrc; openstack endpoint list --long'" \
| grep $admin_ip | sed 's/ /\n/g' | grep ^http | head -1) &> /dev/null
#| grep http | head -1 | cut -d '|' -f 4 | sed 's/v1\/.*/v1\//' | sed 's/ //g') &> /dev/null
#NOTE: this is super ugly sed 's/v1\/.*/v1\//'OS_AUTH_URL
diff --git a/utils/push-test-logs.sh b/utils/push-test-logs.sh
index 331cbd942..68f80874a 100644
--- a/utils/push-test-logs.sh
+++ b/utils/push-test-logs.sh
@@ -36,9 +36,9 @@ if [ -d "$dir_result" ]; then
if [ $? != 0 ]; then
echo "Not possible to push results to artifact: gsutil not installed";
else
- gsutil ls gs://artifacts.opnfv.org/"$project_artifact"/ &>/dev/null
+ gsutil ls gs://artifacts.opnfv.org/"$project"/ &>/dev/null
if [ $? != 0 ]; then
- echo "Not possible to push results to artifact: wrong credentials.";
+ echo "Not possible to push results to artifact: gsutil not installed.";
else
echo "copy result files to artifact $project_artifact"
gsutil -m cp -r "$dir_result" gs://artifacts.opnfv.org/"$project_artifact"/
diff --git a/utils/test/result_collection_api/dashboard/dashboard_utils.py b/utils/test/result_collection_api/dashboard/dashboard_utils.py
index 8d83b006a..3252c3af6 100644
--- a/utils/test/result_collection_api/dashboard/dashboard_utils.py
+++ b/utils/test/result_collection_api/dashboard/dashboard_utils.py
@@ -26,6 +26,8 @@ from bottlenecks2Dashboard import format_bottlenecks_for_dashboard, \
check_bottlenecks_case_exist
from qtip2Dashboard import format_qtip_for_dashboard, \
check_qtip_case_exist
+from promise2Dashboard import format_promise_for_dashboard, \
+ check_promise_case_exist
# any project test project wishing to provide dashboard ready values
# must include at least 2 methods
diff --git a/utils/test/result_collection_api/dashboard/functest2Dashboard.py b/utils/test/result_collection_api/dashboard/functest2Dashboard.py
index 65dbca654..a817f7597 100644
--- a/utils/test/result_collection_api/dashboard/functest2Dashboard.py
+++ b/utils/test/result_collection_api/dashboard/functest2Dashboard.py
@@ -17,12 +17,14 @@
import re
import datetime
+
def get_functest_cases():
"""
get the list of the supported test cases
TODO: update the list when adding a new test case for the dashboard
"""
- return ["status", "vPing", "vPing_userdata", "vIMS", "Tempest", "ODL", "ONOS", "Rally"]
+ return ["status", "vPing", "vPing_userdata", "vIMS", "Tempest", "ODL",
+ "ONOS", "Rally"]
def format_functest_for_dashboard(case, results):
@@ -66,7 +68,9 @@ def format_status_for_dashboard(results):
test_data.append({'nb test suite(s) run': len(testcases)-1})
test_data.append({'vPing': '100%'})
test_data.append({'VIM status': '82%'})
- test_data.append({'SDN Controllers': {'odl':'92%', 'onos':'95%', 'opencontrail':'93%'}})
+ test_data.append({'SDN Controllers': {'odl': '92%',
+ 'onos': '95%',
+ 'ocl': '93%'}})
test_data.append({'VNF deployment': '95%'})
return test_data
@@ -227,7 +231,7 @@ def format_ODL_for_dashboard(results):
nbFailures = 0
for odl in odl_results:
if (odl['test_status']['@status'] == "FAIL"):
- nbFailures+=1
+ nbFailures += 1
new_element.append({'x': data['creation_date'],
'y1': len(odl_results),
'y2': nbFailures})
@@ -246,25 +250,91 @@ def format_ONOS_for_dashboard(results):
Post processing for the odl test case
"""
test_data = [{'description': 'ONOS results for Dashboard'}]
- # Graph 1: (duration)=f(time)
+ # Graph 1: (duration FUNCvirtNet)=f(time)
# ***************************************
new_element = []
# default duration 0:00:08.999904
# consider only seconds => 09
for data in results:
- t = data['details']['duration']
- h,m,s = re.split(':',t)
+ t = data['details']['FUNCvirNet']['duration']
+ h, m, s = re.split(':', t)
s = round(float(s))
- new_duration = int(datetime.timedelta(hours=int(h),minutes=int(m),seconds=int(s)).total_seconds())
+ new_duration = int(datetime.timedelta(hours=int(h),
+ minutes=int(m),
+ seconds=int(s)).total_seconds())
new_element.append({'x': data['creation_date'],
'y': new_duration})
- test_data.append({'name': "ONOS duration",
+ test_data.append({'name': "ONOS FUNCvirNet duration ",
'info': {'type': "graph",
'xlabel': 'time (s)',
'ylabel': 'duration (s)'},
'data_set': new_element})
+
+ # Graph 2: (Nb test, nb failure)FuncvirtNet=f(time)
+ # ***************************************
+ new_element = []
+
+ for data in results:
+ onos_results = data['details']['FUNCvirNet']['status']
+ nbFailures = 0
+ for onos in onos_results:
+ if (onos['Case result'] == "FAIL"):
+ nbFailures += 1
+ new_element.append({'x': data['creation_date'],
+ 'y1': len(onos_results),
+ 'y2': nbFailures})
+
+ test_data.append({'name': "ONOS FUNCvirNet nb tests/nb failures",
+ 'info': {'type': "graph",
+ 'xlabel': 'time',
+ 'y1label': 'Number of tests',
+ 'y2label': 'Number of failures'},
+ 'data_set': new_element})
+
+ # Graph 3: (duration FUNCvirtNetL3)=f(time)
+ # ***************************************
+ new_element = []
+
+ # default duration 0:00:08.999904
+ # consider only seconds => 09
+ for data in results:
+ t = data['details']['FUNCvirNetL3']['duration']
+ h, m, s = re.split(':', t)
+ s = round(float(s))
+ new_duration = int(datetime.timedelta(hours=int(h),
+ minutes=int(m),
+ seconds=int(s)).total_seconds())
+ new_element.append({'x': data['creation_date'],
+ 'y': new_duration})
+
+ test_data.append({'name': "ONOS FUNCvirNetL3 duration",
+ 'info': {'type': "graph",
+ 'xlabel': 'time (s)',
+ 'ylabel': 'duration (s)'},
+ 'data_set': new_element})
+
+ # Graph 4: (Nb test, nb failure)FuncvirtNetL3=f(time)
+ # ***************************************
+ new_element = []
+
+ for data in results:
+ onos_results = data['details']['FUNCvirNetL3']['status']
+ nbFailures = 0
+ for onos in onos_results:
+ if (onos['Case result'] == "FAIL"):
+ nbFailures += 1
+ new_element.append({'x': data['creation_date'],
+ 'y1': len(onos_results),
+ 'y2': nbFailures})
+
+ test_data.append({'name': "ONOS FUNCvirNetL3 nb tests/nb failures",
+ 'info': {'type': "graph",
+ 'xlabel': 'time',
+ 'y1label': 'Number of tests',
+ 'y2label': 'Number of failures'},
+ 'data_set': new_element})
return test_data
@@ -312,6 +382,7 @@ def format_vPing_for_dashboard(results):
return test_data
+
def format_vPing_userdata_for_dashboard(results):
"""
Post processing for the vPing_userdata test case
@@ -347,3 +418,4 @@ def format_vPing_userdata_for_dashboard(results):
'Nb Success': nbTestOk}]})
return test_data
+
diff --git a/utils/test/result_collection_api/dashboard/promise2Dashboard.py b/utils/test/result_collection_api/dashboard/promise2Dashboard.py
new file mode 100644
index 000000000..84f43a7d1
--- /dev/null
+++ b/utils/test/result_collection_api/dashboard/promise2Dashboard.py
@@ -0,0 +1,103 @@
+ #!/usr/bin/python
+#
+# Copyright (c) 2015 Orange
+# morgan.richomme@orange.com
+#
+# This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# This script is used to build dashboard ready json results
+# It may be used for all the test case of the Promise project
+# a new method format_<Test_case>_for_dashboard(results)
+# v0.1: basic example with methods for odl, Tempest, Rally and vPing
+#
+import re
+import datetime
+
+
+def get_promise_cases():
+ """
+ get the list of the supported test cases
+ TODO: update the list when adding a new test case for the dashboard
+ """
+ return ["promise"]
+
+
+def format_promise_for_dashboard(case, results):
+ """
+ generic method calling the method corresponding to the test case
+ check that the testcase is properly declared first
+ then build the call to the specific method
+ """
+ if check_promise_case_exist(case):
+ # note we add _case because testcase and project had the same name
+ # TODO refactoring...looks fine at the beginning wit only 1 project
+ # not very ugly now and clearly not optimized...
+ cmd = "format_" + case + "_case_for_dashboard(results)"
+ res = eval(cmd)
+ else:
+ res = []
+ print "Test cases not declared"
+ return res
+
+
+def check_promise_case_exist(case):
+ """
+ check if the testcase exists
+ if the test case is not defined or not declared in the list
+ return False
+ """
+ promise_cases = get_promise_cases()
+
+ if (case is None or case not in promise_cases):
+ return False
+ else:
+ return True
+
+
+
+
+
+def format_promise_case_for_dashboard(results):
+ """
+ Post processing for the promise test case
+ """
+ test_data = [{'description': 'Promise results for Dashboard'}]
+ # Graph 1: (duration)=f(time)
+ # ***************************************
+ new_element = []
+
+ # default duration 0:00:08.999904
+ # consider only seconds => 09
+ for data in results:
+ t = data['details']['duration']
+ new_element.append({'x': data['creation_date'],
+ 'y': t})
+
+ test_data.append({'name': "Promise duration ",
+ 'info': {'type': "graph",
+ 'xlabel': 'time (s)',
+ 'ylabel': 'duration (s)'},
+ 'data_set': new_element})
+
+ # Graph 2: (Nb test, nb failure)=f(time)
+ # ***************************************
+ new_element = []
+
+ for data in results:
+ promise_results = data['details']
+ new_element.append({'x': data['creation_date'],
+ 'y1': promise_results['tests'],
+ 'y2': promise_results['failures']})
+
+ test_data.append({'name': "Promise nb tests/nb failures",
+ 'info': {'type': "graph",
+ 'xlabel': 'time',
+ 'y1label': 'Number of tests',
+ 'y2label': 'Number of failures'},
+ 'data_set': new_element})
+
+ return test_data