diff options
22 files changed, 565 insertions, 183 deletions
diff --git a/jjb/apex/apex.yml b/jjb/apex/apex.yml index 785056ff0..1ab7f647f 100644 --- a/jjb/apex/apex.yml +++ b/jjb/apex/apex.yml @@ -84,9 +84,6 @@ pattern: 'lib/**' - compare-type: ANT pattern: 'config/**' - forbidden-file-paths: - - compare-type: ANT - pattern: 'docs/**' properties: - build-blocker: @@ -160,16 +157,16 @@ git-revision: false block: true - trigger-builds: - - project: 'functest-apex-opnfv-jump-1-daily-{stream}' - predefined-parameters: - DEPLOY_SCENARIO={scenario} + - project: 'yardstick-apex-opnfv-jump-1-daily-{stream}' block: true block-thresholds: build-step-failure-threshold: 'never' failure-threshold: 'never' unstable-threshold: 'FAILURE' - trigger-builds: - - project: 'yardstick-apex-opnfv-jump-1-daily-{stream}' + - project: 'functest-apex-opnfv-jump-1-daily-{stream}' + predefined-parameters: + DEPLOY_SCENARIO={scenario} block: true block-thresholds: build-step-failure-threshold: 'never' diff --git a/jjb/compass4nfv/compass-ci-jobs.yml b/jjb/compass4nfv/compass-ci-jobs.yml index a4be24ec0..5564d6989 100644 --- a/jjb/compass4nfv/compass-ci-jobs.yml +++ b/jjb/compass4nfv/compass-ci-jobs.yml @@ -168,11 +168,9 @@ builders: - shell: - !include-raw: ./compass-workspace-precleanup.sh + !include-raw-escape: ./compass-download-artifact.sh - shell: - !include-raw: ./compass-download-artifact.sh - - shell: - !include-raw: ./compass-deploy.sh + !include-raw-escape: ./compass-deploy.sh publishers: - archive: diff --git a/jjb/compass4nfv/compass-deploy.sh b/jjb/compass4nfv/compass-deploy.sh index 697a1d573..860f5da1b 100644 --- a/jjb/compass4nfv/compass-deploy.sh +++ b/jjb/compass4nfv/compass-deploy.sh @@ -6,24 +6,38 @@ echo "Starting the deployment on baremetal environment using $INSTALLER_TYPE. Th echo "--------------------------------------------------------" echo +# source the properties file so we get OPNFV vars +source $BUILD_DIRECTORY/latest.properties + +# echo the info about artifact that is used during the deployment +echo "Using ${OPNFV_ARTIFACT_URL/*\/} for deployment" + +if [[ ! "$JOB_NAME" =~ (verify|merge) ]]; then + # for none-merge deployments + # checkout the commit that was used for building the downloaded artifact + # to make sure the ISO and deployment mechanism uses same versions + echo "Checking out $OPNFV_GIT_SHA1" + git checkout $OPNFV_GIT_SHA1 --quiet +fi + echo 1 > /proc/sys/vm/drop_caches export CONFDIR=$WORKSPACE/deploy/conf export ISO_URL=file://$BUILD_DIRECTORY/compass.iso if [[ "$NODE_NAME" =~ "-vm" ]]; then export NETWORK_CONF=$CONFDIR/vm_environment/$NODE_NAME/network.yml - export DHA_CONF=$CONFDIR/vm_environment/${{DEPLOY_SCENARIO}}.yml + export DHA_CONF=$CONFDIR/vm_environment/${DEPLOY_SCENARIO}.yml else export INSTALL_NIC=eth1 export NETWORK_CONF=$CONFDIR/hardware_environment/$NODE_NAME/network.yml - export DHA_CONF=$CONFDIR/hardware_environment/$NODE_NAME/${{DEPLOY_SCENARIO}}.yml + export DHA_CONF=$CONFDIR/hardware_environment/$NODE_NAME/${DEPLOY_SCENARIO}.yml fi cd $WORKSPACE -export OS_VERSION=${{COMPASS_OS_VERSION}} -export OPENSTACK_VERSION=${{COMPASS_OPENSTACK_VERSION}} -./deploy.sh --dha ${{DHA_CONF}} --network ${{NETWORK_CONF}} +export OS_VERSION=${COMPASS_OS_VERSION} +export OPENSTACK_VERSION=${COMPASS_OPENSTACK_VERSION} +./deploy.sh --dha ${DHA_CONF} --network ${NETWORK_CONF} if [ $? -ne 0 ]; then echo "depolyment failed!" deploy_ret=1 @@ -34,6 +48,6 @@ echo "--------------------------------------------------------" echo "Done!" ssh_options="-o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no" -sshpass -p root scp 2>/dev/null $ssh_options root@${{INSTALLER_IP}}:/var/ansible/run/openstack_${{COMPASS_OPENSTACK_VERSION}}-opnfv2/ansible.log ./ &> /dev/null +sshpass -p root scp 2>/dev/null $ssh_options root@${INSTALLER_IP}:/var/ansible/run/openstack_${COMPASS_OPENSTACK_VERSION}-opnfv2/ansible.log ./ &> /dev/null exit $deploy_ret diff --git a/jjb/compass4nfv/compass-project-jobs.yml b/jjb/compass4nfv/compass-project-jobs.yml index 8943afb37..11e1d5f8a 100644 --- a/jjb/compass4nfv/compass-project-jobs.yml +++ b/jjb/compass4nfv/compass-project-jobs.yml @@ -116,9 +116,9 @@ builders: - shell: - !include-raw: ./compass-build.sh + !include-raw-escape: ./compass-build.sh - shell: - !include-raw: ./compass-deploy.sh + !include-raw-escape: ./compass-deploy.sh publishers: - archive: @@ -157,11 +157,11 @@ builders: - shell: - !include-raw: ./compass-build.sh + !include-raw-escape: ./compass-build.sh - shell: - !include-raw: ./compass-upload-artifact.sh + !include-raw-escape: ./compass-upload-artifact.sh - shell: - !include-raw: ./compass-workspace-cleanup.sh + !include-raw-escape: ./compass-workspace-cleanup.sh - job-template: name: 'compass-build-ppa-{stream}' @@ -194,7 +194,7 @@ builders: - shell: - !include-raw: ./compass-makeppa.sh + !include-raw-escape: ./compass-makeppa.sh diff --git a/jjb/doctor/doctor.yml b/jjb/doctor/doctor.yml index 167b696a3..b2cbb0242 100644 --- a/jjb/doctor/doctor.yml +++ b/jjb/doctor/doctor.yml @@ -53,4 +53,4 @@ pattern: 'tests/**' builders: - - shell: "[ -e tests/run.sh ] && ./tests/run.sh" + - shell: "[ -e tests/run.sh ] && bash -n ./tests/run.sh" diff --git a/jjb/fuel/fuel-ci-jobs.yml b/jjb/fuel/fuel-ci-jobs.yml index 33c54eced..6b69afe26 100644 --- a/jjb/fuel/fuel-ci-jobs.yml +++ b/jjb/fuel/fuel-ci-jobs.yml @@ -35,6 +35,8 @@ #-------------------------------- # master #-------------------------------- + - opnfv-jump-2: + <<: *master - ericsson-pod1: <<: *master - ericsson-pod2: @@ -46,23 +48,22 @@ #-------------------------------- scenario: # HA scenarios - - 'os-onos-nofeature-ha': -# auto-trigger-name: 'fuel-{scenario}-{pod}-trigger' - auto-trigger-name: 'brahmaputra-trigger-daily-disabled' - 'os-odl_l2-nofeature-ha': - auto-trigger-name: 'brahmaputra-trigger-daily-disabled' + auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger' - 'os-odl_l3-nofeature-ha': - auto-trigger-name: 'brahmaputra-trigger-daily-disabled' - - 'os-odl_l2-bgpvpn-ha': - auto-trigger-name: 'brahmaputra-trigger-daily-disabled' - - 'os-nosdn-nofeature-ha': - auto-trigger-name: 'brahmaputra-trigger-daily-disabled' + auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger' + - 'os-onos-nofeature-ha': + auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger' - 'os-nosdn-kvm-ha': + auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger' + - 'os-nosdn-ovs-ha': + auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger' + - 'os-nosdn-nofeature-ha': + auto-trigger-name: 'fuel-{scenario}-{pod}-{stream}-trigger' + - 'os-odl_l2-bgpvpn-ha': auto-trigger-name: 'brahmaputra-trigger-daily-disabled' - 'os-nosdn-kvm_ovs-ha': auto-trigger-name: 'brahmaputra-trigger-daily-disabled' - - 'os-nosdn-ovs-ha': - auto-trigger-name: 'brahmaputra-trigger-daily-disabled' - 'os-nosdn-vlan-ha': auto-trigger-name: 'brahmaputra-trigger-daily-disabled' # NOHA scenarios @@ -226,74 +227,207 @@ ######################## # trigger macros ######################## -# os-onos3-nofeature-ha trigger +# os-nosdn-nofeature-ha trigger +- trigger: + name: 'fuel-os-nosdn-nofeature-ha-opnfv-jump-2-master-trigger' + triggers: + - timed: '0 14 * * *' +- trigger: + name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod1-master-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod2-master-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-nofeature-ha-virtual-master-trigger' + triggers: + - timed: '' - trigger: - name: 'fuel-os-onos-nofeature-ha-opnfv-jump-2-trigger' + name: 'fuel-os-nosdn-nofeature-ha-opnfv-jump-2-brahmaputra-trigger' triggers: - - timed: '0 20 * * *' + - timed: '' +- trigger: + name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod1-brahmaputra-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-nofeature-ha-ericsson-pod2-brahmaputra-trigger' + triggers: + - timed: '0 14 * * *' +- trigger: + name: 'fuel-os-nosdn-nofeature-ha-virtual-brahmaputra-trigger' + triggers: + - timed: '' + +# os-nosdn-kvm-ha trigger - trigger: - name: 'fuel-os-onos-nofeature-ha-ericsson-pod1-trigger' + name: 'fuel-os-nosdn-kvm-ha-opnfv-jump-2-master-trigger' + triggers: + - timed: '0 18 * * *' +- trigger: + name: 'fuel-os-nosdn-kvm-ha-ericsson-pod1-master-trigger' triggers: - timed: '' - trigger: - name: 'fuel-os-onos-nofeature-ha-ericsson-pod2-trigger' + name: 'fuel-os-nosdn-kvm-ha-ericsson-pod2-master-trigger' triggers: - timed: '' - trigger: - name: 'fuel-os-onos-nofeature-ha-virtual-trigger' + name: 'fuel-os-nosdn-kvm-ha-virtual-master-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-kvm-ha-opnfv-jump-2-brahmaputra-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-kvm-ha-ericsson-pod1-brahmaputra-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-kvm-ha-ericsson-pod2-brahmaputra-trigger' + triggers: + - timed: '0 18 * * *' +- trigger: + name: 'fuel-os-nosdn-kvm-ha-virtual-brahmaputra-trigger' triggers: - timed: '' # os-odl_l2-nofeature-ha trigger - trigger: - name: 'fuel-os-odl_l2-nofeature-ha-opnfv-jump-2-trigger' + name: 'fuel-os-odl_l2-nofeature-ha-opnfv-jump-2-master-trigger' + triggers: + - timed: '0 2 * * *' +- trigger: + name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod1-master-trigger' triggers: - - timed: '0 0 * * *' + - timed: '' - trigger: - name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod1-trigger' + name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod2-master-trigger' triggers: - timed: '' - trigger: - name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod2-trigger' + name: 'fuel-os-odl_l2-nofeature-ha-virtual-master-trigger' triggers: - timed: '' - trigger: - name: 'fuel-os-odl_l2-nofeature-ha-virtual-trigger' + name: 'fuel-os-odl_l2-nofeature-ha-opnfv-jump-2-brahmaputra-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod1-brahmaputra-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-odl_l2-nofeature-ha-ericsson-pod2-brahmaputra-trigger' + triggers: + - timed: '0 2 * * *' +- trigger: + name: 'fuel-os-odl_l2-nofeature-ha-virtual-brahmaputra-trigger' triggers: - timed: '' # os-odl_l3-nofeature-ha trigger - trigger: - name: 'fuel-os-odl_l3-nofeature-ha-opnfv-jump-2-trigger' + name: 'fuel-os-odl_l3-nofeature-ha-opnfv-jump-2-master-trigger' triggers: - - timed: '0 4 * * *' + - timed: '0 6 * * *' - trigger: - name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod1-trigger' + name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod1-master-trigger' triggers: - timed: '' - trigger: - name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod2-trigger' + name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod2-master-trigger' triggers: - timed: '' - trigger: - name: 'fuel-os-odl_l3-nofeature-ha-virtual-trigger' + name: 'fuel-os-odl_l3-nofeature-ha-virtual-master-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-odl_l3-nofeature-ha-opnfv-jump-2-brahmaputra-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod1-brahmaputra-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-odl_l3-nofeature-ha-ericsson-pod2-brahmaputra-trigger' + triggers: + - timed: '0 6 * * *' +- trigger: + name: 'fuel-os-odl_l3-nofeature-ha-virtual-brahmaputra-trigger' triggers: - timed: '' -# os-odl_l2-bgpvpn-ha trigger +# os-onos-nofeature-ha trigger - trigger: - name: 'fuel-os-odl_l2-bgpvpn-ha-opnfv-jump-2-trigger' + name: 'fuel-os-onos-nofeature-ha-opnfv-jump-2-master-trigger' triggers: - - timed: '0 8 * * *' + - timed: '0 10 * * *' +- trigger: + name: 'fuel-os-onos-nofeature-ha-ericsson-pod1-master-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-onos-nofeature-ha-ericsson-pod2-master-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-onos-nofeature-ha-virtual-master-trigger' + triggers: + - timed: '' - trigger: - name: 'fuel-os-odl_l2-bgpvpn-ha-ericsson-pod1-trigger' + name: 'fuel-os-onos-nofeature-ha-opnfv-jump-2-brahmaputra-trigger' triggers: - timed: '' - trigger: - name: 'fuel-os-odl_l2-bgpvpn-ha-ericsson-pod2-trigger' + name: 'fuel-os-onos-nofeature-ha-ericsson-pod1-brahmaputra-trigger' triggers: - timed: '' - trigger: - name: 'fuel-os-odl_l2-bgpvpn-ha-virtual-trigger' + name: 'fuel-os-onos-nofeature-ha-ericsson-pod2-brahmaputra-trigger' + triggers: + - timed: '0 10 * * *' +- trigger: + name: 'fuel-os-onos-nofeature-ha-virtual-brahmaputra-trigger' + triggers: + - timed: '' + + +# os-nosdn-ovs-ha trigger +- trigger: + name: 'fuel-os-nosdn-ovs-ha-opnfv-jump-2-master-trigger' + triggers: + - timed: '0 22 * * *' +- trigger: + name: 'fuel-os-nosdn-ovs-ha-ericsson-pod1-master-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-ovs-ha-ericsson-pod2-master-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-ovs-ha-virtual-master-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-ovs-ha-opnfv-jump-2-brahmaputra-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-ovs-ha-ericsson-pod1-brahmaputra-trigger' + triggers: + - timed: '' +- trigger: + name: 'fuel-os-nosdn-ovs-ha-ericsson-pod2-brahmaputra-trigger' + triggers: + - timed: '0 22 * * *' +- trigger: + name: 'fuel-os-nosdn-ovs-ha-virtual-brahmaputra-trigger' triggers: - timed: '' diff --git a/jjb/fuel/fuel-download-artifact.sh b/jjb/fuel/fuel-download-artifact.sh index ceac859dd..d78ddc5ba 100755 --- a/jjb/fuel/fuel-download-artifact.sh +++ b/jjb/fuel/fuel-download-artifact.sh @@ -36,7 +36,7 @@ if [[ ! "$JOB_NAME" =~ (verify|merge) ]]; then ISOSTORE="/iso_mount/opnfv_ci/${GIT_BRANCH##*/}" if [[ -f "$ISOSTORE/$OPNFV_ARTIFACT" ]]; then echo "ISO exists locally. Skipping the download and using the file from ISO store" - /bin/cp -f $ISOSTORE/$OPNFV_ARTIFACT $WORKSPACE/opnfv.iso + ln -s $ISOSTORE/$OPNFV_ARTIFACT $WORKSPACE/opnfv.iso echo "--------------------------------------------------------" echo ls -al $WORKSPACE/opnfv.iso diff --git a/jjb/fuel/fuel-upload-artifact.sh b/jjb/fuel/fuel-upload-artifact.sh index 6dd5f34d6..47cf8985c 100755 --- a/jjb/fuel/fuel-upload-artifact.sh +++ b/jjb/fuel/fuel-upload-artifact.sh @@ -29,17 +29,17 @@ if [[ ! "$JOB_NAME" =~ (verify|merge) ]]; then ls -tp | grep -v '/' | tail -n +6 | xargs -d '\n' /bin/rm -f -- # store ISO - echo "Storing $INSTALLER_TYPE artifact on NFS first" + echo "Storing $INSTALLER_TYPE artifact on NFS..." /bin/cp -f $BUILD_DIRECTORY/opnfv-$OPNFV_ARTIFACT_VERSION.iso \ $ISOSTORE/opnfv-$OPNFV_ARTIFACT_VERSION.iso fi fi # log info to console -echo echo "Uploading $INSTALLER_TYPE artifact. This could take some time..." echo +cd $WORKSPACE # upload artifact and additional files to google storage gsutil cp $BUILD_DIRECTORY/opnfv-$OPNFV_ARTIFACT_VERSION.iso \ gs://$GS_URL/opnfv-$OPNFV_ARTIFACT_VERSION.iso > gsutil.iso.log 2>&1 diff --git a/jjb/functest/functest-ci-jobs.yml b/jjb/functest/functest-ci-jobs.yml index de0592209..1165b2a2c 100644 --- a/jjb/functest/functest-ci-jobs.yml +++ b/jjb/functest/functest-ci-jobs.yml @@ -53,6 +53,9 @@ #-------------------------------- # master #-------------------------------- + - opnfv-jump-2: + installer: fuel + <<: *master - intel-pod6: installer: joid <<: *master diff --git a/jjb/opnfv/installer-params.yml b/jjb/opnfv/installer-params.yml index e03b4749e..57ca9864e 100644 --- a/jjb/opnfv/installer-params.yml +++ b/jjb/opnfv/installer-params.yml @@ -47,7 +47,7 @@ description: 'Installer used for deploying OPNFV on this POD' - string: name: EXTERNAL_NETWORK - default: 'net04_ext' + default: 'admin_floating_net' description: 'external network for test' - parameter: diff --git a/jjb/opnfv/slave-params.yml b/jjb/opnfv/slave-params.yml index b2c6e7cf5..5e0274d84 100644 --- a/jjb/opnfv/slave-params.yml +++ b/jjb/opnfv/slave-params.yml @@ -81,10 +81,6 @@ name: GIT_BASE default: https://gerrit.opnfv.org/gerrit/$PROJECT description: 'Git URL to use on this Jenkins Slave' - - string: - name: EXTERNAL_NETWORK - default: 'admin_floating_net' - description: 'external network' - parameter: name: 'intelpod2-jumphost-defaults' @@ -424,3 +420,41 @@ name: SSH_KEY default: /root/.ssh/id_rsa description: 'SSH key to use for Apex' + +- parameter: + name: 'dell-us-testing-bm-1-defaults' + parameters: + - node: + name: SLAVE_NAME + description: 'Slave name on Jenkins' + allowed-slaves: + - dell-us-testing-bm-1 + default-slaves: + - dell-us-testing-bm-1 + - string: + name: INSTALLER_VERSION + default: latest + description: 'Version of the installer to deploy' + - string: + name: GIT_BASE + default: https://gerrit.opnfv.org/gerrit/$PROJECT + description: 'Git URL to use on this Jenkins Slave' + +- parameter: + name: 'dell-us-deploying-bm3-defaults' + parameters: + - node: + name: SLAVE_NAME + description: 'Slave name on Jenkins' + allowed-slaves: + - dell-us-deploying-bm3 + default-slaves: + - dell-us-deploying-bm3 + - string: + name: INSTALLER_VERSION + default: latest + description: 'Version of the installer to deploy' + - string: + name: GIT_BASE + default: https://gerrit.opnfv.org/gerrit/$PROJECT + description: 'Git URL to use on this Jenkins Slave' diff --git a/jjb/qtip/qtip.yml b/jjb/qtip/qtip-ci-jobs.yml index 450ba938e..d33bdda32 100644 --- a/jjb/qtip/qtip.yml +++ b/jjb/qtip/qtip-ci-jobs.yml @@ -1,105 +1,74 @@ +#################################### +# job configuration for qtip +#################################### - project: - name: qtip - - project: '{name}' - - jobs: - - 'qtip-verify-{stream}' - - 'qtip-{installer_type}-{pod}-daily-{stream}' - + name: qtip-ci-jobs + + project: 'qtip' + +#-------------------------------- +# BRANCH ANCHORS +#-------------------------------- + master: &master + stream: master + branch: '{stream}' + gs-pathname: '' + brahmaputra: &brahmaputra + stream: brahmaputra + branch: 'stable/{stream}' + gs-pathname: '{stream}' +#-------------------------------- +# POD, INSTALLER, AND BRANCH MAPPING +#-------------------------------- +# brahmaputra +#-------------------------------- pod: - dell-us-testing-bm-1: - node: 'dell-us-testing-bm-1' - installer_type: 'fuel' - installer_ip: '10.20.0.2' + installer: compass + <<: *brahmaputra + - orange-pod2: + installer: joid + <<: *brahmaputra - dell-us-deploying-bm3: - node: 'dell-us-deploying-bm3' - installer_type: 'fuel' - installer_ip: '10.20.0.2' - -# only master branch is enabled at the moment to keep no of jobs sane - stream: - - master: - branch: 'master' - gs-pathname: '' -# - brahmaputra: -# branch: 'stable/brahmaputra' -# gs-pathname: '/brahmaputra' - -- job-template: - name: 'qtip-verify-{stream}' - - parameters: - - project-parameter: - project: '{project}' - - gerrit-parameter: - branch: '{branch}' - - 'opnfv-build-defaults' - - scm: - - gerrit-trigger-scm: - credentials-id: '{ssh-credentials}' - refspec: '$GERRIT_REFSPEC' - choosing-strategy: 'gerrit' - - triggers: - - gerrit: - trigger-on: - - patchset-created-event: - exclude-drafts: 'false' - exclude-trivial-rebase: 'false' - exclude-no-code-change: 'false' - - draft-published-event - - comment-added-contains-event: - comment-contains-value: 'recheck' - - comment-added-contains-event: - comment-contains-value: 'reverify' - projects: - - project-compare-type: 'ANT' - project-pattern: '{project}' - branches: - - branch-compare-type: 'ANT' - branch-pattern: '**/{branch}' - forbidden-file-paths: - - compare-type: ANT - pattern: 'docs/**|.gitignore' - - builders: - - shell: | - echo "Nothing to verify!" + installer: fuel + <<: *brahmaputra + +#-------------------------------- +# master +#-------------------------------- + - juniper-pod1: + installer: joid + <<: *master +# - zte-build-1: #would be confirmed with the ZTE lab by tomorrow +# installer: fuel +# <<: *master + +#-------------------------------- + jobs: + - 'qtip-{installer}-{pod}-daily-{stream}' +################################ +# job templates +################################ - job-template: - name: 'qtip-{installer_type}-{pod}-daily-{stream}' + name: 'qtip-{installer}-{pod}-daily-{stream}' disabled: false - node: '{node}' - parameters: - project-parameter: project: '{project}' + - '{pod}-defaults' + - '{installer}-defaults' - string: - name: INSTALLER_TYPE - default: '{installer_type}' - description: "Installer name that is used for deployment." - - string: - name: INSTALLER_IP - default: '{installer_ip}' - description: "Installer IP" - - string: - name: NODE_NAME - default: '{node}' - description: "Name of the POD" - - string: - name: GIT_BASE - default: https://gerrit.opnfv.org/gerrit/$PROJECT - description: "Used for overriding the GIT URL coming from Global Jenkins configuration in case if the stuff is done on none-LF HW." + name: DEPLOY_SCENARIO + default: 'os-nosdn-nofeature-ha' scm: - git-scm: credentials-id: '{ssh-credentials}' refspec: '' - branch: master + branch: '{branch}' builders: - 'qtip-cleanup' @@ -107,9 +76,9 @@ - 'qtip-run-suite' - 'qtip-pushtoDB' - triggers: - - timed: '0 0,13,15,20 * * *' - + publishers: + - email: + recipients: nauman.ahad@xflowresearch.com, mofassir.arif@xflowresearch.com, vikram@nvirters.com ########################### #biuilder macros @@ -123,7 +92,6 @@ envs="INSTALLER_TYPE=${INSTALLER_TYPE} -e INSTALLER_IP=${INSTALLER_IP} -e NODE_NAME=${NODE_NAME}" suite="TEST_CASE=all" docker pull opnfv/qtip:latest - cmd=" docker run -id -e $envs -e $suite opnfv/qtip:latest /bin/bash" echo "Qtip: Running docker run command: ${cmd}" ${cmd} @@ -190,17 +158,3 @@ docker rmi opnfv/qtip:$tag done fi - -- builder: - name: qtip-fetch-os-cred - builders: - - shell: | - #!/bin/bash - - echo $HOME - echo $WORKSPACE - cd $WORKSPACE - git clone https://gerrit.opnfv.org/gerrit/releng - cd $WORKSPACE/releng/utils - #./fetch_os_creds.sh #To test the problem with creating the file in $HOME - ./fetch_os_creds.sh -d $WORKSPACE/opnfv-stack.sh #To test if the file can be created in $WORKSPACE diff --git a/jjb/qtip/qtip-project-jobs.yml b/jjb/qtip/qtip-project-jobs.yml new file mode 100644 index 000000000..c77bac722 --- /dev/null +++ b/jjb/qtip/qtip-project-jobs.yml @@ -0,0 +1,58 @@ +- project: + name: qtip + + project: '{name}' + + jobs: + - 'qtip-verify-{stream}' + +# only master branch is enabled at the moment to keep no of jobs sane + stream: + - master: + branch: '{stream}' + gs-pathname: '' + - brahmaputra: + branch: 'stable/{stream}' + gs-pathname: '/{stream}' + +- job-template: + name: 'qtip-verify-{stream}' + + parameters: + - project-parameter: + project: '{project}' + - gerrit-parameter: + branch: '{branch}' + - 'opnfv-build-defaults' + + scm: + - gerrit-trigger-scm: + credentials-id: '{ssh-credentials}' + refspec: '$GERRIT_REFSPEC' + choosing-strategy: 'gerrit' + + triggers: + - gerrit: + trigger-on: + - patchset-created-event: + exclude-drafts: 'false' + exclude-trivial-rebase: 'false' + exclude-no-code-change: 'false' + - draft-published-event + - comment-added-contains-event: + comment-contains-value: 'recheck' + - comment-added-contains-event: + comment-contains-value: 'reverify' + projects: + - project-compare-type: 'ANT' + project-pattern: '{project}' + branches: + - branch-compare-type: 'ANT' + branch-pattern: '**/{branch}' + forbidden-file-paths: + - compare-type: ANT + pattern: 'docs/**|.gitignore' + + builders: + - shell: | + echo "Nothing to verify!" diff --git a/jjb/releng-macros.yaml b/jjb/releng-macros.yaml index 5d9570705..b43a76d5f 100644 --- a/jjb/releng-macros.yaml +++ b/jjb/releng-macros.yaml @@ -195,7 +195,7 @@ echo "###########################" echo - gs_base="artifacts.opnfv.org/review" + gs_base="artifacts.opnfv.org/$PROJECT/review" gs_path="$gs_base/$GERRIT_CHANGE_NUMBER" local_path="upload/$GERRIT_CHANGE_NUMBER" @@ -283,6 +283,13 @@ [[ $GERRIT_CHANGE_NUMBER =~ .+ ]] + gs_path="artifacts.opnfv.org/$PROJECT/review/$GERRIT_CHANGE_NUMBER" + + if gsutil ls "gs://$gs_path" > /dev/null 2>&1 ; then + echo + echo "Deleting Out-of-dated Documents..." + gsutil -m rm -r "gs://$gs_path" + fi gs_path="artifacts.opnfv.org/review/$GERRIT_CHANGE_NUMBER" if gsutil ls "gs://$gs_path" > /dev/null 2>&1 ; then diff --git a/jjb/yardstick/yardstick-ci-jobs.yml b/jjb/yardstick/yardstick-ci-jobs.yml index ec5c547ee..e7221e70c 100644 --- a/jjb/yardstick/yardstick-ci-jobs.yml +++ b/jjb/yardstick/yardstick-ci-jobs.yml @@ -71,6 +71,11 @@ #-------------------------------- # master #-------------------------------- + - opnfv-jump-2: + installer: fuel + suite: daily + auto-trigger-name: 'brahmaputra-trigger-daily-disabled' + <<: *master - ericsson-pod1: installer: fuel suite: daily @@ -328,7 +333,7 @@ parameters: - string: name: YARDSTICK_DB_BACKEND - default: '-r 213.77.62.197/results' + default: '-i 10.118.36.90:8086' description: 'Arguments to use in order to choose the backend DB' - parameter: @@ -336,7 +341,7 @@ parameters: - string: name: YARDSTICK_DB_BACKEND - default: '-r 213.77.62.197/results' + default: '-i 10.118.36.90:8086' description: 'Arguments to use in order to choose the backend DB' - parameter: diff --git a/jjb/yardstick/yardstick-project-jobs.yml b/jjb/yardstick/yardstick-project-jobs.yml index 749ebf574..f8f1abbff 100644 --- a/jjb/yardstick/yardstick-project-jobs.yml +++ b/jjb/yardstick/yardstick-project-jobs.yml @@ -120,12 +120,13 @@ # install python packages easy_install -U setuptools - pip install Sphinx==1.3.1 doc8 docutils sphinxcontrib-httpdomain python setup.py develop # unit tests ./run_tests.sh + pip install Sphinx==1.3.1 doc8 docutils sphinxcontrib-httpdomain + # generate api doc sources sphinx-apidoc -o docs/apidocs yardstick diff --git a/utils/docs-build.sh b/utils/docs-build.sh index e5d6b4a7f..3c0e8bd80 100755 --- a/utils/docs-build.sh +++ b/utils/docs-build.sh @@ -41,7 +41,7 @@ revision="$(git rev-parse --short HEAD)" rev_full="$(git rev-parse HEAD)" version="$(git describe --abbrev=0 2> /dev/null || echo draft) ($revision)" project="$(basename $(git rev-parse --show-toplevel))" -html_notes="\n Revision: $rev_full\n\n Build date: |today|" +html_notes=" Revision: $rev_full\n Build date: |today|" default_conf='releng/docs/etc/conf.py' opnfv_logo='releng/docs/etc/opnfv-logo.png' diff --git a/utils/fetch_os_creds.sh b/utils/fetch_os_creds.sh index d1b192dcb..a5c601b82 100755 --- a/utils/fetch_os_creds.sh +++ b/utils/fetch_os_creds.sh @@ -94,7 +94,7 @@ if [ "$installer_type" == "fuel" ]; then #This file contains the mgmt keystone API, we need the public one for our rc file admin_ip=$(cat $dest_path | grep "OS_AUTH_URL" | sed 's/^.*\=//' | sed "s/^\([\"']\)\(.*\)\1\$/\2/g" | sed s'/\/$//') public_ip=$(sshpass -p r00tme ssh $ssh_options root@${installer_ip} \ - "ssh ${controller_ip} 'source openrc; keystone endpoint-list'" \ + "ssh ${controller_ip} 'source openrc; openstack endpoint list --long'" \ | grep $admin_ip | sed 's/ /\n/g' | grep ^http | head -1) &> /dev/null #| grep http | head -1 | cut -d '|' -f 4 | sed 's/v1\/.*/v1\//' | sed 's/ //g') &> /dev/null #NOTE: this is super ugly sed 's/v1\/.*/v1\//'OS_AUTH_URL diff --git a/utils/push-test-logs.sh b/utils/push-test-logs.sh index 331cbd942..68f80874a 100644 --- a/utils/push-test-logs.sh +++ b/utils/push-test-logs.sh @@ -36,9 +36,9 @@ if [ -d "$dir_result" ]; then if [ $? != 0 ]; then echo "Not possible to push results to artifact: gsutil not installed"; else - gsutil ls gs://artifacts.opnfv.org/"$project_artifact"/ &>/dev/null + gsutil ls gs://artifacts.opnfv.org/"$project"/ &>/dev/null if [ $? != 0 ]; then - echo "Not possible to push results to artifact: wrong credentials."; + echo "Not possible to push results to artifact: gsutil not installed."; else echo "copy result files to artifact $project_artifact" gsutil -m cp -r "$dir_result" gs://artifacts.opnfv.org/"$project_artifact"/ diff --git a/utils/test/result_collection_api/dashboard/dashboard_utils.py b/utils/test/result_collection_api/dashboard/dashboard_utils.py index 8d83b006a..3252c3af6 100644 --- a/utils/test/result_collection_api/dashboard/dashboard_utils.py +++ b/utils/test/result_collection_api/dashboard/dashboard_utils.py @@ -26,6 +26,8 @@ from bottlenecks2Dashboard import format_bottlenecks_for_dashboard, \ check_bottlenecks_case_exist from qtip2Dashboard import format_qtip_for_dashboard, \ check_qtip_case_exist +from promise2Dashboard import format_promise_for_dashboard, \ + check_promise_case_exist # any project test project wishing to provide dashboard ready values # must include at least 2 methods diff --git a/utils/test/result_collection_api/dashboard/functest2Dashboard.py b/utils/test/result_collection_api/dashboard/functest2Dashboard.py index 65dbca654..a817f7597 100644 --- a/utils/test/result_collection_api/dashboard/functest2Dashboard.py +++ b/utils/test/result_collection_api/dashboard/functest2Dashboard.py @@ -17,12 +17,14 @@ import re import datetime + def get_functest_cases(): """ get the list of the supported test cases TODO: update the list when adding a new test case for the dashboard """ - return ["status", "vPing", "vPing_userdata", "vIMS", "Tempest", "ODL", "ONOS", "Rally"] + return ["status", "vPing", "vPing_userdata", "vIMS", "Tempest", "ODL", + "ONOS", "Rally"] def format_functest_for_dashboard(case, results): @@ -66,7 +68,9 @@ def format_status_for_dashboard(results): test_data.append({'nb test suite(s) run': len(testcases)-1}) test_data.append({'vPing': '100%'}) test_data.append({'VIM status': '82%'}) - test_data.append({'SDN Controllers': {'odl':'92%', 'onos':'95%', 'opencontrail':'93%'}}) + test_data.append({'SDN Controllers': {'odl': '92%', + 'onos': '95%', + 'ocl': '93%'}}) test_data.append({'VNF deployment': '95%'}) return test_data @@ -227,7 +231,7 @@ def format_ODL_for_dashboard(results): nbFailures = 0 for odl in odl_results: if (odl['test_status']['@status'] == "FAIL"): - nbFailures+=1 + nbFailures += 1 new_element.append({'x': data['creation_date'], 'y1': len(odl_results), 'y2': nbFailures}) @@ -246,25 +250,91 @@ def format_ONOS_for_dashboard(results): Post processing for the odl test case """ test_data = [{'description': 'ONOS results for Dashboard'}] - # Graph 1: (duration)=f(time) + # Graph 1: (duration FUNCvirtNet)=f(time) # *************************************** new_element = [] # default duration 0:00:08.999904 # consider only seconds => 09 for data in results: - t = data['details']['duration'] - h,m,s = re.split(':',t) + t = data['details']['FUNCvirNet']['duration'] + h, m, s = re.split(':', t) s = round(float(s)) - new_duration = int(datetime.timedelta(hours=int(h),minutes=int(m),seconds=int(s)).total_seconds()) + new_duration = int(datetime.timedelta(hours=int(h), + minutes=int(m), + seconds=int(s)).total_seconds()) new_element.append({'x': data['creation_date'], 'y': new_duration}) - test_data.append({'name': "ONOS duration", + test_data.append({'name': "ONOS FUNCvirNet duration ", 'info': {'type': "graph", 'xlabel': 'time (s)', 'ylabel': 'duration (s)'}, 'data_set': new_element}) + + # Graph 2: (Nb test, nb failure)FuncvirtNet=f(time) + # *************************************** + new_element = [] + + for data in results: + onos_results = data['details']['FUNCvirNet']['status'] + nbFailures = 0 + for onos in onos_results: + if (onos['Case result'] == "FAIL"): + nbFailures += 1 + new_element.append({'x': data['creation_date'], + 'y1': len(onos_results), + 'y2': nbFailures}) + + test_data.append({'name': "ONOS FUNCvirNet nb tests/nb failures", + 'info': {'type': "graph", + 'xlabel': 'time', + 'y1label': 'Number of tests', + 'y2label': 'Number of failures'}, + 'data_set': new_element}) + + # Graph 3: (duration FUNCvirtNetL3)=f(time) + # *************************************** + new_element = [] + + # default duration 0:00:08.999904 + # consider only seconds => 09 + for data in results: + t = data['details']['FUNCvirNetL3']['duration'] + h, m, s = re.split(':', t) + s = round(float(s)) + new_duration = int(datetime.timedelta(hours=int(h), + minutes=int(m), + seconds=int(s)).total_seconds()) + new_element.append({'x': data['creation_date'], + 'y': new_duration}) + + test_data.append({'name': "ONOS FUNCvirNetL3 duration", + 'info': {'type': "graph", + 'xlabel': 'time (s)', + 'ylabel': 'duration (s)'}, + 'data_set': new_element}) + + # Graph 4: (Nb test, nb failure)FuncvirtNetL3=f(time) + # *************************************** + new_element = [] + + for data in results: + onos_results = data['details']['FUNCvirNetL3']['status'] + nbFailures = 0 + for onos in onos_results: + if (onos['Case result'] == "FAIL"): + nbFailures += 1 + new_element.append({'x': data['creation_date'], + 'y1': len(onos_results), + 'y2': nbFailures}) + + test_data.append({'name': "ONOS FUNCvirNetL3 nb tests/nb failures", + 'info': {'type': "graph", + 'xlabel': 'time', + 'y1label': 'Number of tests', + 'y2label': 'Number of failures'}, + 'data_set': new_element}) return test_data @@ -312,6 +382,7 @@ def format_vPing_for_dashboard(results): return test_data + def format_vPing_userdata_for_dashboard(results): """ Post processing for the vPing_userdata test case @@ -347,3 +418,4 @@ def format_vPing_userdata_for_dashboard(results): 'Nb Success': nbTestOk}]}) return test_data + diff --git a/utils/test/result_collection_api/dashboard/promise2Dashboard.py b/utils/test/result_collection_api/dashboard/promise2Dashboard.py new file mode 100644 index 000000000..84f43a7d1 --- /dev/null +++ b/utils/test/result_collection_api/dashboard/promise2Dashboard.py @@ -0,0 +1,103 @@ + #!/usr/bin/python +# +# Copyright (c) 2015 Orange +# morgan.richomme@orange.com +# +# This program and the accompanying materials +# are made available under the terms of the Apache License, Version 2.0 +# which accompanies this distribution, and is available at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# This script is used to build dashboard ready json results +# It may be used for all the test case of the Promise project +# a new method format_<Test_case>_for_dashboard(results) +# v0.1: basic example with methods for odl, Tempest, Rally and vPing +# +import re +import datetime + + +def get_promise_cases(): + """ + get the list of the supported test cases + TODO: update the list when adding a new test case for the dashboard + """ + return ["promise"] + + +def format_promise_for_dashboard(case, results): + """ + generic method calling the method corresponding to the test case + check that the testcase is properly declared first + then build the call to the specific method + """ + if check_promise_case_exist(case): + # note we add _case because testcase and project had the same name + # TODO refactoring...looks fine at the beginning wit only 1 project + # not very ugly now and clearly not optimized... + cmd = "format_" + case + "_case_for_dashboard(results)" + res = eval(cmd) + else: + res = [] + print "Test cases not declared" + return res + + +def check_promise_case_exist(case): + """ + check if the testcase exists + if the test case is not defined or not declared in the list + return False + """ + promise_cases = get_promise_cases() + + if (case is None or case not in promise_cases): + return False + else: + return True + + + + + +def format_promise_case_for_dashboard(results): + """ + Post processing for the promise test case + """ + test_data = [{'description': 'Promise results for Dashboard'}] + # Graph 1: (duration)=f(time) + # *************************************** + new_element = [] + + # default duration 0:00:08.999904 + # consider only seconds => 09 + for data in results: + t = data['details']['duration'] + new_element.append({'x': data['creation_date'], + 'y': t}) + + test_data.append({'name': "Promise duration ", + 'info': {'type': "graph", + 'xlabel': 'time (s)', + 'ylabel': 'duration (s)'}, + 'data_set': new_element}) + + # Graph 2: (Nb test, nb failure)=f(time) + # *************************************** + new_element = [] + + for data in results: + promise_results = data['details'] + new_element.append({'x': data['creation_date'], + 'y1': promise_results['tests'], + 'y2': promise_results['failures']}) + + test_data.append({'name': "Promise nb tests/nb failures", + 'info': {'type': "graph", + 'xlabel': 'time', + 'y1label': 'Number of tests', + 'y2label': 'Number of failures'}, + 'data_set': new_element}) + + return test_data |