summaryrefslogtreecommitdiffstats
path: root/jjb
diff options
context:
space:
mode:
Diffstat (limited to 'jjb')
-rwxr-xr-xjjb/apex/apex-build.sh52
-rwxr-xr-xjjb/apex/apex-deploy.sh32
-rwxr-xr-xjjb/apex/apex-download-artifact.sh43
-rwxr-xr-xjjb/apex/apex-iso-verify.sh5
-rwxr-xr-xjjb/apex/apex-upload-artifact.sh30
-rw-r--r--jjb/apex/scenarios.yaml.hidden4
-rw-r--r--jjb/armband/armband-verify-jobs.yml56
-rw-r--r--jjb/compass4nfv/compass-ci-jobs.yml6
-rwxr-xr-xjjb/dovetail/dovetail-run.sh6
-rw-r--r--jjb/fuel/fuel-verify-jobs.yml54
-rw-r--r--jjb/releng/opnfv-docker.yml4
-rwxr-xr-xjjb/yardstick/yardstick-daily.sh1
12 files changed, 159 insertions, 134 deletions
diff --git a/jjb/apex/apex-build.sh b/jjb/apex/apex-build.sh
index cf5999832..aabd20e30 100755
--- a/jjb/apex/apex-build.sh
+++ b/jjb/apex/apex-build.sh
@@ -18,10 +18,18 @@ elif echo $BUILD_TAG | grep "csit" 1> /dev/null; then
export BUILD_ARGS="-r $OPNFV_ARTIFACT_VERSION -c $CACHE_DIRECTORY"
elif [ "$ARTIFACT_VERSION" == "daily" ]; then
export OPNFV_ARTIFACT_VERSION=$(date -u +"%Y-%m-%d")
- export BUILD_ARGS="-r $OPNFV_ARTIFACT_VERSION -c $CACHE_DIRECTORY --iso"
+ if [ "$BRANCH" == 'master' ]; then
+ export BUILD_ARGS="-r $OPNFV_ARTIFACT_VERSION -c $CACHE_DIRECTORY"
+ else
+ export BUILD_ARGS="-r $OPNFV_ARTIFACT_VERSION -c $CACHE_DIRECTORY --iso"
+ fi
else
export OPNFV_ARTIFACT_VERSION=${ARTIFACT_VERSION}
- export BUILD_ARGS="-r $OPNFV_ARTIFACT_VERSION -c $CACHE_DIRECTORY --iso"
+ if [ "$BRANCH" == 'master' ]; then
+ export BUILD_ARGS="-r $OPNFV_ARTIFACT_VERSION -c $CACHE_DIRECTORY"
+ else
+ export BUILD_ARGS="-r $OPNFV_ARTIFACT_VERSION -c $CACHE_DIRECTORY --iso"
+ fi
fi
# Temporary hack until we fix apex build script
@@ -46,7 +54,7 @@ echo "Cache Directory Contents:"
echo "-------------------------"
ls -al $CACHE_DIRECTORY
-if [[ "$BUILD_ARGS" =~ '--iso' ]]; then
+if [[ "$BUILD_ARGS" =~ '--iso' && "$BRANCH" != 'master' ]]; then
mkdir -p /tmp/apex-iso/
rm -f /tmp/apex-iso/*.iso
cp -f $BUILD_DIRECTORY/../.build/release/OPNFV-CentOS-7-x86_64-$OPNFV_ARTIFACT_VERSION.iso /tmp/apex-iso/
@@ -54,18 +62,32 @@ fi
if ! echo $ARTIFACT_VERSION | grep "dev" 1> /dev/null; then
echo "Writing opnfv.properties file"
- # save information regarding artifact into file
- (
- echo "OPNFV_ARTIFACT_VERSION=$OPNFV_ARTIFACT_VERSION"
- echo "OPNFV_GIT_URL=$(git config --get remote.origin.url)"
- echo "OPNFV_GIT_SHA1=$(git rev-parse HEAD)"
- echo "OPNFV_ARTIFACT_URL=$GS_URL/opnfv-$OPNFV_ARTIFACT_VERSION.iso"
- echo "OPNFV_ARTIFACT_SHA512SUM=$(sha512sum $BUILD_DIRECTORY/../.build/release/OPNFV-CentOS-7-x86_64-$OPNFV_ARTIFACT_VERSION.iso | cut -d' ' -f1)"
- echo "OPNFV_SRPM_URL=$GS_URL/opnfv-apex-$RPM_VERSION.src.rpm"
- echo "OPNFV_RPM_URL=$GS_URL/opnfv-apex-$RPM_VERSION.noarch.rpm"
- echo "OPNFV_RPM_SHA512SUM=$(sha512sum $BUILD_DIRECTORY/../.build/noarch/opnfv-apex-$RPM_VERSION.noarch.rpm | cut -d' ' -f1)"
- echo "OPNFV_BUILD_URL=$BUILD_URL"
- ) > $WORKSPACE/opnfv.properties
+ if [ "$BRANCH" != master ]; then
+ # save information regarding artifact into file
+ (
+ echo "OPNFV_ARTIFACT_VERSION=$OPNFV_ARTIFACT_VERSION"
+ echo "OPNFV_GIT_URL=$(git config --get remote.origin.url)"
+ echo "OPNFV_GIT_SHA1=$(git rev-parse HEAD)"
+ echo "OPNFV_ARTIFACT_URL=$GS_URL/opnfv-$OPNFV_ARTIFACT_VERSION.iso"
+ echo "OPNFV_ARTIFACT_SHA512SUM=$(sha512sum $BUILD_DIRECTORY/../.build/release/OPNFV-CentOS-7-x86_64-$OPNFV_ARTIFACT_VERSION.iso | cut -d' ' -f1)"
+ echo "OPNFV_SRPM_URL=$GS_URL/opnfv-apex-$RPM_VERSION.src.rpm"
+ echo "OPNFV_RPM_URL=$GS_URL/opnfv-apex-$RPM_VERSION.noarch.rpm"
+ echo "OPNFV_RPM_SHA512SUM=$(sha512sum $BUILD_DIRECTORY/../.build/noarch/opnfv-apex-$RPM_VERSION.noarch.rpm | cut -d' ' -f1)"
+ echo "OPNFV_BUILD_URL=$BUILD_URL"
+ ) > $WORKSPACE/opnfv.properties
+ else
+ # save information regarding artifact into file
+ # we only generate the python package for master
+ (
+ echo "OPNFV_ARTIFACT_VERSION=$OPNFV_ARTIFACT_VERSION"
+ echo "OPNFV_GIT_URL=$(git config --get remote.origin.url)"
+ echo "OPNFV_GIT_SHA1=$(git rev-parse HEAD)"
+ echo "OPNFV_SRPM_URL=$GS_URL/python34-opnfv-apex-$RPM_VERSION.src.rpm"
+ echo "OPNFV_RPM_URL=$GS_URL/python34-opnfv-apex-$RPM_VERSION.noarch.rpm"
+ echo "OPNFV_RPM_SHA512SUM=$(sha512sum $BUILD_DIRECTORY/../.build/noarch/python34-opnfv-apex-$RPM_VERSION.noarch.rpm | cut -d' ' -f1)"
+ echo "OPNFV_BUILD_URL=$BUILD_URL"
+ ) > $WORKSPACE/opnfv.properties
+ fi
fi
echo "--------------------------------------------------------"
echo "Done!"
diff --git a/jjb/apex/apex-deploy.sh b/jjb/apex/apex-deploy.sh
index 958e4bf5e..35c2b8544 100755
--- a/jjb/apex/apex-deploy.sh
+++ b/jjb/apex/apex-deploy.sh
@@ -18,20 +18,26 @@ else
fi
# Dev or RPM/ISO build
-# For upstream deployments we currently only use git repo and not RPM
-# Need to decide after Fraser if we want to use RPM or not for upstream
-if [[ "$ARTIFACT_VERSION" =~ dev || "$DEPLOY_SCENARIO" =~ "upstream" ]]; then
+if [[ "$ARTIFACT_VERSION" =~ dev ]]; then
# Settings for deploying from git workspace
DEPLOY_SETTINGS_DIR="${WORKSPACE}/config/deploy"
NETWORK_SETTINGS_DIR="${WORKSPACE}/config/network"
- DEPLOY_CMD="opnfv-deploy --image-dir ${WORKSPACE}/.build"
CLEAN_CMD="opnfv-clean"
- RESOURCES="${WORKSPACE}/.build/"
+ # if we are using master, then we are downloading/caching upstream images
+ # we want to use that built in mechanism to avoid re-downloading every job
+ # so we use a dedicated folder to hold the upstream cache
+ UPSTREAM_CACHE=$HOME/upstream_cache
+ if [ "$BRANCH" == 'master' ]; then
+ mkdir -p ${UPSTREAM_CACHE}
+ RESOURCES=$UPSTREAM_CACHE
+ else
+ RESOURCES="${WORKSPACE}/.build/"
+ fi
CONFIG="${WORKSPACE}/build"
BASE=$CONFIG
IMAGES=$RESOURCES
LIB="${WORKSPACE}/lib"
-
+ DEPLOY_CMD="opnfv-deploy --image-dir ${RESOURCES}"
# Ensure artifacts were downloaded and extracted correctly
# TODO(trozet) add verification here
@@ -43,9 +49,17 @@ if [[ "$ARTIFACT_VERSION" =~ dev || "$DEPLOY_SCENARIO" =~ "upstream" ]]; then
else
DEPLOY_SETTINGS_DIR="/etc/opnfv-apex/"
NETWORK_SETTINGS_DIR="/etc/opnfv-apex/"
- DEPLOY_CMD="opnfv-deploy"
CLEAN_CMD="opnfv-clean"
- RESOURCES="/var/opt/opnfv/images"
+ # set to use different directory here because upon RPM removal this
+ # directory will be wiped in daily
+ UPSTREAM_CACHE=$HOME/upstream_cache
+ if [ "$BRANCH" == 'master' ]; then
+ mkdir -p ${UPSTREAM_CACHE}
+ RESOURCES=$UPSTREAM_CACHE
+ else
+ RESOURCES="/var/opt/opnfv/images"
+ fi
+ DEPLOY_CMD="opnfv-deploy --image-dir ${RESOURCES}"
CONFIG="/var/opt/opnfv"
BASE=$CONFIG
IMAGES=$RESOURCES
@@ -123,7 +137,7 @@ else
DEPLOY_CMD="${DEPLOY_CMD} -i ${INVENTORY_FILE}"
fi
-if [[ "$DEPLOY_SCENARIO" =~ "upstream" ]]; then
+if [[ "$BRANCH" == "master" ]]; then
echo "Upstream deployment detected"
DEPLOY_CMD="${DEPLOY_CMD} --upstream"
fi
diff --git a/jjb/apex/apex-download-artifact.sh b/jjb/apex/apex-download-artifact.sh
index 6c0f8fe26..3efe1cbc7 100755
--- a/jjb/apex/apex-download-artifact.sh
+++ b/jjb/apex/apex-download-artifact.sh
@@ -17,19 +17,20 @@ else
echo "Deploy scenario: ${DEPLOY_SCENARIO}"
fi
-# if upstream we do not need to download anything
-if [[ "$DEPLOY_SCENARIO" =~ upstream ]]; then
- echo "Upstream deployment detected, skipping download artifact"
-elif [[ "$ARTIFACT_VERSION" =~ dev ]]; then
- # dev build
- GERRIT_PATCHSET_NUMBER=$(echo $GERRIT_REFSPEC | grep -Eo '[0-9]+$')
- export OPNFV_ARTIFACT_VERSION="dev${GERRIT_CHANGE_NUMBER}_${GERRIT_PATCHSET_NUMBER}"
- # get build artifact
- pushd ${BUILD_DIRECTORY} > /dev/null
- echo "Downloading packaged dev build: apex-${OPNFV_ARTIFACT_VERSION}.tar.gz"
- curl --fail -s -o $BUILD_DIRECTORY/apex-${OPNFV_ARTIFACT_VERSION}.tar.gz http://$GS_URL/apex-${OPNFV_ARTIFACT_VERSION}.tar.gz
- tar -xvf apex-${OPNFV_ARTIFACT_VERSION}.tar.gz
- popd > /dev/null
+if [[ "$ARTIFACT_VERSION" =~ dev ]]; then
+ if [ "$BRANCH" == 'master' ]; then
+ echo "Skipping download of artifacts for master branch"
+ else
+ # dev build
+ GERRIT_PATCHSET_NUMBER=$(echo $GERRIT_REFSPEC | grep -Eo '[0-9]+$')
+ export OPNFV_ARTIFACT_VERSION="dev${GERRIT_CHANGE_NUMBER}_${GERRIT_PATCHSET_NUMBER}"
+ # get build artifact
+ pushd ${BUILD_DIRECTORY} > /dev/null
+ echo "Downloading packaged dev build: apex-${OPNFV_ARTIFACT_VERSION}.tar.gz"
+ curl --fail -s -o $BUILD_DIRECTORY/apex-${OPNFV_ARTIFACT_VERSION}.tar.gz http://$GS_URL/apex-${OPNFV_ARTIFACT_VERSION}.tar.gz
+ tar -xvf apex-${OPNFV_ARTIFACT_VERSION}.tar.gz
+ popd > /dev/null
+ fi
else
echo "Will use RPMs..."
@@ -45,20 +46,16 @@ else
RPM_INSTALL_PATH=$(echo "http://"$OPNFV_RPM_URL | sed 's/\/'"$(basename $OPNFV_RPM_URL)"'//')
RPM_LIST=$(basename $OPNFV_RPM_URL)
- # find version of RPM
- VERSION_EXTENSION=$(echo $(basename $RPM_LIST) | grep -Eo '[0-9]+\.[0-9]+-([0-9]{8}|[a-z]+-[0-9]\.[0-9]+)')
- # build RPM List which already includes base Apex RPM
- RPM_LIST+=" opnfv-apex-undercloud-${VERSION_EXTENSION}.noarch.rpm"
-
- # add back legacy support for danube
- if [ "$BRANCH" == 'stable/danube' ]; then
- RPM_LIST+=" opnfv-apex-common-${VERSION_EXTENSION}.noarch.rpm"
- else
+ if [ "$BRANCH" != 'master' ]; then
+ # find version of RPM
+ VERSION_EXTENSION=$(echo $(basename $RPM_LIST) | grep -Eo '[0-9]+\.[0-9]+-([0-9]{8}|[a-z]+-[0-9]\.[0-9]+)')
+ # build RPM List which already includes base Apex RPM
+ RPM_LIST+=" opnfv-apex-undercloud-${VERSION_EXTENSION}.noarch.rpm"
RPM_LIST+=" python34-opnfv-apex-${VERSION_EXTENSION}.noarch.rpm"
fi
# remove old / install new RPMs
- if rpm -q opnfv-apex > /dev/null; then
+ if rpm -q python34-opnfv-apex > /dev/null; then
INSTALLED_RPMS=$(rpm -qa | grep apex)
if [ -n "$INSTALLED_RPMS" ]; then
sudo yum remove -y ${INSTALLED_RPMS}
diff --git a/jjb/apex/apex-iso-verify.sh b/jjb/apex/apex-iso-verify.sh
index 4faeb607e..f34937619 100755
--- a/jjb/apex/apex-iso-verify.sh
+++ b/jjb/apex/apex-iso-verify.sh
@@ -8,6 +8,11 @@ echo "Starting the Apex iso verify."
echo "--------------------------------------------------------"
echo
+if [ "$BRANCH" == 'master' ]; then
+ echo "Skipping Apex iso verify for master branch"
+ exit 0
+fi
+
# Must be RPMs/ISO
echo "Downloading latest properties file"
diff --git a/jjb/apex/apex-upload-artifact.sh b/jjb/apex/apex-upload-artifact.sh
index 4037d25ad..8743368fb 100755
--- a/jjb/apex/apex-upload-artifact.sh
+++ b/jjb/apex/apex-upload-artifact.sh
@@ -109,8 +109,8 @@ fi
if [ "$ARTIFACT_TYPE" == 'snapshot' ]; then
uploadsnap
elif [ "$ARTIFACT_TYPE" == 'iso' ]; then
- if [[ "$ARTIFACT_VERSION" =~ dev ]]; then
- echo "Skipping artifact upload for ${ARTIFACT_TYPE} due to dev build"
+ if [[ "$ARTIFACT_VERSION" =~ dev || "$BRANCH" == 'master' ]]; then
+ echo "Skipping ISO artifact upload for ${ARTIFACT_TYPE} due to dev/master build"
exit 0
fi
if [[ -n "$SIGN_ARTIFACT" && "$SIGN_ARTIFACT" == "true" ]]; then
@@ -119,20 +119,28 @@ elif [ "$ARTIFACT_TYPE" == 'iso' ]; then
uploadiso
elif [ "$ARTIFACT_TYPE" == 'rpm' ]; then
if [[ "$ARTIFACT_VERSION" =~ dev ]]; then
- echo "dev build detected, will upload image tarball"
- ARTIFACT_TYPE=tarball
- uploadimages
+ if [ "$BRANCH" == 'master' ]; then
+ echo "will not upload artifacts, master uses upstream"
+ ARTIFACT_TYPE=none
+ else
+ echo "dev build detected, will upload image tarball"
+ ARTIFACT_TYPE=tarball
+ uploadimages
+ fi
else
RPM_INSTALL_PATH=$BUILD_DIRECTORY/noarch
+ # RPM URL should be python package for master, and is only package we need
RPM_LIST=$RPM_INSTALL_PATH/$(basename $OPNFV_RPM_URL)
- VERSION_EXTENSION=$(echo $(basename $OPNFV_RPM_URL) | sed 's/opnfv-apex-//')
- RPM_LIST+=" ${RPM_INSTALL_PATH}/opnfv-apex-undercloud-${VERSION_EXTENSION}"
- RPM_LIST+=" ${RPM_INSTALL_PATH}/python34-opnfv-apex-${VERSION_EXTENSION}"
SRPM_INSTALL_PATH=$BUILD_DIRECTORY
SRPM_LIST=$SRPM_INSTALL_PATH/$(basename $OPNFV_SRPM_URL)
- VERSION_EXTENSION=$(echo $(basename $OPNFV_SRPM_URL) | sed 's/opnfv-apex-//')
- SRPM_LIST+=" ${SRPM_INSTALL_PATH}/opnfv-apex-undercloud-${VERSION_EXTENSION}"
- SRPM_LIST+=" ${SRPM_INSTALL_PATH}/python34-opnfv-apex-${VERSION_EXTENSION}"
+ if [ "$BRANCH" != 'master' ]; then
+ VERSION_EXTENSION=$(echo $(basename $OPNFV_RPM_URL) | sed 's/opnfv-apex-//')
+ RPM_LIST+=" ${RPM_INSTALL_PATH}/opnfv-apex-undercloud-${VERSION_EXTENSION}"
+ RPM_LIST+=" ${RPM_INSTALL_PATH}/python34-opnfv-apex-${VERSION_EXTENSION}"
+ VERSION_EXTENSION=$(echo $(basename $OPNFV_SRPM_URL) | sed 's/opnfv-apex-//')
+ SRPM_LIST+=" ${SRPM_INSTALL_PATH}/opnfv-apex-undercloud-${VERSION_EXTENSION}"
+ SRPM_LIST+=" ${SRPM_INSTALL_PATH}/python34-opnfv-apex-${VERSION_EXTENSION}"
+ fi
if [[ -n "$SIGN_ARTIFACT" && "$SIGN_ARTIFACT" == "true" ]]; then
signrpm
diff --git a/jjb/apex/scenarios.yaml.hidden b/jjb/apex/scenarios.yaml.hidden
index 91cd789c5..2650eafbb 100644
--- a/jjb/apex/scenarios.yaml.hidden
+++ b/jjb/apex/scenarios.yaml.hidden
@@ -1,6 +1,6 @@
master:
- - 'os-odl-master_upstream-noha'
- - 'os-odl-queens_upstream-noha'
+ - 'os-odl-nofeature-noha'
+ - 'os-odl-queens-noha'
fraser:
- 'os-nosdn-nofeature-noha'
- 'os-nosdn-nofeature-ha'
diff --git a/jjb/armband/armband-verify-jobs.yml b/jjb/armband/armband-verify-jobs.yml
index c800872c7..08cf3c084 100644
--- a/jjb/armband/armband-verify-jobs.yml
+++ b/jjb/armband/armband-verify-jobs.yml
@@ -21,12 +21,8 @@
# patch verification phases
#####################################
phase:
- - 'basic':
- slave-label: 'armband-virtual'
- 'deploy-virtual':
slave-label: 'armband-virtual'
- - 'smoke-test':
- slave-label: 'armband-virtual'
#####################################
# jobs
#####################################
@@ -105,15 +101,18 @@
installer: '{installer}'
- '{installer}-defaults':
gs-pathname: '{gs-pathname}'
+ - string:
+ name: DEPLOY_SCENARIO
+ default: 'os-nosdn-nofeature-ha'
builders:
- description-setter:
description: "Built on $NODE_NAME"
- multijob:
- name: basic
+ name: deploy-virtual
condition: SUCCESSFUL
projects:
- - name: 'armband-verify-basic-{stream}'
+ - name: 'armband-verify-deploy-virtual-{stream}'
current-parameters: false
predefined-parameters: |
BRANCH=$BRANCH
@@ -123,33 +122,39 @@
node-parameters: false
kill-phase-on: FAILURE
abort-all-job: true
+
- multijob:
- name: deploy-virtual
+ name: smoke-test
condition: SUCCESSFUL
projects:
- - name: 'armband-verify-deploy-virtual-{stream}'
+ # Use Functest job definition from jjb/functest/functest-daily-jobs
+ - name: 'functest-fuel-armband-virtual-suite-{stream}'
current-parameters: false
predefined-parameters: |
+ FUNCTEST_MODE=tier
+ FUNCTEST_TIER=healthcheck
+ # Should be in sync with fuel-deploy.sh default scenario
+ DEPLOY_SCENARIO=$DEPLOY_SCENARIO
BRANCH=$BRANCH
GERRIT_REFSPEC=$GERRIT_REFSPEC
GERRIT_CHANGE_NUMBER=$GERRIT_CHANGE_NUMBER
GERRIT_CHANGE_COMMIT_MESSAGE=$GERRIT_CHANGE_COMMIT_MESSAGE
- node-parameters: false
- kill-phase-on: FAILURE
+ node-parameters: true
+ kill-phase-on: NEVER
abort-all-job: true
- - multijob:
- name: smoke-test
- condition: SUCCESSFUL
- projects:
- - name: 'armband-verify-smoke-test-{stream}'
+ - name: 'functest-fuel-armband-virtual-suite-{stream}'
current-parameters: false
predefined-parameters: |
+ FUNCTEST_MODE=testcase
+ FUNCTEST_SUITE_NAME=vping_ssh
+ # Should be in sync with fuel-deploy.sh default scenario
+ DEPLOY_SCENARIO=$DEPLOY_SCENARIO
BRANCH=$BRANCH
GERRIT_REFSPEC=$GERRIT_REFSPEC
GERRIT_CHANGE_NUMBER=$GERRIT_CHANGE_NUMBER
GERRIT_CHANGE_COMMIT_MESSAGE=$GERRIT_CHANGE_COMMIT_MESSAGE
- node-parameters: false
- kill-phase-on: FAILURE
+ node-parameters: true
+ kill-phase-on: NEVER
abort-all-job: true
- job-template:
@@ -170,7 +175,6 @@
use-build-blocker: true
blocking-jobs:
- 'armband-verify-deploy-.*'
- - 'armband-verify-test-.*'
block-level: 'NODE'
scm:
@@ -200,23 +204,7 @@
# builder macros
#####################################
- builder:
- name: 'armband-verify-basic-macro'
- builders:
- - shell: |
- #!/bin/bash
-
- echo "Not activated!"
-
-- builder:
name: 'armband-verify-deploy-virtual-macro'
builders:
- shell:
!include-raw: ../fuel/fuel-deploy.sh
-
-- builder:
- name: 'armband-verify-smoke-test-macro'
- builders:
- - shell: |
- #!/bin/bash
-
- echo "Not activated!"
diff --git a/jjb/compass4nfv/compass-ci-jobs.yml b/jjb/compass4nfv/compass-ci-jobs.yml
index 5f82803b6..88fb02605 100644
--- a/jjb/compass4nfv/compass-ci-jobs.yml
+++ b/jjb/compass4nfv/compass-ci-jobs.yml
@@ -276,7 +276,7 @@
label: '{scenario}'
steps:
- trigger-builds:
- - project: 'bottlenecks-compass-posca_stress_ping-{pod}-daily-{stream}'
+ - project: 'bottlenecks-compass-posca_factor_multistack_storage_parallel-{pod}-daily-{stream}'
current-parameters: false
predefined-parameters:
DEPLOY_SCENARIO={scenario}
@@ -286,7 +286,7 @@
build-step-failure-threshold: 'never'
failure-threshold: 'never'
unstable-threshold: 'FAILURE'
- - project: 'bottlenecks-compass-posca_factor_multistack_storage_parallel-{pod}-daily-{stream}'
+ - project: 'bottlenecks-compass-posca_factor_soak_throughputs-{pod}-daily-{stream}'
current-parameters: false
predefined-parameters:
DEPLOY_SCENARIO={scenario}
@@ -296,7 +296,7 @@
build-step-failure-threshold: 'never'
failure-threshold: 'never'
unstable-threshold: 'FAILURE'
- - project: 'bottlenecks-compass-posca_factor_soak_throughputs-{pod}-daily-{stream}'
+ - project: 'bottlenecks-compass-posca_stress_ping-{pod}-daily-{stream}'
current-parameters: false
predefined-parameters:
DEPLOY_SCENARIO={scenario}
diff --git a/jjb/dovetail/dovetail-run.sh b/jjb/dovetail/dovetail-run.sh
index f74b10754..a5148cde9 100755
--- a/jjb/dovetail/dovetail-run.sh
+++ b/jjb/dovetail/dovetail-run.sh
@@ -323,9 +323,9 @@ docker exec $container_id ${run_cmd}
sudo cp -r ${DOVETAIL_HOME}/results ./
# To make sure the file owner is the current user, for the copied results files in the above line
# if not, there will be error when next time to wipe workspace
-# CURRENT_USER=${SUDO_USER:-$USER}
-# PRIMARY_GROUP=$(id -gn $CURRENT_USER)
-# sudo chown -R ${CURRENT_USER}:${PRIMARY_GROUP} ${WORKSPACE}/results
+CURRENT_USER=${SUDO_USER:-$USER}
+PRIMARY_GROUP=$(id -gn $CURRENT_USER)
+sudo chown -R ${CURRENT_USER}:${PRIMARY_GROUP} ${WORKSPACE}/results
#remove useless files to save disk space
sudo rm -rf ./results/workspace
diff --git a/jjb/fuel/fuel-verify-jobs.yml b/jjb/fuel/fuel-verify-jobs.yml
index d81b0ce92..bef6a5094 100644
--- a/jjb/fuel/fuel-verify-jobs.yml
+++ b/jjb/fuel/fuel-verify-jobs.yml
@@ -21,12 +21,8 @@
# patch verification phases
#####################################
phase:
- - 'basic':
- slave-label: 'fuel-virtual'
- 'deploy-virtual':
slave-label: 'fuel-virtual'
- - 'smoke-test':
- slave-label: 'fuel-virtual'
#####################################
# jobs
#####################################
@@ -105,15 +101,18 @@
installer: '{installer}'
- '{installer}-defaults':
gs-pathname: '{gs-pathname}'
+ - string:
+ name: DEPLOY_SCENARIO
+ default: 'os-nosdn-nofeature-ha'
builders:
- description-setter:
description: "Built on $NODE_NAME"
- multijob:
- name: basic
+ name: deploy-virtual
condition: SUCCESSFUL
projects:
- - name: 'fuel-verify-basic-{stream}'
+ - name: 'fuel-verify-deploy-virtual-{stream}'
current-parameters: false
predefined-parameters: |
BRANCH=$BRANCH
@@ -123,33 +122,37 @@
node-parameters: false
kill-phase-on: FAILURE
abort-all-job: true
+
- multijob:
- name: deploy-virtual
+ name: smoke-test
condition: SUCCESSFUL
projects:
- - name: 'fuel-verify-deploy-virtual-{stream}'
+ # Use Functest job definition from jjb/functest/functest-daily-jobs
+ - name: 'functest-fuel-virtual-suite-{stream}'
current-parameters: false
predefined-parameters: |
+ FUNCTEST_MODE=tier
+ FUNCTEST_TIER=healthcheck
+ DEPLOY_SCENARIO=$DEPLOY_SCENARIO
BRANCH=$BRANCH
GERRIT_REFSPEC=$GERRIT_REFSPEC
GERRIT_CHANGE_NUMBER=$GERRIT_CHANGE_NUMBER
GERRIT_CHANGE_COMMIT_MESSAGE=$GERRIT_CHANGE_COMMIT_MESSAGE
- node-parameters: false
- kill-phase-on: FAILURE
+ node-parameters: true
+ kill-phase-on: NEVER
abort-all-job: true
- - multijob:
- name: smoke-test
- condition: SUCCESSFUL
- projects:
- - name: 'fuel-verify-smoke-test-{stream}'
+ - name: 'functest-fuel-virtual-suite-{stream}'
current-parameters: false
predefined-parameters: |
+ FUNCTEST_MODE=testcase
+ FUNCTEST_SUITE_NAME=vping_ssh
+ DEPLOY_SCENARIO=$DEPLOY_SCENARIO
BRANCH=$BRANCH
GERRIT_REFSPEC=$GERRIT_REFSPEC
GERRIT_CHANGE_NUMBER=$GERRIT_CHANGE_NUMBER
GERRIT_CHANGE_COMMIT_MESSAGE=$GERRIT_CHANGE_COMMIT_MESSAGE
- node-parameters: false
- kill-phase-on: FAILURE
+ node-parameters: true
+ kill-phase-on: NEVER
abort-all-job: true
- job-template:
@@ -170,7 +173,6 @@
use-build-blocker: true
blocking-jobs:
- 'fuel-verify-deploy-.*'
- - 'fuel-verify-test-.*'
block-level: 'NODE'
scm:
@@ -200,23 +202,7 @@
# builder macros
#####################################
- builder:
- name: 'fuel-verify-basic-macro'
- builders:
- - shell: |
- #!/bin/bash
-
- echo "Not activated!"
-
-- builder:
name: 'fuel-verify-deploy-virtual-macro'
builders:
- shell:
!include-raw: ./fuel-deploy.sh
-
-- builder:
- name: 'fuel-verify-smoke-test-macro'
- builders:
- - shell: |
- #!/bin/bash
-
- echo "Not activated!"
diff --git a/jjb/releng/opnfv-docker.yml b/jjb/releng/opnfv-docker.yml
index 0fe76f029..b7d1ce66f 100644
--- a/jjb/releng/opnfv-docker.yml
+++ b/jjb/releng/opnfv-docker.yml
@@ -219,6 +219,10 @@
<<: *danube
<<: *other-receivers
# projects with jobs for fraser
+ - 'bottlenecks':
+ project: 'bottlenecks'
+ <<: *fraser
+ <<: *other-receivers
- 'clover':
project: 'clover'
dockerdir: '.'
diff --git a/jjb/yardstick/yardstick-daily.sh b/jjb/yardstick/yardstick-daily.sh
index 2fd68484e..783c64ee1 100755
--- a/jjb/yardstick/yardstick-daily.sh
+++ b/jjb/yardstick/yardstick-daily.sh
@@ -63,6 +63,7 @@ if [ "$(uname -m)" = 'aarch64' ]; then
fi
echo "Yardstick: Pulling image ${DOCKER_REPO}:${DOCKER_TAG}"
docker pull ${DOCKER_REPO}:$DOCKER_TAG >$redirect
+docker images
# map log directory
branch=${BRANCH##*/}