diff options
Diffstat (limited to 'ci')
-rwxr-xr-x | ci/deploy.sh | 203 |
1 files changed, 116 insertions, 87 deletions
diff --git a/ci/deploy.sh b/ci/deploy.sh index f6af1b57a..f6c3591d2 100755 --- a/ci/deploy.sh +++ b/ci/deploy.sh @@ -33,11 +33,15 @@ $(notify "$(basename "$0"): Deploy the Fuel@OPNFV MCP stack" 3) $(notify "USAGE:" 2) $(basename "$0") -b base-uri -l lab-name -p pod-name -s deploy-scenario \\ [-B PXE Bridge [-B Mgmt Bridge [-B Internal Bridge [-B Public Bridge]]]] \\ - [-S storage-dir] [-L /path/to/log/file.tar.gz] + [-S storage-dir] [-L /path/to/log/file.tar.gz] [-f] [-F] [-e] [-d] $(notify "OPTIONS:" 2) -b Base-uri for the stack-configuration structure -B Bridge(s): 1st usage = PXE, 2nd = Mgmt, 3rd = Internal, 4th = Public + -d Dry-run + -e Do not launch environment deployment + -f Deploy on existing Salt master + -F Do only create a Salt master -h Print this message and exit -l Lab-name -p Pod-name @@ -45,14 +49,6 @@ $(notify "OPTIONS:" 2) -S Storage dir for VM images -L Deployment log path and file name -$(notify "DISABLED OPTIONS (not yet supported with MCP):" 3) - -d (disabled) Dry-run - -e (disabled) Do not launch environment deployment - -f (disabled) Deploy on existing Salt master - -F (disabled) Do only create a Salt master - -i (disabled) iso url - -T (disabled) Timeout, in minutes, for the deploy. - $(notify "Description:" 2) Deploys the Fuel@OPNFV stack on the indicated lab resource. @@ -61,8 +57,11 @@ It depends on the OPNFV official configuration directory/file structure and provides a fairly simple mechanism to execute a deployment. $(notify "Input parameters to the build script are:" 2) --b Base URI to the configuration directory (needs to be provided in a URI - style, it can be a local resource: file:// or a remote resource http(s)://) +-b Base URI to the configuration directory (needs to be provided in URI style, + it can be a local resource: file:// or a remote resource http(s)://). + A POD Descriptor File (PDF) should be available at: + <base-uri>/labs/<lab-name>/<pod-name>.yaml + The default is './mcp/config'. -B Bridges to be used by deploy script. It can be specified several times, or as a comma separated list of bridges, or both: -B br1 -B br2,br3 First occurence sets PXE Brige, next Mgmt, then Internal and Public. @@ -74,30 +73,24 @@ $(notify "Input parameters to the build script are:" 2) For baremetal deploys, PXE bridge is used for baremetal node provisioning, while "mcpcontrol" is used to provision the infrastructure VMs only. The default is 'pxebr'. +-d Dry-run - Produce deploy config files, but do not execute deploy +-e Do not launch environment deployment +-f Deploy on existing Salt master +-F Do only create a Salt master -h Print this message and exit -L Deployment log path and name, eg. -L /home/jenkins/job.log.tar.gz -l Lab name as defined in the configuration directory, e.g. lf --p POD name as defined in the configuration directory, e.g. pod-1 +-p POD name as defined in the configuration directory, e.g. pod2 -s Deployment-scenario, this points to a short deployment scenario name, which has to be defined in config directory (e.g. os-odl-nofeature-ha). -S Storage dir for VM images, default is mcp/deploy/images -$(notify "Disabled input parameters (not yet supported with MCP):" 3) --d (disabled) Dry-run - Produce deploy config files, but do not execute deploy --f (disabled) Deploy on existing Salt master --e (disabled) Do not launch environment deployment --F (disabled) Do only create a Salt master --T (disabled) Timeout, in minutes, for the deploy. - It defaults to using the DEPLOY_TIMEOUT environment variable when defined. --i (disabled) .iso image to be deployed (needs to be provided in a URI - style, it can be a local resource: file:// or a remote resource http(s)://) - $(notify "[NOTE] sudo & virsh priviledges are needed for this script to run" 3) Example: $(notify "sudo $(basename "$0") \\ - -b file:///home/jenkins/lab-config \\ + -b file:///home/jenkins/securedlab \\ -l lf -p pod2 \\ -s os-odl-nofeature-ha" 2) EOF @@ -132,38 +125,23 @@ clean() { ############################################################################## # BEGIN of variables to customize # -SCRIPT_PATH=$(readlink -f "$(dirname "${BASH_SOURCE[0]}")") -DEPLOY_DIR=$(cd "${SCRIPT_PATH}/../mcp/scripts"; pwd) -STORAGE_DIR=$(cd "${SCRIPT_PATH}/../mcp/deploy/images"; pwd) -RECLASS_CLUSTER_DIR=$(cd "${SCRIPT_PATH}/../mcp/reclass/classes/cluster"; pwd) +REPO_ROOT_PATH=$(readlink -f "$(dirname "${BASH_SOURCE[0]}")/..") +DEPLOY_DIR=$(cd "${REPO_ROOT_PATH}/mcp/scripts"; pwd) +STORAGE_DIR=$(cd "${REPO_ROOT_PATH}/mcp/deploy/images"; pwd) +RECLASS_CLUSTER_DIR=$(cd "${REPO_ROOT_PATH}/mcp/reclass/classes/cluster"; pwd) DEPLOY_TYPE='baremetal' OPNFV_BRIDGES=('pxebr' 'mgmt' 'internal' 'public') URI_REGEXP='(file|https?|ftp)://.*' +BASE_CONFIG_URI="file://${REPO_ROOT_PATH}/mcp/config" -export SSH_KEY=${SSH_KEY:-"/var/lib/opnfv/mcp.rsa"} -export SALT_MASTER=${INSTALLER_IP:-10.20.0.2} -export SALT_MASTER_USER=${SALT_MASTER_USER:-ubuntu} -export MAAS_IP=${MAAS_IP:-${SALT_MASTER%.*}.3} -export MAAS_PXE_NETWORK=${MAAS_PXE_NETWORK:-192.168.11.0} +# Customize deploy workflow +DRY_RUN=${DRY_RUN:-0} +USE_EXISTING_INFRA=${USE_EXISTING_INFRA:-0} +INFRA_CREATION_ONLY=${INFRA_CREATION_ONLY:-0} +NO_DEPLOY_ENVIRONMENT=${NO_DEPLOY_ENVIRONMENT:-0} -# Derivated from above global vars -export MCP_CTRL_NETWORK_ROOTSTR=${SALT_MASTER%.*} -export MAAS_PXE_NETWORK_ROOTSTR=${MAAS_PXE_NETWORK%.*} -export SSH_OPTS="-o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -i ${SSH_KEY}" -export SSH_SALT="${SALT_MASTER_USER}@${SALT_MASTER}" +source "${DEPLOY_DIR}/globals.sh" -# Variables below are disabled for now, to be re-introduced or removed later -set +x -USE_EXISTING_FUEL='' -FUEL_CREATION_ONLY='' -NO_DEPLOY_ENVIRONMENT='' -DRY_RUN=0 -if ! [ -z "${DEPLOY_TIMEOUT}" ]; then - DEPLOY_TIMEOUT="-dt ${DEPLOY_TIMEOUT}" -else - DEPLOY_TIMEOUT="" -fi -set -x # # END of variables to customize ############################################################################## @@ -173,7 +151,7 @@ set -x # set +x OPNFV_BRIDGE_IDX=0 -while getopts "b:B:dfFl:L:p:s:S:T:i:he" OPTION +while getopts "b:B:dfFl:L:p:s:S:he" OPTION do case $OPTION in b) @@ -197,20 +175,16 @@ do IFS=${OIFS} ;; d) - notify '' 3 "${OPTION}"; continue DRY_RUN=1 ;; f) - notify '' 3 "${OPTION}"; continue - USE_EXISTING_FUEL='-nf' + USE_EXISTING_INFRA=1 ;; F) - notify '' 3 "${OPTION}"; continue - FUEL_CREATION_ONLY='-fo' + INFRA_CREATION_ONLY=1 ;; e) - notify '' 3 "${OPTION}"; continue - NO_DEPLOY_ENVIRONMENT='-nde' + NO_DEPLOY_ENVIRONMENT=1 ;; l) TARGET_LAB=${OPTARG} @@ -232,19 +206,6 @@ do STORAGE_DIR="${OPTARG}" fi ;; - T) - notify '' 3 "${OPTION}"; continue - DEPLOY_TIMEOUT="-dt ${OPTARG}" - ;; - i) - notify '' 3 "${OPTION}"; continue - ISO=${OPTARG} - if [[ ! $ISO =~ ${URI_REGEXP} ]]; then - notify "[ERROR] -i $ISO - invalid URI\n" - usage - exit 1 - fi - ;; h) usage exit 0 @@ -300,6 +261,32 @@ if [ "$(uname -i)" = "aarch64" ]; then [ -n "$(command -v yum)" ] && sudo yum install -y --skip-broken vgabios fi +# Clone git submodules and apply our patches +make -C "${REPO_ROOT_PATH}/mcp/patches" deepclean patches-import + +# Convert Pharos-compatible POD Descriptor File (PDF) to reclass model input +PHAROS_GEN_CONFIG_SCRIPT="./pharos/config/utils/generate_config.py" +PHAROS_INSTALLER_ADAPTER="./pharos/config/installers/fuel/pod_config.yml.j2" +BASE_CONFIG_PDF="${BASE_CONFIG_URI}/labs/${TARGET_LAB}/${TARGET_POD}.yaml" +BASE_CONFIG_IDF="${BASE_CONFIG_URI}/labs/${TARGET_LAB}/idf-${TARGET_POD}.yaml" +LOCAL_PDF="${STORAGE_DIR}/$(basename "${BASE_CONFIG_PDF}")" +LOCAL_IDF="${STORAGE_DIR}/$(basename "${BASE_CONFIG_IDF}")" +LOCAL_PDF_RECLASS="${STORAGE_DIR}/pod_config.yml" +if ! curl --create-dirs -o "${LOCAL_PDF}" "${BASE_CONFIG_PDF}"; then + if [ "${DEPLOY_TYPE}" = 'baremetal' ]; then + notify "[ERROR] Could not retrieve PDF (Pod Descriptor File)!\n" 1>&2 + exit 1 + else + notify "[WARN] Could not retrieve PDF (Pod Descriptor File)!\n" 3 + fi +elif ! curl -o "${LOCAL_IDF}" "${BASE_CONFIG_IDF}"; then + notify "[WARN] POD has no IDF (Installer Descriptor File)!\n" 3 +elif ! "${PHAROS_GEN_CONFIG_SCRIPT}" -y "${LOCAL_PDF}" \ + -j "${PHAROS_INSTALLER_ADAPTER}" > "${LOCAL_PDF_RECLASS}"; then + notify "[ERROR] Could not convert PDF to reclass model input!\n" 1>&2 + exit 1 +fi + # Check scenario file existence SCENARIO_DIR="../config/scenario" if [ ! -f "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml" ]; then @@ -319,9 +306,12 @@ if [ ! -f "${SCENARIO_DIR}/defaults-$(uname -i).yaml" ]; then fi # Get required infra deployment data +set +x source lib.sh eval "$(parse_yaml "${SCENARIO_DIR}/defaults-$(uname -i).yaml")" eval "$(parse_yaml "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml")" +eval "$(parse_yaml "${LOCAL_PDF_RECLASS}")" +set -x export CLUSTER_DOMAIN=${cluster_domain} @@ -335,27 +325,66 @@ done # Expand reclass and virsh network templates for tp in "${RECLASS_CLUSTER_DIR}/all-mcp-ocata-common/opnfv/"*.template \ - net_*.template; do envsubst < "${tp}" > "${tp%.template}"; done + net_*.template; do + eval "cat <<-EOF + $(<"${tp}") + EOF" 2> /dev/null > "${tp%.template}" +done -# Infra setup -generate_ssh_key -prepare_vms virtual_nodes "${base_image}" "${STORAGE_DIR}" -create_networks OPNFV_BRIDGES -create_vms virtual_nodes virtual_nodes_ram virtual_nodes_vcpus \ - OPNFV_BRIDGES "${STORAGE_DIR}" -update_mcpcontrol_network -start_vms virtual_nodes -check_connection +# Map PDF networks 'admin', 'mgmt', 'private' and 'public' to bridge names +BR_NAMES=('admin' 'mgmt' 'private' 'public') +BR_NETS=( \ + "${parameters__param_opnfv_maas_pxe_address}" \ + "${parameters__param_opnfv_infra_config_address}" \ + "${parameters__param_opnfv_openstack_compute_node01_tenant_address}" \ + "${parameters__param_opnfv_openstack_compute_node01_external_address}" \ +) +for ((i = 0; i < ${#BR_NETS[@]}; i++)); do + br_jump=$(eval echo "\$parameters__param_opnfv_jump_bridge_${BR_NAMES[i]}") + if [ -n "${br_jump}" ] && [ "${br_jump}" != 'None' ] && \ + [ -d "/sys/class/net/${br_jump}/bridge" ]; then + notify "[OK] Bridge found for '${BR_NAMES[i]}': ${br_jump}\n" 2 + OPNFV_BRIDGES[${i}]="${br_jump}" + elif [ -n "${BR_NETS[i]}" ]; then + bridge=$(ip addr | awk "/${BR_NETS[i]%.*}./ {print \$NF; exit}") + if [ -n "${bridge}" ] && [ -d "/sys/class/net/${bridge}/bridge" ]; then + notify "[OK] Bridge found for net ${BR_NETS[i]%.*}.0: ${bridge}\n" 2 + OPNFV_BRIDGES[${i}]="${bridge}" + fi + fi +done +notify "[NOTE] Using bridges: ${OPNFV_BRIDGES[*]}\n" 2 -./salt.sh +# Infra setup +if [ ${DRY_RUN} -eq 1 ]; then + notify "Dry run, skipping all deployment tasks\n" 2 1>&2 + exit 0 +elif [ ${USE_EXISTING_INFRA} -eq 1 ]; then + notify "Use existing infra\n" 2 1>&2 + check_connection +else + generate_ssh_key + prepare_vms virtual_nodes "${base_image}" "${STORAGE_DIR}" + create_networks OPNFV_BRIDGES + create_vms virtual_nodes virtual_nodes_ram virtual_nodes_vcpus \ + OPNFV_BRIDGES "${STORAGE_DIR}" + update_mcpcontrol_network + start_vms virtual_nodes + check_connection + ./salt.sh "${LOCAL_PDF_RECLASS}" +fi # Openstack cluster setup -for state in "${cluster_states[@]}"; do - notify "STATE: ${state}\n" 2 - # shellcheck disable=SC2086,2029 - ssh ${SSH_OPTS} "ubuntu@${SALT_MASTER}" \ - sudo "/root/fuel/mcp/config/states/${state} || true" -done +if [ ${INFRA_CREATION_ONLY} -eq 1 ] || [ ${NO_DEPLOY_ENVIRONMENT} -eq 1 ]; then + notify "Skip openstack cluster setup\n" 2 +else + for state in "${cluster_states[@]}"; do + notify "STATE: ${state}\n" 2 + # shellcheck disable=SC2086,2029 + ssh ${SSH_OPTS} "${SSH_SALT}" \ + sudo "/root/fuel/mcp/config/states/${state} || true" + done +fi ./log.sh "${DEPLOY_LOG}" |