blob: 441bf9059d30f30ff2bc5a8574145c5fbef3349b (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
|
#!/bin/bash
set -o errexit
set -o nounset
set -o pipefail
IPV6_FLAG=False
ALLINONE_FLAG=False
CSIT_ENV_FLAG=False
# log info to console
echo "Starting the Apex deployment."
echo "--------------------------------------------------------"
echo
if [ -z "$DEPLOY_SCENARIO" ]; then
echo "Deploy scenario not set!"
exit 1
else
echo "Deploy scenario: ${DEPLOY_SCENARIO}"
fi
# Dev or RPM/ISO build
if [[ "$ARTIFACT_VERSION" =~ dev ]]; then
# Settings for deploying from git workspace
DEPLOY_SETTINGS_DIR="${WORKSPACE}/config/deploy"
NETWORK_SETTINGS_DIR="${WORKSPACE}/config/network"
CLEAN_CMD="opnfv-clean"
# if we are using master, then we are downloading/caching upstream images
# we want to use that built in mechanism to avoid re-downloading every job
# so we use a dedicated folder to hold the upstream cache
UPSTREAM_CACHE=$HOME/upstream_cache
if [ "$BRANCH" == 'master' ]; then
mkdir -p ${UPSTREAM_CACHE}
RESOURCES=$UPSTREAM_CACHE
else
RESOURCES="${WORKSPACE}/.build/"
fi
CONFIG="${WORKSPACE}/build"
BASE=$CONFIG
IMAGES=$RESOURCES
LIB="${WORKSPACE}/lib"
DEPLOY_CMD="opnfv-deploy --image-dir ${RESOURCES}"
# Ensure artifacts were downloaded and extracted correctly
# TODO(trozet) add verification here
# Install dev build
sudo rm -rf /tmp/.build
mv -f .build /tmp/
sudo pip3 install --upgrade --force-reinstall .
mv -f /tmp/.build ${WORKSPACE}/
else
DEPLOY_SETTINGS_DIR="/etc/opnfv-apex/"
NETWORK_SETTINGS_DIR="/etc/opnfv-apex/"
CLEAN_CMD="opnfv-clean"
# set to use different directory here because upon RPM removal this
# directory will be wiped in daily
UPSTREAM_CACHE=$HOME/upstream_cache
if [ "$BRANCH" == 'master' ]; then
mkdir -p ${UPSTREAM_CACHE}
RESOURCES=$UPSTREAM_CACHE
else
RESOURCES="/var/opt/opnfv/images"
fi
DEPLOY_CMD="opnfv-deploy --image-dir ${RESOURCES}"
CONFIG="/var/opt/opnfv"
BASE=$CONFIG
IMAGES=$RESOURCES
LIB="/var/opt/opnfv/lib"
sudo mkdir -p /var/log/apex
sudo chmod 777 /var/log/apex
cd /var/log/apex
fi
# Install Dependencies
# Make sure python34 dependencies are installed
dependencies="epel-release python34 python34-devel libvirt-devel python34-pip \
ansible python34-PyYAML python34-jinja2 python34-setuptools python-tox ansible"
for dep_pkg in $dependencies; do
if ! rpm -q ${dep_pkg} > /dev/null; then
if ! sudo yum install -y ${dep_pkg}; then
echo "Failed to install ${dep_pkg}"
exit 1
fi
fi
done
if [[ "$JOB_NAME" =~ "virtual" ]]; then
# Make sure ipxe-roms-qemu package is updated to latest.
# This package is needed for multi virtio nic PXE boot in virtual environment.
sudo yum update -y ipxe-roms-qemu
fi
if [ "$OPNFV_CLEAN" == 'yes' ]; then
if sudo test -e '/root/inventory/pod_settings.yaml'; then
clean_opts='-i /root/inventory/pod_settings.yaml'
else
clean_opts=''
fi
sudo ${CLEAN_CMD} ${clean_opts}
fi
# These are add-ons to regular scenarios where you can do like
# os-nosdn-nofeature-noha-ipv6, or os-nosdn-nofeature-noha-allinone
if echo ${DEPLOY_SCENARIO} | grep ipv6; then
IPV6_FLAG=True
DEPLOY_SCENARIO=$(echo ${DEPLOY_SCENARIO} | sed 's/-ipv6//')
echo "INFO: IPV6 Enabled"
fi
if echo ${DEPLOY_SCENARIO} | grep allinone; then
ALLINONE_FLAG=True
DEPLOY_SCENARIO=$(echo ${DEPLOY_SCENARIO} | sed 's/-allinone//')
echo "INFO: All in one deployment detected"
fi
if echo ${DEPLOY_SCENARIO} | grep csit; then
CSIT_ENV_FLAG=True
DEPLOY_SCENARIO=$(echo ${DEPLOY_SCENARIO} | sed 's/-csit//')
echo "INFO: CSIT env requested in deploy scenario"
fi
echo "Deploy Scenario set to ${DEPLOY_SCENARIO}"
DEPLOY_FILE="${DEPLOY_SETTINGS_DIR}/${DEPLOY_SCENARIO}.yaml"
if [ ! -e "$DEPLOY_FILE" ]; then
echo "ERROR: Required settings file missing: Deploy settings file ${DEPLOY_FILE}"
fi
if [[ "$JOB_NAME" =~ "virtual" ]]; then
# settings for virtual deployment
DEPLOY_CMD="${DEPLOY_CMD} -v"
if [[ "${DEPLOY_SCENARIO}" =~ fdio|ovs ]]; then
DEPLOY_CMD="${DEPLOY_CMD} --virtual-default-ram 12 --virtual-compute-ram 7"
fi
if [[ "$ALLINONE_FLAG" == "True" ]]; then
DEPLOY_CMD="${DEPLOY_CMD} --virtual-computes 0"
elif [[ "$PROMOTE" == "True" ]]; then
DEPLOY_CMD="${DEPLOY_CMD} --virtual-computes 2"
fi
if [[ "$PROMOTE" == "True" || "$CSIT_ENV_FLAG" == "True" ]]; then
if [[ "$DEPLOY_SCENARIO" =~ "queens" ]]; then
CSIT_ENV="csit-queens-environment.yaml"
else
CSIT_ENV="csit-environment.yaml"
fi
DEPLOY_CMD="${DEPLOY_CMD} -e ${CSIT_ENV}"
fi
else
# settings for bare metal deployment
NETWORK_SETTINGS_DIR="/root/network"
INVENTORY_FILE="/root/inventory/pod_settings.yaml"
if ! sudo test -e "$INVENTORY_FILE"; then
echo "ERROR: Required settings file missing: Inventory settings file ${INVENTORY_FILE}"
exit 1
fi
# include inventory file for bare metal deployment
DEPLOY_CMD="${DEPLOY_CMD} -i ${INVENTORY_FILE}"
fi
if [[ "$BRANCH" == "master" ]]; then
echo "Upstream deployment detected"
DEPLOY_CMD="${DEPLOY_CMD} --upstream"
fi
if [ "$IPV6_FLAG" == "True" ]; then
NETWORK_FILE="${NETWORK_SETTINGS_DIR}/network_settings_v6.yaml"
elif [[ "$PROMOTE" == "True" ]]; then
NETWORK_FILE="${NETWORK_SETTINGS_DIR}/network_settings_csit.yaml"
else
NETWORK_FILE="${NETWORK_SETTINGS_DIR}/network_settings.yaml"
fi
# Check that network settings file exists
if ! sudo test -e "$NETWORK_FILE"; then
echo "ERROR: Required settings file missing: Network Settings file ${NETWORK_FILE}"
exit 1
fi
# start deployment
sudo ${DEPLOY_CMD} -d ${DEPLOY_FILE} -n ${NETWORK_FILE} --debug
echo
echo "--------------------------------------------------------"
echo "Done!"
|