diff options
author | Alexandru Avadanii <Alexandru.Avadanii@enea.com> | 2018-01-29 06:02:10 +0100 |
---|---|---|
committer | Alexandru Avadanii <Alexandru.Avadanii@enea.com> | 2018-02-05 06:03:35 +0100 |
commit | 9c3ecbb5238fa815e3f6b46c1933d69d588a5430 (patch) | |
tree | 793f75e8e13cfe82d125f82c5d6732dcf3df539f /ci | |
parent | 5f95f52021fa08de3a0a46bdd98b0b711fec1ecc (diff) |
[virtual] Split 'pxebr' from 'mcpcontrol' net
- add new virsh managed network 'pxebr' (to mimic baremetal behavior
on virtual PODs, this will be the equivalent of PXE/admin network);
- connect 'pxebr' to 3rd interface for cfg01, mas01 for all deploys
(used to be baremetal-specific), replacing 'internal';
- keep 'mcpcontrol' connected only to 'cfg01' (+ 'mas01' if present)
for initial infrastructure bring-up (1st interface);
- switch all virtual cluster nodes to 'pxebr' (1st interface);
- use 'pxebr' for all Salt cluster nodes traffic, 'mcpcontrol' only
for mas01<=>cfg01 Salt traffic;
- convert <user-data.template> to jinja2 and expand it based on PDF
instead of using `envsubst`;
- split <user-data.sh.j2> into two versions, one for each network
used for Salt traffic;
- ci/deploy.sh: Read scenario data before template parsing for
cluster domain variable, needed in virsh network def;
- leave docs diagram refresh to later after all possible deploy types
have settled;
- limit keyserver proxy usage to nodes where the configured http proxy
matches the first nameserver (true for all MaaS-provisioned nodes),
so we can re-use the same pillar for FN VMs and baremetal nodes;
- add PXE/admin IP on cfg01's 3rd interface and switch other vnodes
`salt_master_host` to point to it;
JIRA: FUEL-322
Change-Id: Ie4f7aedddf2ef81046f1127b377d88dce79f0fda
Signed-off-by: Alexandru Avadanii <Alexandru.Avadanii@enea.com>
Diffstat (limited to 'ci')
-rwxr-xr-x | ci/deploy.sh | 11 |
1 files changed, 6 insertions, 5 deletions
diff --git a/ci/deploy.sh b/ci/deploy.sh index ef941b005..fb22cc45a 100755 --- a/ci/deploy.sh +++ b/ci/deploy.sh @@ -277,19 +277,20 @@ if [ ! -f "${SCENARIO_DIR}/defaults-$(uname -i).yaml" ]; then notify_e "[ERROR] Scenario defaults file is missing!" fi +# Get scenario data and (jumpserver) arch defaults +eval "$(parse_yaml "${SCENARIO_DIR}/defaults-$(uname -i).yaml")" +eval "$(parse_yaml "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml")" +export CLUSTER_DOMAIN=${cluster_domain} + # Expand jinja2 templates based on PDF data and env vars do_templates "${REPO_ROOT_PATH}" "${STORAGE_DIR}" "${TARGET_LAB}" \ "${TARGET_POD}" "${BASE_CONFIG_URI}" -# Get required infra deployment data +# Get required infra deployment data based on PDF/IDF (after template parsing) set +x -eval "$(parse_yaml "${SCENARIO_DIR}/defaults-$(uname -i).yaml")" -eval "$(parse_yaml "${SCENARIO_DIR}/${DEPLOY_TYPE}/${DEPLOY_SCENARIO}.yaml")" eval "$(parse_yaml "${LOCAL_PDF_RECLASS}")" [[ "${CI_DEBUG}" =~ (false|0) ]] || set -x -export CLUSTER_DOMAIN=${cluster_domain} - # Serialize vnode data as '<name0>,<ram0>,<vcpu0>|<name1>,<ram1>,<vcpu1>[...]' for node in "${virtual_nodes[@]}"; do virtual_custom_ram="virtual_${node}_ram" |