diff options
Diffstat (limited to 'ci')
l---------[-rwxr-xr-x] | ci/00-maasdeploy.sh | 385 | ||||
-rwxr-xr-x | ci/03-maasdeploy.sh | 36 |
2 files changed, 26 insertions, 395 deletions
diff --git a/ci/00-maasdeploy.sh b/ci/00-maasdeploy.sh index 38bf3537..33b6f4dd 100755..120000 --- a/ci/00-maasdeploy.sh +++ b/ci/00-maasdeploy.sh @@ -1,384 +1 @@ -#!/bin/bash -#placeholder for deployment script. -set -ex - -echo "Note: This script is deprecated. Use 03-maasdeploy.sh instead." -echo "Are you sure you want to execute this script? [y/N] " -read a -[[ "$a" = "y" ]] || exit - - -virtinstall=0 -labname=$1 - -#install the packages needed -sudo apt-add-repository ppa:opnfv-team/proposed -y -sudo apt-add-repository ppa:maas-deployers/stable -y -sudo apt-add-repository ppa:juju/stable -y -sudo apt-add-repository ppa:maas/stable -y -sudo apt-add-repository cloud-archive:mitaka -y -sudo apt-get update -y -sudo apt-get dist-upgrade -y -sudo pip install --upgrade pip -sudo apt-get install openssh-server bzr git maas-deployer juju juju-deployer \ - maas-cli python-pip python-psutil python-openstackclient \ - python-congressclient gsutil charm-tools pastebinit -y - -sudo apt-get purge juju -y -wget https://launchpad.net/~juju/+archive/ubuntu/stable/+files/juju-core_1.25.5-0ubuntu1~14.04.2~juju1_amd64.deb -sudo dpkg -i juju-core_1.25.5-0ubuntu1~14.04.2~juju1_amd64.deb - -#first parameter should be custom and second should be either -# absolute location of file (including file name) or url of the -# file to download. - -labname=$1 -labfile=$2 - -# -# Config preparation -# - -# Get labconfig and generate deployment.yaml for MAAS and deployconfig.yaml -case "$labname" in - intelpod[569]|orangepod[12]|cengnpod[12] ) - array=(${labname//pod/ }) - cp ../labconfig/${array[0]}/pod${array[1]}/labconfig.yaml . - python genMAASConfig.py -l labconfig.yaml > deployment.yaml - python genDeploymentConfig.py -l labconfig.yaml > deployconfig.yaml - ;; - 'attvirpod1' ) - cp ../labconfig/att/virpod1/labconfig.yaml . - python genMAASConfig.py -l labconfig.yaml > deployment.yaml - python genDeploymentConfig.py -l labconfig.yaml > deployconfig.yaml - ;; - 'juniperpod1' ) - cp maas/juniper/pod1/deployment.yaml ./deployment.yaml - ;; - 'custom') - if [ -e $labfile ]; then - cp $labfile ./labconfig.yaml || true - else - wget $labconfigfile -t 3 -T 10 -O ./labconfig.yaml || true - count=`wc -l labconfig.yaml | cut -d " " -f 1` - if [ $count -lt 10 ]; then - rm -rf labconfig.yaml - fi - fi - if [ ! -e ./labconfig.yaml ]; then - virtinstall=1 - else - python genMAASConfig.py -l labconfig.yaml > deployment.yaml - python genDeploymentConfig.py -l labconfig.yaml > deployconfig.yaml - labname=`grep "maas_name" deployment.yaml | cut -d ':' -f 2 | sed -e 's/ //'` - fi - ;; - * ) - virtinstall=1 - ;; -esac - -# In the case of a virtual deployment get deployment.yaml and deployconfig.yaml -if [ "$virtinstall" -eq 1 ]; then - labname="default" - ./cleanvm.sh || true - cp ../labconfig/default/deployment.yaml ./ - cp ../labconfig/default/labconfig.yaml ./ - cp ../labconfig/default/deployconfig.yaml ./ -fi - -# -# Prepare local environment to avoid password asking -# - -# make sure no password asked during the deployment. -echo "$USER ALL=(ALL) NOPASSWD:ALL" > 90-joid-init - -if [ -e /etc/sudoers.d/90-joid-init ]; then - sudo cp /etc/sudoers.d/90-joid-init 91-joid-init - sudo chown $USER:$USER 91-joid-init - sudo chmod 660 91-joid-init - sudo cat 90-joid-init >> 91-joid-init - sudo chown root:root 91-joid-init - sudo mv 91-joid-init /etc/sudoers.d/ -else - sudo chown root:root 90-joid-init - sudo mv 90-joid-init /etc/sudoers.d/ -fi - -if [ ! -e $HOME/.ssh/id_rsa ]; then - ssh-keygen -N '' -f $HOME/.ssh/id_rsa -fi - -echo "... Deployment of maas Started ...." - -# -# Virsh preparation -# - -# define the pool and try to start even though its already exist. -# For fresh install this may or may not there. -sudo apt-get install libvirt-bin -y -sudo adduser $USER libvirtd -sudo virsh pool-define-as default --type dir --target /var/lib/libvirt/images/ || true -sudo virsh pool-start default || true -sudo virsh pool-autostart default || true - -# In case of virtual install set network -if [ "$virtinstall" -eq 1 ]; then - sudo virsh net-dumpxml default > default-net-org.xml - sudo sed -i '/dhcp/d' default-net-org.xml - sudo sed -i '/range/d' default-net-org.xml - sudo virsh net-define default-net-org.xml - sudo virsh net-destroy default - sudo virsh net-start default -fi - -# Ensure virsh can connect without ssh auth -cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys - - -# -# Cleanup, juju init and config backup -# - -# To avoid problem between apiclient/maas_client and apiclient from google -# we remove the package google-api-python-client from yardstick installer -if [ $(pip list |grep google-api-python-client |wc -l) == 1 ]; then - sudo pip uninstall google-api-python-client -fi - -#create backup directory -mkdir ~/joid_config/ || true -mkdir ~/.juju/ || true - -# Init Juju -juju init -f || true - -# -# MAAS deploy -# - -sudo maas-deployer -c deployment.yaml -d --force - -sudo chown $USER:$USER environments.yaml - -echo "... Deployment of maas finish ...." - -# Backup deployment.yaml and deployconfig.yaml in .juju folder - -cp ./environments.yaml ~/.juju/ -cp ./environments.yaml ~/joid_config/ - -if [ -e ./deployconfig.yaml ]; then - cp ./deployconfig.yaml ~/.juju/ - cp ./labconfig.yaml ~/.juju/ - cp ./deployconfig.yaml ~/joid_config/ - cp ./labconfig.yaml ~/joid_config/ -fi - -if [ -e ./deployment.yaml ]; then - cp ./deployment.yaml ~/.juju/ - cp ./deployment.yaml ~/joid_config/ -fi - -# -# MAAS Customization -# - -maas_ip=`grep " ip_address" deployment.yaml | cut -d ':' -f 2 | sed -e 's/ //'` -apikey=`grep maas-oauth: environments.yaml | cut -d "'" -f 2` -maas login maas http://${maas_ip}/MAAS/api/1.0 ${apikey} -maas maas sshkeys new key="`cat $HOME/.ssh/id_rsa.pub`" - -#Added the Qtip public to run the Qtip test after install on bare metal nodes. -#maas maas sshkeys new key="`cat ./maas/sshkeys/QtipKey.pub`" -#maas maas sshkeys new key="`cat ./maas/sshkeys/DominoKey.pub`" - -#adding compute and control nodes VM to MAAS for virtual deployment purpose. -if [ "$virtinstall" -eq 1 ]; then - # create two more VMs to do the deployment. - sudo virt-install --connect qemu:///system --name node1-control --ram 8192 --cpu host --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node1-control - - sudo virt-install --connect qemu:///system --name node2-compute --ram 8192 --cpu host --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node2-compute - - sudo virt-install --connect qemu:///system --name node5-compute --ram 8192 --cpu host --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node5-compute - - node1controlmac=`grep "mac address" node1-control | head -1 | cut -d "'" -f 2` - node2computemac=`grep "mac address" node2-compute | head -1 | cut -d "'" -f 2` - node5computemac=`grep "mac address" node5-compute | head -1 | cut -d "'" -f 2` - - sudo virsh -c qemu:///system define --file node1-control - sudo virsh -c qemu:///system define --file node2-compute - sudo virsh -c qemu:///system define --file node5-compute - - maas maas tags new name='control' - maas maas tags new name='compute' - - controlnodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node1-control' tags='control' hostname='node1-control' power_type='virsh' mac_addresses=$node1controlmac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node1-control' | grep system_id | cut -d '"' -f 4 ` - - maas maas tag update-nodes control add=$controlnodeid - - computenodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node2-compute' tags='compute' hostname='node2-compute' power_type='virsh' mac_addresses=$node2computemac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node2-compute' | grep system_id | cut -d '"' -f 4 ` - - maas maas tag update-nodes compute add=$computenodeid - - computenodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node5-compute' tags='compute' hostname='node5-compute' power_type='virsh' mac_addresses=$node5computemac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node5-compute' | grep system_id | cut -d '"' -f 4 ` - - maas maas tag update-nodes compute add=$computenodeid -fi - -# -# Functions for MAAS network customization -# - -#Below function will mark the interfaces in Auto mode to enbled by MAAS -enableautomode() { - listofnodes=`maas maas nodes list | grep system_id | cut -d '"' -f 4` - for nodes in $listofnodes - do - maas maas interface link-subnet $nodes $1 mode=$2 subnet=$3 - done -} - -#Below function will mark the interfaces in Auto mode to enbled by MAAS -# using hostname of the node added into MAAS -enableautomodebyname() { - if [ ! -z "$4" ]; then - for i in `seq 1 7`; - do - nodes=`maas maas nodes list | grep system_id | cut -d '"' -f 4` - if [ ! -z "$nodes" ]; then - maas maas interface link-subnet $nodes $1 mode=$2 subnet=$3 - fi - done - fi -} - -#Below function will create vlan and update interface with the new vlan -# will return the vlan id created -crvlanupdsubnet() { - newvlanid=`maas maas vlans create $2 name=$3 vid=$4 | grep resource | cut -d '/' -f 6 ` - maas maas subnet update $5 vlan=$newvlanid - eval "$1"="'$newvlanid'" -} - -#Below function will create interface with new vlan and bind to physical interface -crnodevlanint() { - listofnodes=`maas maas nodes list | grep system_id | cut -d '"' -f 4` - - for nodes in $listofnodes - do - parentid=`maas maas interface read $nodes $2 | grep interfaces | cut -d '/' -f 8` - maas maas interfaces create-vlan $nodes vlan=$1 parent=$parentid - done - } - -#function for JUJU envronment - -addcredential() { - controllername=`awk 'NR==1{print $2}' environments.yaml` - cloudname=`awk 'NR==1{print $2}' environments.yaml` - - echo "credentials:" > credential.yaml - echo " $controllername:" >> credential.yaml - echo " opnfv-credentials:" >> credential.yaml - echo " auth-type: oauth1" >> credential.yaml - echo " maas-oauth: $apikey" >> credential.yaml - - juju add-credential $controllername -f credential.yaml --replace -} - -addcloud() { - controllername=`awk 'NR==1{print $2}' environments.yaml` - cloudname=`awk 'NR==1{print $2}' environments.yaml` - - echo "clouds:" > maas-cloud.yaml - echo " $cloudname:" >> maas-cloud.yaml - echo " type: maas" >> maas-cloud.yaml - echo " auth-types: [oauth1]" >> maas-cloud.yaml - echo " endpoint: http://$maas_ip/MAAS" >> maas-cloud.yaml - - juju add-cloud $cloudname maas-cloud.yaml --replace -} - - -# -# VLAN customization -# - -case "$labname" in - 'intelpod9' ) - maas refresh - crvlanupdsubnet vlan904 fabric-1 "MgmtNetwork" 904 2 || true - crvlanupdsubnet vlan905 fabric-2 "PublicNetwork" 905 3 || true - crnodevlanint $vlan905 eth1 || true - crnodevlanint $vlan905 eth3 || true - enableautomodebyname eth1.905 AUTO "10.9.15.0/24" || true - enableautomodebyname eth3.905 AUTO "10.9.15.0/24" || true - enableautomodebyname eth0 AUTO "10.9.12.0/24" || true - enableautomodebyname eth2 AUTO "10.9.12.0/24" || true - ;; -esac - -# -# Enable MAAS nodes interfaces -# - -#read interface needed in Auto mode and enable it. Will be rmeoved once auto enablement will be implemented in the maas-deployer. -if [ -e ~/joid_config/deployconfig.yaml ]; then - cp ~/joid_config/deployconfig.yaml ./deployconfig.yaml -elif [ -e ~/.juju/deployconfig.yaml ]; then - cp ~/.juju/deployconfig.yaml ./deployconfig.yaml -fi - -if [ -e ./deployconfig.yaml ]; then - enableiflist=`grep "interface-enable" deployconfig.yaml | cut -d ' ' -f 4 ` - datanet=`grep "dataNetwork" deployconfig.yaml | cut -d ' ' -f 4 | sed -e 's/ //'` - stornet=`grep "storageNetwork" deployconfig.yaml | cut -d ' ' -f 4 | sed -e 's/ //'` - pubnet=`grep "publicNetwork" deployconfig.yaml | cut -d ' ' -f 4 | sed -e 's/ //'` - - # split EXTERNAL_NETWORK=first ip;last ip; gateway;network - - if [ "$datanet" != "''" ]; then - EXTNET=(${enableiflist//,/ }) - i="0" - while [ ! -z "${EXTNET[i]}" ]; - do - enableautomode ${EXTNET[i]} AUTO $datanet || true - i=$[$i+1] - done - fi - if [ "$stornet" != "''" ]; then - EXTNET=(${enableiflist//,/ }) - i="0" - while [ ! -z "${EXTNET[i]}" ]; - do - enableautomode ${EXTNET[i]} AUTO $stornet || true - i=$[$i+1] - done - fi - if [ "$pubnet" != "''" ]; then - EXTNET=(${enableiflist//,/ }) - i="0" - while [ ! -z "${EXTNET[i]}" ]; - do - enableautomode ${EXTNET[i]} AUTO $pubnet || true - i=$[$i+1] - done - fi -fi - - -# Add the cloud and controller credentials for MAAS for that lab. -jujuver=`juju --version` - -if [[ "$jujuver" > "2" ]]; then - addcloud - addcredential -fi - -# -# End of scripts -# -echo " .... MAAS deployment finished successfully ...." +03-maasdeploy.sh
\ No newline at end of file diff --git a/ci/03-maasdeploy.sh b/ci/03-maasdeploy.sh index e5e203e7..479ab039 100755 --- a/ci/03-maasdeploy.sh +++ b/ci/03-maasdeploy.sh @@ -16,10 +16,14 @@ sudo apt-add-repository ppa:maas/stable -y sudo apt-add-repository cloud-archive:ocata -y sudo apt-get update -y #sudo apt-get dist-upgrade -y -sudo apt-get install bridge-utils openssh-server bzr git virtinst qemu-kvm libvirt-bin juju \ - maas maas-region-controller python-pip python-psutil python-openstackclient \ +sudo apt-get install bridge-utils openssh-server bzr git virtinst qemu-kvm libvirt-bin \ + maas maas-region-controller juju python-pip python-psutil python-openstackclient \ python-congressclient gsutil charm-tools pastebinit python-jinja2 sshpass \ - openssh-server vlan ipmitool jq expect -y + openssh-server vlan ipmitool jq expect snap -y + +#sudo apt-get install snap -y +#sudo snap install maas --classic +#sudo snap install juju --classic sudo pip install --upgrade pip @@ -130,7 +134,11 @@ echo "... Deployment of maas Started ...." # define the pool and try to start even though its already exist. # For fresh install this may or may not there. -sudo adduser $USER libvirtd +#some system i am seeing libvirt and some have libvirtd looks like libvirt-bin is +#keep switching so lets try both. + +sudo adduser $USER libvirtd || true +sudo adduser $USER libvirt || true sudo virsh pool-define-as default --type dir --target /var/lib/libvirt/images/ || true sudo virsh pool-start default || true sudo virsh pool-autostart default || true @@ -138,10 +146,10 @@ sudo virsh pool-autostart default || true # In case of virtual install set network if [ "$virtinstall" -eq 1 ]; then sudo virsh net-dumpxml default > default-net-org.xml - sudo sed -i '/dhcp/d' default-net-org.xml - sudo sed -i '/range/d' default-net-org.xml - sudo virsh net-define default-net-org.xml + sed -i '/dhcp/d' default-net-org.xml + sed -i '/range/d' default-net-org.xml sudo virsh net-destroy default + sudo virsh net-define default-net-org.xml sudo virsh net-start default rm -f default-net-org.xml fi @@ -294,8 +302,12 @@ setupspacenetwork(){ *) JUJU_SPACE='default'; DHCP='OFF'; echo " >>> Unknown SPACE" ;; esac JUJU_SPACE_ID=$(maas $PROFILE spaces read | jq -r ".[] | select(.name==\"$JUJU_SPACE\")".id) - if ([ $NET_FABRIC_NAME ] && [ $NET_FABRIC_NAME != "null" ]); then - maas $PROFILE subnet update $SPACE_CIDR space=$JUJU_SPACE_ID + JUJU_VLAN_VID=$(maas $PROFILE subnets read | jq -r ".[] | select(.name==\"$SPACE_CIDR\")".vlan.vid) + NET_FABRIC_ID=$(maas $PROFILE fabric read $NET_FABRIC_NAME | jq -r ".id") + if ([ $NET_FABRIC_ID ] && [ $NET_FABRIC_ID != "null" ]); then + if ([ $JUJU_VLAN_VID ] && [ $JUJU_VLAN_VID != "null" ]); then + maas $PROFILE vlan update $NET_FABRIC_ID $JUJU_VLAN_VID space=$JUJU_SPACE_ID + fi fi if ([ $type == "admin" ]); then # If we have a network, we create it @@ -397,12 +409,12 @@ addnodes(){ units=$(($units - 1)); NODE_NAME=`cat labconfig.json | jq ".lab.racks[].nodes[$units].name" | cut -d \" -f 2 ` - sudo virt-install --connect $VIRSHURL --name $NODE_NAME --ram 8192 --cpu host --vcpus 4 \ + virt-install --connect $VIRSHURL --name $NODE_NAME --ram 8192 --cpu host --vcpus 4 \ --disk size=120,format=qcow2,bus=virtio,cache=directsync,io=native,pool=default \ $netw $netw --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee $NODE_NAME nodemac=`grep "mac address" $NODE_NAME | head -1 | cut -d '"' -f 2` - sudo virsh -c $VIRSHURL define --file $NODE_NAME + virsh -c $VIRSHURL define --file $NODE_NAME rm -f $NODE_NAME maas $PROFILE machines create autodetect_nodegroup='yes' name=$NODE_NAME \ tags='control compute' hostname=$NODE_NAME power_type='virsh' mac_addresses=$nodemac \ @@ -431,6 +443,8 @@ addnodes(){ done fi + maas $PROFILE pods create type=virsh power_address="$VIRSHURL" power_user=$USER + # make sure nodes are added into MAAS and none of them is in commisoning state while [ "$(maas $PROFILE nodes read | grep Commissioning )" ]; do |