summaryrefslogtreecommitdiffstats
path: root/ci
diff options
context:
space:
mode:
Diffstat (limited to 'ci')
l---------[-rwxr-xr-x]ci/00-maasdeploy.sh385
-rwxr-xr-xci/03-maasdeploy.sh36
2 files changed, 26 insertions, 395 deletions
diff --git a/ci/00-maasdeploy.sh b/ci/00-maasdeploy.sh
index 38bf3537..33b6f4dd 100755..120000
--- a/ci/00-maasdeploy.sh
+++ b/ci/00-maasdeploy.sh
@@ -1,384 +1 @@
-#!/bin/bash
-#placeholder for deployment script.
-set -ex
-
-echo "Note: This script is deprecated. Use 03-maasdeploy.sh instead."
-echo "Are you sure you want to execute this script? [y/N] "
-read a
-[[ "$a" = "y" ]] || exit
-
-
-virtinstall=0
-labname=$1
-
-#install the packages needed
-sudo apt-add-repository ppa:opnfv-team/proposed -y
-sudo apt-add-repository ppa:maas-deployers/stable -y
-sudo apt-add-repository ppa:juju/stable -y
-sudo apt-add-repository ppa:maas/stable -y
-sudo apt-add-repository cloud-archive:mitaka -y
-sudo apt-get update -y
-sudo apt-get dist-upgrade -y
-sudo pip install --upgrade pip
-sudo apt-get install openssh-server bzr git maas-deployer juju juju-deployer \
- maas-cli python-pip python-psutil python-openstackclient \
- python-congressclient gsutil charm-tools pastebinit -y
-
-sudo apt-get purge juju -y
-wget https://launchpad.net/~juju/+archive/ubuntu/stable/+files/juju-core_1.25.5-0ubuntu1~14.04.2~juju1_amd64.deb
-sudo dpkg -i juju-core_1.25.5-0ubuntu1~14.04.2~juju1_amd64.deb
-
-#first parameter should be custom and second should be either
-# absolute location of file (including file name) or url of the
-# file to download.
-
-labname=$1
-labfile=$2
-
-#
-# Config preparation
-#
-
-# Get labconfig and generate deployment.yaml for MAAS and deployconfig.yaml
-case "$labname" in
- intelpod[569]|orangepod[12]|cengnpod[12] )
- array=(${labname//pod/ })
- cp ../labconfig/${array[0]}/pod${array[1]}/labconfig.yaml .
- python genMAASConfig.py -l labconfig.yaml > deployment.yaml
- python genDeploymentConfig.py -l labconfig.yaml > deployconfig.yaml
- ;;
- 'attvirpod1' )
- cp ../labconfig/att/virpod1/labconfig.yaml .
- python genMAASConfig.py -l labconfig.yaml > deployment.yaml
- python genDeploymentConfig.py -l labconfig.yaml > deployconfig.yaml
- ;;
- 'juniperpod1' )
- cp maas/juniper/pod1/deployment.yaml ./deployment.yaml
- ;;
- 'custom')
- if [ -e $labfile ]; then
- cp $labfile ./labconfig.yaml || true
- else
- wget $labconfigfile -t 3 -T 10 -O ./labconfig.yaml || true
- count=`wc -l labconfig.yaml | cut -d " " -f 1`
- if [ $count -lt 10 ]; then
- rm -rf labconfig.yaml
- fi
- fi
- if [ ! -e ./labconfig.yaml ]; then
- virtinstall=1
- else
- python genMAASConfig.py -l labconfig.yaml > deployment.yaml
- python genDeploymentConfig.py -l labconfig.yaml > deployconfig.yaml
- labname=`grep "maas_name" deployment.yaml | cut -d ':' -f 2 | sed -e 's/ //'`
- fi
- ;;
- * )
- virtinstall=1
- ;;
-esac
-
-# In the case of a virtual deployment get deployment.yaml and deployconfig.yaml
-if [ "$virtinstall" -eq 1 ]; then
- labname="default"
- ./cleanvm.sh || true
- cp ../labconfig/default/deployment.yaml ./
- cp ../labconfig/default/labconfig.yaml ./
- cp ../labconfig/default/deployconfig.yaml ./
-fi
-
-#
-# Prepare local environment to avoid password asking
-#
-
-# make sure no password asked during the deployment.
-echo "$USER ALL=(ALL) NOPASSWD:ALL" > 90-joid-init
-
-if [ -e /etc/sudoers.d/90-joid-init ]; then
- sudo cp /etc/sudoers.d/90-joid-init 91-joid-init
- sudo chown $USER:$USER 91-joid-init
- sudo chmod 660 91-joid-init
- sudo cat 90-joid-init >> 91-joid-init
- sudo chown root:root 91-joid-init
- sudo mv 91-joid-init /etc/sudoers.d/
-else
- sudo chown root:root 90-joid-init
- sudo mv 90-joid-init /etc/sudoers.d/
-fi
-
-if [ ! -e $HOME/.ssh/id_rsa ]; then
- ssh-keygen -N '' -f $HOME/.ssh/id_rsa
-fi
-
-echo "... Deployment of maas Started ...."
-
-#
-# Virsh preparation
-#
-
-# define the pool and try to start even though its already exist.
-# For fresh install this may or may not there.
-sudo apt-get install libvirt-bin -y
-sudo adduser $USER libvirtd
-sudo virsh pool-define-as default --type dir --target /var/lib/libvirt/images/ || true
-sudo virsh pool-start default || true
-sudo virsh pool-autostart default || true
-
-# In case of virtual install set network
-if [ "$virtinstall" -eq 1 ]; then
- sudo virsh net-dumpxml default > default-net-org.xml
- sudo sed -i '/dhcp/d' default-net-org.xml
- sudo sed -i '/range/d' default-net-org.xml
- sudo virsh net-define default-net-org.xml
- sudo virsh net-destroy default
- sudo virsh net-start default
-fi
-
-# Ensure virsh can connect without ssh auth
-cat $HOME/.ssh/id_rsa.pub >> $HOME/.ssh/authorized_keys
-
-
-#
-# Cleanup, juju init and config backup
-#
-
-# To avoid problem between apiclient/maas_client and apiclient from google
-# we remove the package google-api-python-client from yardstick installer
-if [ $(pip list |grep google-api-python-client |wc -l) == 1 ]; then
- sudo pip uninstall google-api-python-client
-fi
-
-#create backup directory
-mkdir ~/joid_config/ || true
-mkdir ~/.juju/ || true
-
-# Init Juju
-juju init -f || true
-
-#
-# MAAS deploy
-#
-
-sudo maas-deployer -c deployment.yaml -d --force
-
-sudo chown $USER:$USER environments.yaml
-
-echo "... Deployment of maas finish ...."
-
-# Backup deployment.yaml and deployconfig.yaml in .juju folder
-
-cp ./environments.yaml ~/.juju/
-cp ./environments.yaml ~/joid_config/
-
-if [ -e ./deployconfig.yaml ]; then
- cp ./deployconfig.yaml ~/.juju/
- cp ./labconfig.yaml ~/.juju/
- cp ./deployconfig.yaml ~/joid_config/
- cp ./labconfig.yaml ~/joid_config/
-fi
-
-if [ -e ./deployment.yaml ]; then
- cp ./deployment.yaml ~/.juju/
- cp ./deployment.yaml ~/joid_config/
-fi
-
-#
-# MAAS Customization
-#
-
-maas_ip=`grep " ip_address" deployment.yaml | cut -d ':' -f 2 | sed -e 's/ //'`
-apikey=`grep maas-oauth: environments.yaml | cut -d "'" -f 2`
-maas login maas http://${maas_ip}/MAAS/api/1.0 ${apikey}
-maas maas sshkeys new key="`cat $HOME/.ssh/id_rsa.pub`"
-
-#Added the Qtip public to run the Qtip test after install on bare metal nodes.
-#maas maas sshkeys new key="`cat ./maas/sshkeys/QtipKey.pub`"
-#maas maas sshkeys new key="`cat ./maas/sshkeys/DominoKey.pub`"
-
-#adding compute and control nodes VM to MAAS for virtual deployment purpose.
-if [ "$virtinstall" -eq 1 ]; then
- # create two more VMs to do the deployment.
- sudo virt-install --connect qemu:///system --name node1-control --ram 8192 --cpu host --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node1-control
-
- sudo virt-install --connect qemu:///system --name node2-compute --ram 8192 --cpu host --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node2-compute
-
- sudo virt-install --connect qemu:///system --name node5-compute --ram 8192 --cpu host --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node5-compute
-
- node1controlmac=`grep "mac address" node1-control | head -1 | cut -d "'" -f 2`
- node2computemac=`grep "mac address" node2-compute | head -1 | cut -d "'" -f 2`
- node5computemac=`grep "mac address" node5-compute | head -1 | cut -d "'" -f 2`
-
- sudo virsh -c qemu:///system define --file node1-control
- sudo virsh -c qemu:///system define --file node2-compute
- sudo virsh -c qemu:///system define --file node5-compute
-
- maas maas tags new name='control'
- maas maas tags new name='compute'
-
- controlnodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node1-control' tags='control' hostname='node1-control' power_type='virsh' mac_addresses=$node1controlmac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node1-control' | grep system_id | cut -d '"' -f 4 `
-
- maas maas tag update-nodes control add=$controlnodeid
-
- computenodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node2-compute' tags='compute' hostname='node2-compute' power_type='virsh' mac_addresses=$node2computemac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node2-compute' | grep system_id | cut -d '"' -f 4 `
-
- maas maas tag update-nodes compute add=$computenodeid
-
- computenodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node5-compute' tags='compute' hostname='node5-compute' power_type='virsh' mac_addresses=$node5computemac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node5-compute' | grep system_id | cut -d '"' -f 4 `
-
- maas maas tag update-nodes compute add=$computenodeid
-fi
-
-#
-# Functions for MAAS network customization
-#
-
-#Below function will mark the interfaces in Auto mode to enbled by MAAS
-enableautomode() {
- listofnodes=`maas maas nodes list | grep system_id | cut -d '"' -f 4`
- for nodes in $listofnodes
- do
- maas maas interface link-subnet $nodes $1 mode=$2 subnet=$3
- done
-}
-
-#Below function will mark the interfaces in Auto mode to enbled by MAAS
-# using hostname of the node added into MAAS
-enableautomodebyname() {
- if [ ! -z "$4" ]; then
- for i in `seq 1 7`;
- do
- nodes=`maas maas nodes list | grep system_id | cut -d '"' -f 4`
- if [ ! -z "$nodes" ]; then
- maas maas interface link-subnet $nodes $1 mode=$2 subnet=$3
- fi
- done
- fi
-}
-
-#Below function will create vlan and update interface with the new vlan
-# will return the vlan id created
-crvlanupdsubnet() {
- newvlanid=`maas maas vlans create $2 name=$3 vid=$4 | grep resource | cut -d '/' -f 6 `
- maas maas subnet update $5 vlan=$newvlanid
- eval "$1"="'$newvlanid'"
-}
-
-#Below function will create interface with new vlan and bind to physical interface
-crnodevlanint() {
- listofnodes=`maas maas nodes list | grep system_id | cut -d '"' -f 4`
-
- for nodes in $listofnodes
- do
- parentid=`maas maas interface read $nodes $2 | grep interfaces | cut -d '/' -f 8`
- maas maas interfaces create-vlan $nodes vlan=$1 parent=$parentid
- done
- }
-
-#function for JUJU envronment
-
-addcredential() {
- controllername=`awk 'NR==1{print $2}' environments.yaml`
- cloudname=`awk 'NR==1{print $2}' environments.yaml`
-
- echo "credentials:" > credential.yaml
- echo " $controllername:" >> credential.yaml
- echo " opnfv-credentials:" >> credential.yaml
- echo " auth-type: oauth1" >> credential.yaml
- echo " maas-oauth: $apikey" >> credential.yaml
-
- juju add-credential $controllername -f credential.yaml --replace
-}
-
-addcloud() {
- controllername=`awk 'NR==1{print $2}' environments.yaml`
- cloudname=`awk 'NR==1{print $2}' environments.yaml`
-
- echo "clouds:" > maas-cloud.yaml
- echo " $cloudname:" >> maas-cloud.yaml
- echo " type: maas" >> maas-cloud.yaml
- echo " auth-types: [oauth1]" >> maas-cloud.yaml
- echo " endpoint: http://$maas_ip/MAAS" >> maas-cloud.yaml
-
- juju add-cloud $cloudname maas-cloud.yaml --replace
-}
-
-
-#
-# VLAN customization
-#
-
-case "$labname" in
- 'intelpod9' )
- maas refresh
- crvlanupdsubnet vlan904 fabric-1 "MgmtNetwork" 904 2 || true
- crvlanupdsubnet vlan905 fabric-2 "PublicNetwork" 905 3 || true
- crnodevlanint $vlan905 eth1 || true
- crnodevlanint $vlan905 eth3 || true
- enableautomodebyname eth1.905 AUTO "10.9.15.0/24" || true
- enableautomodebyname eth3.905 AUTO "10.9.15.0/24" || true
- enableautomodebyname eth0 AUTO "10.9.12.0/24" || true
- enableautomodebyname eth2 AUTO "10.9.12.0/24" || true
- ;;
-esac
-
-#
-# Enable MAAS nodes interfaces
-#
-
-#read interface needed in Auto mode and enable it. Will be rmeoved once auto enablement will be implemented in the maas-deployer.
-if [ -e ~/joid_config/deployconfig.yaml ]; then
- cp ~/joid_config/deployconfig.yaml ./deployconfig.yaml
-elif [ -e ~/.juju/deployconfig.yaml ]; then
- cp ~/.juju/deployconfig.yaml ./deployconfig.yaml
-fi
-
-if [ -e ./deployconfig.yaml ]; then
- enableiflist=`grep "interface-enable" deployconfig.yaml | cut -d ' ' -f 4 `
- datanet=`grep "dataNetwork" deployconfig.yaml | cut -d ' ' -f 4 | sed -e 's/ //'`
- stornet=`grep "storageNetwork" deployconfig.yaml | cut -d ' ' -f 4 | sed -e 's/ //'`
- pubnet=`grep "publicNetwork" deployconfig.yaml | cut -d ' ' -f 4 | sed -e 's/ //'`
-
- # split EXTERNAL_NETWORK=first ip;last ip; gateway;network
-
- if [ "$datanet" != "''" ]; then
- EXTNET=(${enableiflist//,/ })
- i="0"
- while [ ! -z "${EXTNET[i]}" ];
- do
- enableautomode ${EXTNET[i]} AUTO $datanet || true
- i=$[$i+1]
- done
- fi
- if [ "$stornet" != "''" ]; then
- EXTNET=(${enableiflist//,/ })
- i="0"
- while [ ! -z "${EXTNET[i]}" ];
- do
- enableautomode ${EXTNET[i]} AUTO $stornet || true
- i=$[$i+1]
- done
- fi
- if [ "$pubnet" != "''" ]; then
- EXTNET=(${enableiflist//,/ })
- i="0"
- while [ ! -z "${EXTNET[i]}" ];
- do
- enableautomode ${EXTNET[i]} AUTO $pubnet || true
- i=$[$i+1]
- done
- fi
-fi
-
-
-# Add the cloud and controller credentials for MAAS for that lab.
-jujuver=`juju --version`
-
-if [[ "$jujuver" > "2" ]]; then
- addcloud
- addcredential
-fi
-
-#
-# End of scripts
-#
-echo " .... MAAS deployment finished successfully ...."
+03-maasdeploy.sh \ No newline at end of file
diff --git a/ci/03-maasdeploy.sh b/ci/03-maasdeploy.sh
index e5e203e7..479ab039 100755
--- a/ci/03-maasdeploy.sh
+++ b/ci/03-maasdeploy.sh
@@ -16,10 +16,14 @@ sudo apt-add-repository ppa:maas/stable -y
sudo apt-add-repository cloud-archive:ocata -y
sudo apt-get update -y
#sudo apt-get dist-upgrade -y
-sudo apt-get install bridge-utils openssh-server bzr git virtinst qemu-kvm libvirt-bin juju \
- maas maas-region-controller python-pip python-psutil python-openstackclient \
+sudo apt-get install bridge-utils openssh-server bzr git virtinst qemu-kvm libvirt-bin \
+ maas maas-region-controller juju python-pip python-psutil python-openstackclient \
python-congressclient gsutil charm-tools pastebinit python-jinja2 sshpass \
- openssh-server vlan ipmitool jq expect -y
+ openssh-server vlan ipmitool jq expect snap -y
+
+#sudo apt-get install snap -y
+#sudo snap install maas --classic
+#sudo snap install juju --classic
sudo pip install --upgrade pip
@@ -130,7 +134,11 @@ echo "... Deployment of maas Started ...."
# define the pool and try to start even though its already exist.
# For fresh install this may or may not there.
-sudo adduser $USER libvirtd
+#some system i am seeing libvirt and some have libvirtd looks like libvirt-bin is
+#keep switching so lets try both.
+
+sudo adduser $USER libvirtd || true
+sudo adduser $USER libvirt || true
sudo virsh pool-define-as default --type dir --target /var/lib/libvirt/images/ || true
sudo virsh pool-start default || true
sudo virsh pool-autostart default || true
@@ -138,10 +146,10 @@ sudo virsh pool-autostart default || true
# In case of virtual install set network
if [ "$virtinstall" -eq 1 ]; then
sudo virsh net-dumpxml default > default-net-org.xml
- sudo sed -i '/dhcp/d' default-net-org.xml
- sudo sed -i '/range/d' default-net-org.xml
- sudo virsh net-define default-net-org.xml
+ sed -i '/dhcp/d' default-net-org.xml
+ sed -i '/range/d' default-net-org.xml
sudo virsh net-destroy default
+ sudo virsh net-define default-net-org.xml
sudo virsh net-start default
rm -f default-net-org.xml
fi
@@ -294,8 +302,12 @@ setupspacenetwork(){
*) JUJU_SPACE='default'; DHCP='OFF'; echo " >>> Unknown SPACE" ;;
esac
JUJU_SPACE_ID=$(maas $PROFILE spaces read | jq -r ".[] | select(.name==\"$JUJU_SPACE\")".id)
- if ([ $NET_FABRIC_NAME ] && [ $NET_FABRIC_NAME != "null" ]); then
- maas $PROFILE subnet update $SPACE_CIDR space=$JUJU_SPACE_ID
+ JUJU_VLAN_VID=$(maas $PROFILE subnets read | jq -r ".[] | select(.name==\"$SPACE_CIDR\")".vlan.vid)
+ NET_FABRIC_ID=$(maas $PROFILE fabric read $NET_FABRIC_NAME | jq -r ".id")
+ if ([ $NET_FABRIC_ID ] && [ $NET_FABRIC_ID != "null" ]); then
+ if ([ $JUJU_VLAN_VID ] && [ $JUJU_VLAN_VID != "null" ]); then
+ maas $PROFILE vlan update $NET_FABRIC_ID $JUJU_VLAN_VID space=$JUJU_SPACE_ID
+ fi
fi
if ([ $type == "admin" ]); then
# If we have a network, we create it
@@ -397,12 +409,12 @@ addnodes(){
units=$(($units - 1));
NODE_NAME=`cat labconfig.json | jq ".lab.racks[].nodes[$units].name" | cut -d \" -f 2 `
- sudo virt-install --connect $VIRSHURL --name $NODE_NAME --ram 8192 --cpu host --vcpus 4 \
+ virt-install --connect $VIRSHURL --name $NODE_NAME --ram 8192 --cpu host --vcpus 4 \
--disk size=120,format=qcow2,bus=virtio,cache=directsync,io=native,pool=default \
$netw $netw --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee $NODE_NAME
nodemac=`grep "mac address" $NODE_NAME | head -1 | cut -d '"' -f 2`
- sudo virsh -c $VIRSHURL define --file $NODE_NAME
+ virsh -c $VIRSHURL define --file $NODE_NAME
rm -f $NODE_NAME
maas $PROFILE machines create autodetect_nodegroup='yes' name=$NODE_NAME \
tags='control compute' hostname=$NODE_NAME power_type='virsh' mac_addresses=$nodemac \
@@ -431,6 +443,8 @@ addnodes(){
done
fi
+ maas $PROFILE pods create type=virsh power_address="$VIRSHURL" power_user=$USER
+
# make sure nodes are added into MAAS and none of them is in commisoning state
while [ "$(maas $PROFILE nodes read | grep Commissioning )" ];
do