summaryrefslogtreecommitdiffstats
path: root/docs/release
diff options
context:
space:
mode:
authorNarinder Gupta <narinder.gupta@canonical.com>2017-03-24 15:13:17 +0000
committerGerrit Code Review <gerrit@opnfv.org>2017-03-24 15:13:17 +0000
commitb2c77af4a4d474b23428575f801a87b374fc45be (patch)
tree3abfa0aed339783d56854c2b7172acee5427c384 /docs/release
parent4afffb76ced3bafae18e92a80137c25f969898f2 (diff)
parent36a495e8d20149811d13f6ba944ecc30a393aed4 (diff)
Merge "changing the user guide draft version."
Diffstat (limited to 'docs/release')
-rw-r--r--docs/release/userguide/userguide.rst744
1 files changed, 271 insertions, 473 deletions
diff --git a/docs/release/userguide/userguide.rst b/docs/release/userguide/userguide.rst
index ce949819..d010a488 100644
--- a/docs/release/userguide/userguide.rst
+++ b/docs/release/userguide/userguide.rst
@@ -1,9 +1,9 @@
Introduction
============
-This document will explain how to install OPNFV Brahmaputra with JOID including installing JOID, configuring JOID for your environment, and deploying OPNFV with different SDN solutions in HA, or non-HA mode. Prerequisites include
+This document will explain how to install OPNFV Danube with JOID including installing JOID, configuring JOID for your environment, and deploying OPNFV with different SDN solutions in HA, or non-HA mode. Prerequisites include
-- An Ubuntu 14.04 LTS Server Jumphost
+- An Ubuntu 16.04 LTS Server Jumphost
- Minimum 2 Networks per Pharos requirement
- One for the administrative network with gateway to access the Internet
@@ -19,21 +19,14 @@ This document will explain how to install OPNFV Brahmaputra with JOID including
- Hard Disk: 1 (250GB)
- NIC: eth0 (Admin, Management), eth1 (external network)
- - Control Node x 3, minimum H/W configuration:
+ - Control and Compute Nodes x 5, minimum H/W configuration:
- CPU cores: 16
- Memory: 32GB
- - Hard Disk: 1 (500GB)
+ - Hard Disk: 2 (500GB) prefer SSD
- NIC: eth0 (Admin, Management), eth1 (external network)
- - Compute Node x 2, minimum H/W configuration:
-
- - CPU cores: 16
- - Memory: 32GB
- - Hard Disk: 1 (1TB), this includes the space for Ceph.
- - NIC: eth0 (Admin, Management), eth1 (external network)
-
-**OTE**: Above configuration is minimum. For better performance and usage of the OpenStack, please consider higher specs for all nodes.
+**NOTE**: Above configuration is minimum. For better performance and usage of the OpenStack, please consider higher specs for all nodes.
Make sure all servers are connected to top of rack switch and configured accordingly. No DHCP server should be up and configured. Configure gateways only on eth0 and eth1 networks to access the network outside your lab.
@@ -61,22 +54,22 @@ For more info on Juju and MAAS, please visit https://jujucharms.com/ and http://
Typical JOID Setup
^^^^^^^^^^^^^^^^^^
-The MAAS server is installed and configured in a VM on the Ubuntu 14.04 LTS Jump Host with
+The MAAS server is installed and configured on Jumphost with Ubuntu 16.04 LTS with
access to the Internet. Another VM is created to be managed by MAAS as a bootstrap node
for Juju. The rest of the resources, bare metal or virtual, will be registered and
provisioned in MAAS. And finally the MAAS environment details are passed to Juju for use.
Installation
============
-We will use MAAS-deployer to automate the deployment of MAAS clusters for use as a Juju provider. MAAS-deployer uses a set of configuration files and simple commands to build a MAAS cluster using virtual machines for the region controller and bootstrap hosts and automatically commission nodes as required so that the only remaining step is to deploy services with Juju. For more information about the maas-deployer, please see https://launchpad.net/maas-deployer.
+We will use 03-maasdeploy.sh to automate the deployment of MAAS clusters for use as a Juju provider. MAAS-deployer uses a set of configuration files and simple commands to build a MAAS cluster using virtual machines for the region controller and bootstrap hosts and automatically commission nodes as required so that the only remaining step is to deploy services with Juju. For more information about the maas-deployer, please see https://launchpad.net/maas-deployer.
Configuring the Jump Host
^^^^^^^^^^^^^^^^^^^^^^^^^
Let's get started on the Jump Host node.
-The MAAS server is going to be installed and configured in a virtual machine. We need to create bridges on the Jump Host prior to setting up the MAAS-deployer.
+The MAAS server is going to be installed and configured on a Jumphost machine. We need to create bridges on the Jump Host prior to setting up the MAAS.
-**OTE**: For all the commands in this document, please do not use a ‘root’ user account to run. Please create a non root user account. We recommend using the ‘ubuntu’ user.
+**NOTE**: For all the commands in this document, please do not use a ‘root’ user account to run. Please create a non root user account. We recommend using the ‘ubuntu’ user.
Install the bridge-utils package on the Jump Host and configure a minimum of two bridges, one for the Admin network, the other for the Public network:
@@ -114,7 +107,7 @@ Install the bridge-utils package on the Jump Host and configure a minimum of two
**NOTE**: The Ethernet device names can vary from one installation to another. Please change the Ethernet device names according to your environment.
-MAAS-deployer has been integrated in the JOID project. To get the JOID code, please run
+MAAS has been integrated in the JOID project. To get the JOID code, please run
::
@@ -128,248 +121,104 @@ To set up your own environment, create a directory in joid/ci/maas/<company name
::
$ cd joid/ci
- $ mkdir -p maas/myown/pod
- $ cp maas/juniper/pod1/deployment.yaml maas/myown/pod/
-
-Now let's configure MAAS-deployer by editing the deployment.yaml file. Let's review each section. We will use the Juniper pod deployment.yaml as an example.
-
-::
-
- # This file defines the deployment for the MAAS environment which is to be
- # deployed and automated.
- demo-maas:
- maas:
- # Defines the general setup for the MAAS environment, including the
- # username and password for the host as well as the MAAS server.
- user: ubuntu
- password: ubuntu
-
-'demo-maas' is the environment name we set, it will be used by Juju. The username and password will be the login credentials for the MAAS server VM and also for the MAAS server web UI.
+ $ mkdir -p ../labconfig/myown/pod
+ $ cp ../labconfig/cengn/pod2/labconfig.yaml ../labconfig/myown/pod/
-::
-
- # Contains the virtual machine parameters for creating the MAAS virtual
- # server. Here you can configure the name of the virsh domain, the
- # parameters for how the network is attached.
- name: opnfv-maas-juniper
- interfaces: ['bridge=brAdm,model=virtio', 'bridge=brPublic,model=virtio']
- memory: 4096
- vcpus: 1
- arch: amd64
- pool: default
- disk_size: 160G
-
-When it's configured, you will see a KVM VM created and named 'opnfv-maas-juniper' on the
-Jump Host with 2 network interfaces configured and connected to brAdm and brPublic on the
-host. You may want to increase the vcpu number and disk size for the VM depending on the
-resources.
-
-::
-
- # Apt http proxy setting(s)
- apt_http_proxy:
-
- apt_sources:
- - ppa:maas/stable
- - ppa:juju/stable
-
-If in your environment uses an http proxy, please enter its information here. In addition, add the MAAS and Juju PPA locations here.
-
-::
-
- # Virsh power settings
- # Specifies the uri and keys to use for virsh power control of the
- # juju virtual machine. If the uri is omitted, the value for the
- # --remote is used. If no power settings are desired, then do not
- # supply the virsh block.
- virsh:
- rsa_priv_key: /home/ubuntu/.ssh/id_rsa
- rsa_pub_key: /home/ubuntu/.ssh/id_rsa.pub
- uri: qemu+ssh://ubuntu@172.16.50.51/system
-
- # Defines the IP Address that the configuration script will use
- # to access the MAAS controller via SSH.
- ip_address: 172.16.50.50
-
-This section defines MAAS server IP (172.16.50.50) and the virsh power settings. The Juju bootstrap VM is defined later.
+Now let's configure labconfig.yaml file. Please modify the sections in the labconfig as per your lab configuration.
::
- # This section allows the user to set a series of options on the
- # MAAS server itself. The list of config options can be found in
- # the upstream MAAS documentation:
- # - http://maas.ubuntu.com/docs/api.html#maas-server
- settings:
- main_archive: http://us.archive.ubuntu.com/ubuntu
- upstream_dns: 8.8.8.8
- maas_name: juniperpod1
- # kernel_opts: "console=tty0 console=ttyS1,115200n8"
- # ntp_server: ntp.ubuntu.com
+lab:
+ ## Change the name of the lab you want maas name will get firmat as per location and rack name ##
+ location: myown
+ racks:
+ - rack: pod
-Here we specify some settings for the MAAS server itself. Once MAAS is deployed, you will find these settings on http://172.16.50.50/MAAS/settings/.
+ ## based on your lab hardware please fill it accoridngly. ##
+ # Define one network and control and two control, compute and storage
+ # and rest for compute and storage for backward compaibility. again
+ # server with more disks should be used for compute and storage only.
+ nodes:
+ # DCOMP4-B, 24cores, 64G, 2disk, 4TBdisk
+ - name: rack-2-m1
+ architecture: x86_64
+ roles: [network,control]
+ nics:
+ - ifname: eth0
+ spaces: [admin]
+ mac: ["0c:c4:7a:3a:c5:b6"]
+ - ifname: eth1
+ spaces: [floating]
+ mac: ["0c:c4:7a:3a:c5:b7"]
+ power:
+ type: ipmi
+ address: <bmc ip>
+ user: <bmc username>
+ pass: <bmc password>
+
+ ## repeate the above section for number of hardware nodes you have it.
+
+ ## define the floating IP range along with gateway IP to be used during the instance floating ips ##
+ floating-ip-range: 172.16.120.20,172.16.120.62,172.16.120.254,172.16.120.0/24
+ # Mutiple MACs seperated by space where MACs are from ext-ports across all network nodes.
+
+ ## interface name to be used for floating ips ##
+ # eth1 of m4 since tags for networking are not yet implemented.
+ ext-port: "eth1"
+ dns: 8.8.8.8
+ osdomainname:
+
+opnfv:
+ release: d
+ distro: xenial
+ type: nonha
+ openstack: newton
+ sdncontroller:
+ - type: nosdn
+ storage:
+ - type: ceph
+ ## define the maximum disk possible in your environment ##
+ disk: /dev/sdb
+ feature: odl_l2
+ ## Ensure the following configuration matches the bridge configuration on your jumphost
+ spaces:
+ - type: admin
+ bridge: brAdm
+ cidr: 10.120.0.0/24
+ gateway: 10.120.0.254
+ vlan:
+ - type: floating
+ bridge: brPublic
+ cidr: 172.16.120.0/24
+ gateway: 172.16.120.254
+
+::
+
+
+Next we will use the 03-maasdeploy.sh in joid/ci to kick off maas deployment.
+
+Starting MAAS depoyment
+^^^^^^^^^^^^^^^^^^^^^^^
+Now run the 03-maasdeploy.sh script with the environment you just created
::
- # This section is used to define the networking parameters for when
- # the node first comes up. It is fed into the meta-data cloud-init
- # configuration and is used to configure the networking piece of the
- # service. The contents of this section are written directly to the
- # /etc/network/interfaces file.
- #
- # Please note, this is slightly different than the
- # node-group-interfaces section below. This will configure the
- # machine's networking params, and the node-group-interfaces will
- # configure the maas node-group interfaces which is used for
- # controlling the dhcp, dns, etc.
- network_config: |
- auto lo
- iface lo inet loopback
-
- auto eth0
- iface eth0 inet static
- address 172.16.50.50
- netmask 255.255.255.0
- network 172.16.50.0
- broadcast 172.16.50.255
- dns-nameservers 8.8.8.8 127.0.0.1
-
- auto eth1
- iface eth1 inet static
- address 10.10.15.50
- netmask 255.255.240.0
- network 10.10.0.0
- broadcast 10.10.15.255
- gateway 10.10.10.1
-
-This section defines the MAAS server's network interfaces. Once MAAS is deployed, you will find this setting at /etc/network/interfaces in the MAAS VM.
-
-::
-
- # The node-group-interfaces section is used to configure the MAAS
- # network interfaces. Basic configuration is supported, such as which
- # device should be bound, the range of IP addresses, etc.
- # Note: this may contain the special identifiers:
- # ${maas_net} - the first 3 octets of the ipv4 address
- # ${maas_ip} - the ip address of the MAAS controller
- node_group_ifaces:
- - device: eth0
- ip: 172.16.50.50
- subnet_mask: 255.255.255.0
- broadcast_ip: 172.16.50.255
- router_ip: 172.16.50.50
- static_range:
- low: 172.16.50.60
- high: 172.16.50.90
- dynamic_range:
- low: 172.16.50.91
- high: 172.16.50.254
+ ~/joid/ci$ ./03-maasdeploy.sh custom ../labconfig/mylab/pod/labconfig.yaml
-This section configures the MAAS cluster controller. Here it configures the MAAS cluster
-to provide DHCP and DNS services on the eth0 interface with dynamic and static IP ranges
-defined. You should allocate enough IP addresses for bare metal hosts in the static IP
-range, and allocate as many as possible in the dynamic IP range.
-
-::
-
- # Defines the physical nodes which are added to the MAAS cluste
- # controller upon startup of the node.
- nodes:
- - name: 2-R4N4B2-control
- tags: control
- architecture: amd64/generic
- mac_addresses:
- - "0c:c4:7a:16:2a:70"
- power:
- type: ipmi
- address: 10.10.7.92
- user: ADMIN
- pass: ADMIN
- driver: LAN_2_0
-
- - name: 3-R4N3B1-compute
- tags: compute
- architecture: amd64/generic
- mac_addresses:
- - "0c:c4:7a:53:57:c2"
- power:
- type: ipmi
- address: 10.10.7.84
- user: ADMIN
- pass: ADMIN
- driver: LAN_2_0
- <snip>
-
-This section defines the physical nodes to be added to the MAAS cluster controller. For
-example, the first node here is named ‘2-R4N4B2-control’, with a tag 'control' and
-architecture specified as amd64/generic. You will need to know the MAC address of the
-network interface of the node where it can reach MAAS server; it's the network interface
-of the node to PXE boot on. You need to tell MAAS how to power control the node by
-providing the the BMC IP address and BMC admin credentials. MAAS power control not only
-supports IPMI v2.0, but also supports virsh, Cisco UCS manager, HP moonshot iLO, and
-Microsoft OCS, among others. Tag is used here with Juju constraints to make sure that a
-particular service gets deployed only on hardware with the tag you created. Later when we
-go through the Juju deploy bundle, you will see the constraints setting.
-
-::
-
- # Contains the virtual machine parameters for creating the Juju bootstrap
- # node virtual machine
- juju-bootstrap:
- name: bootstrap
- interfaces: ['bridge=brAdm,model=virtio', 'bridge=brPublic,model=virtio']
- memory: 4096
- vcpus: 2
- arch: amd64
- pool: default
- disk_size: 120G
-
-The last section of the example deployment.yaml file defines the Juju bootstrap VM node.
-When it's configured, you will see a KVM VM created and named 'juju-boostrap' on the Jump
-Host with 2 network interfaces configured and connected to brAdm and brPublic on the host.
-You may want to increase the vcpu number and disk size for the VM depending on the resources.
-
-We are now done providing all the information regarding the MAAS VM and Juju VM, and how
-nodes and how many of them will be registered in MAAS. This information is very important,
-if you have questions, please hop on to #opnfv-joid IRC channel on freenode to ask.
-
-Next we will use the 02-maasdeploy.sh in joid/ci to kick off maas-deployer. Before we do
-that, we will create an entry to tell maas-deployer what deployment.yaml file to use. Use
-your favorite editor to add an entry under the section case $1. In our example, this is
-what we add::
-
- 'juniperpod1' )
- cp maas/juniper/pod1/deployment.yaml ./deployment.yaml
- ;;
-
-**NOTE**: If your username is different from ‘ubuntu’, please change the ssh key section accordingly::
-
- #just make sure the ssh keys are added into maas for the current user
- sed --i "s@/home/ubuntu@$HOME@g" ./deployment.yaml
- sed --i "s@qemu+ssh://ubuntu@qemu+ssh://$USER@g" ./deployment.yaml
-
-Starting MAAS-deployer
-^^^^^^^^^^^^^^^^^^^^^^
-Now run the 02-maasdeploy.sh script with the environment you just created
-
-::
-
- ~/joid/ci$ ./02-maasdeploy.sh juniperpod1
-
-This will take approximately 40 minutes to couple of hours depending on your environment. This script will do the following:
-1. Create 2 VMs (KVM).
-2. Install MAAS in one of the VMs.
+This will take approximately 30 minutes to couple of hours depending on your environment. This script will do the following:
+1. Create 1 VM (KVM).
+2. Install MAAS on the Jumphost.
3. Configure MAAS to enlist and commission a VM for Juju bootstrap node.
4. Configure MAAS to enlist and commission bare metal servers.
+5. Download and load 16.04 images to be used by MAAS.
-When it's done, you should be able to view the MAAS webpage (in our example http://172.16.50.50/MAAS) and see 1 bootstrap node and bare metal servers in the 'Ready' state on the nodes page.
-
-Here is an example output of running 02-maasdeploy.sh: http://pastebin.ubuntu.com/15117137/
+When it's done, you should be able to view the MAAS webpage (in our example http://172.16.50.2/MAAS) and see 1 bootstrap node and bare metal servers in the 'Ready' state on the nodes page.
-Troubleshooting MAAS deployer
-^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+Troubleshooting MAAS deployment
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
During the installation process, please carefully review the error messages.
-Join IRC channel #opnfv-joid on freenode to ask question. After the issues are resolved, re-running 02-maasdeploy.sh will clean up the VMs created previously. There is no need to manually undo what’s been done.
+Join IRC channel #opnfv-joid on freenode to ask question. After the issues are resolved, re-running 03-maasdeploy.sh will clean up the VMs created previously. There is no need to manually undo what’s been done.
Deploying OPNFV
^^^^^^^^^^^^^^^
@@ -378,13 +227,17 @@ HA or non-HA mode. For OpenStack, it supports Juno and Liberty. For SDN, it supp
vSwitch, OpenContrail, OpenDaylight and ONOS (Open Network Operating System). In addition
to HA or non-HA mode, it also supports deploying the latest from the development tree (tip).
-The deploy.sh script in the joid/ci directoy will do all the work for you. For example, the following deploys OpenStack Liberty with OpenDaylight in a HA mode in the Intelpod7.
+The deploy.sh script in the joid/ci directoy will do all the work for you. For example, the following deploys OpenStack Newton with OpenvSwitch in a HA mode.
::
- ~/joid/ci$ ./deploy.sh -o liberty -s odl -t ha -l intelpod7 -f none
+ ~/joid/ci$ ./deploy.sh -o newton -s nosdn -t ha -l custom -f none -m openstack
+
+The deploy.sh script in the joid/ci directoy will do all the work for you. For example, the following deploys Kubernetes with Load balancer on the pod.
+
+::
-**NOTE: You will need to modify ~/joid/ci/01-deploybundle.sh to deploy to your own environment, explained later.**
+ ~/joid/ci$ ./deploy.sh -m openstack -f lb
Take a look at the deploy.sh script. You will find we support the following for each option::
@@ -398,69 +251,43 @@ Take a look at the deploy.sh script. You will find we support the following for
ha: HA mode of OpenStack.
tip: The tip of the development.
[-o]
- juno: OpenStack Juno version.
- liberty: OpenStack Liberty version.
+ mitak: OpenStack Mitaka version.
+ newton: OpenStack Newton version.
[-l]
- default: For virtual deployment where installation will be done on KVM created using ./02-maasdeploy.sh
- intelpod5: Install on bare metal OPNFV pod5 of the Intel lab.
- intelpod6: Install on bare metal OPNFV pod6 of the Intel lab.
- orangepod2: Install on bare metal OPNFV pod2 of the Orange lab.
- (other pods)
- Note: if you make changes as per your pod above then please use your pod.
+ default: For virtual deployment where installation will be done on KVM created using ./03-maasdeploy.sh
+ custom: Install on bare metal OPNFV defined by labconfig.yaml
[-f]
none: no special feature will be enabled.
ipv6: IPv6 will be enabled for tenant in OpenStack.
+ dpdk: dpdk will be enabled.
+ lxd: virt-type will be lxd.
+ dvr: DVR will be enabled.
+ lb: Load balancing in case of Kubernetes will be enabled.
+ [-d]
+ xenial: distro to be used is Xenial 16.04
+ [-a]
+ amd64: Only x86 architecture will be used. Future version will support arm64 as well.
+ [-m]
+ openstack: Openstack model will be deployed.
+ kubernetes: Kubernetes model will be deployed.
-The script will call 00-bootstrap.sh to bootstrap the Juju VM node, then it will call 01-deploybundle.sh with the corrosponding parameter values.
+The script will call 01-bootstrap.sh to bootstrap the Juju VM node, then it will call 02-deploybundle.sh with the corrosponding parameter values.
::
- ./01-deploybundle.sh $opnfvtype $openstack $opnfvlab $opnfvsdn $opnfvfeature
+ ./02-deploybundle.sh $opnfvtype $openstack $opnfvlab $opnfvsdn $opnfvfeature $opnfvdistro
-You will notice in the 01-deploybundle.sh, it copies over the charm bundle file based on the ha/nonha/tip setting::
- case "$1" in
- 'nonha' )
- cp $4/juju-deployer/ovs-$4-nonha.yaml ./bundles.yaml
- ;;
- 'ha' )
- cp $4/juju-deployer/ovs-$4-ha.yaml ./bundles.yaml
- ;;
- 'tip' )
- cp $4/juju-deployer/ovs-$4-tip.yaml ./bundles.yaml
- cp common/source/* ./
- sed -i -- "s|branch: master|branch: stable/$2|g" ./*.yaml
- ;;
- * )
- cp $4/juju-deployer/ovs-$4-nonha.yaml ./bundles.yaml
- ;;
- esac
+Python script GenBundle.py would be used to create bundle.yaml based on the template
+defined in the config_tpl/juju2/ directory.
-After the respective yaml file is copied over and renamed to bundle.yaml, in the next
-section, it will update the bundle.yaml based on your network configuration and
-environment. For example, for the Juniper pod 1, we need to change vip suffix from
-10.4.1.1 to 172.16.50.1, which is our admin network, and eth1 is on the public network.
-
-::
-
- 'juniperpod1' )
- sed -i -- 's/10.4.1.1/172.16.50.1/g' ./bundles.yaml
- sed -i -- 's/#ext-port: "eth1"/ext-port: "eth1"/g' ./bundles.yaml
- ;;
-
-**NOTE**: If you are using a separate data network, then add this line below along with other changes, which signify that network 10.4.9.0/24 will be used as the data network for openstack.
-
-::
-
- sed -i -- 's/#os-data-network: 10.4.8.0\/21/os-data-network: 10.4.9.0\/24/g' ./bundles.yaml
-
-By default debug is enabled in the deploy.sh script and error messages will be printed on the SSH terminal where you are running the scripts. It could take an hour to a couple of hours (maximum) to complete. Here is an example output of the deployment: http://pastebin.ubuntu.com/15006924/
+By default debug is enabled in the deploy.sh script and error messages will be printed on the SSH terminal where you are running the scripts. It could take an hour to a couple of hours (maximum) to complete.
You can check the status of the deployment by running this command in another terminal::
$ watch juju status --format tabular
-This will refresh the juju status output in tabular format every 2 seconds. Here is an example output of juju status --format tabular: http://pastebin.ubuntu.com/15134109/
+This will refresh the juju status output in tabular format every 2 seconds.
Next we will show you what Juju is deploying and to where, and how you can modify based on your own needs.
@@ -477,60 +304,19 @@ service, they can be used to deploy an entire workload, with working relations a
configuration. The use of bundles allows for easy repeatability and for sharing of
complex, multi-service deployments.
-For OPNFV, we have collected the charm bundles for each SDN deployment. They are stored in
-each SDN directory in ~/joid/ci. In each SDN folder, there are 3 bundle.yaml files, one
-for HA, one for non-HA, and the other for tip. For example for OpenDaylight::
-
- ~/joid/ci/odl/juju-deployer$ ls
- ovs-odl-ha.yaml ovs-odl-nonha.yaml ovs-odl-tip.yaml scripts
- ~/joid/ci/odl/juju-deployer$
-
-We use Juju-deployer to deploy a set of charms via a yaml configuration file. You can find the complete format guide for the Juju-deployer configuration file here: http://pythonhosted.org/juju-deployer/config.html
+For OPNFV, we have created the charm bundles for each SDN deployment. They are stored in
+each directory in ~/joid/ci.
-Let’s take a quick look at the ovs-odl-nonha.yaml to give you an idea about the charm bundle.
+We use Juju to deploy a set of charms via a yaml configuration file. You can find the complete format guide for the Juju configuration file here: http://pythonhosted.org/juju-deployer/config.html
-Assuming we are deploying OpenDayling with OpenStack Liberty in non-HA mode, according to the deploy.sh, we know it will run these two commands::
-
- juju-deployer -vW -d -t 3600 -c bundles.yaml trusty-liberty-nodes
- juju-deployer -vW -d -t 7200 -r 5 -c bundles.yaml trusty-liberty
-
-In the ovs-odl-nonha.yaml file, find the section of ‘trusty-liberty-nodes’ close to the bottom of the file::
-
- trusty-liberty-nodes:
- inherits: openstack-phase1
- overrides:
- series: trusty
-
-It inherits ‘openstack-phase1’, which you will find in the beginning of the file::
-
- openstack-phase1:
- series: trusty
- services:
- nodes-api:
- charm: "cs:trusty/ubuntu"
- num_units: 1
- constraints: tags=control
- nodes-compute:
- charm: "cs:trusty/ubuntu"
- num_units: 1
- constraints: tags=compute
- ntp:
- charm: "cs:trusty/ntp"
- relations:
- - - "ntp:juju-info"
- - "nodes-api:juju-info"
- - - "ntp:juju-info"
- - "nodes-compute:juju-info"
-
-In the ‘services’ subsection, here we deploy the ‘Ubuntu Trusty charm from the charm
-store,’ name the service ‘nodes-api,’ deploy just one unit, and assign a tag of ‘control’
-to this service. You can deploy the same charm and name it differently such as the second
+In the ‘services’ subsection, here we deploy the ‘Ubuntu Xenial charm from the charm
+store,’ You can deploy the same charm and name it differently such as the second
service ‘nodes-compute.’ The third service we deploy is named ‘ntp’ and is deployed from
the NTP Trusty charm from the Charm Store. The NTP charm is a subordinate charm, which is
designed for and deployed to the running space of another service unit.
The tag here is related to what we define in the deployment.yaml file for the
-MAAS-deployer. When ‘constraints’ is set, Juju will ask its provider, in this case MAAS,
+MAAS. When ‘constraints’ is set, Juju will ask its provider, in this case MAAS,
to provide a resource with the tags. In this case, Juju is asking one resource tagged with
control and one resource tagged with compute from MAAS. Once the resource information is
passed to Juju, Juju will start the installation of the specified version of Ubuntu.
@@ -540,15 +326,13 @@ and charms is you can define the relation of two services and all the service un
deployed will set up the relations accordingly. This makes scaling out a very easy task.
Here we add the relation between NTP and the two bare metal services.
-Once the relations are established, Juju-deployer considers the deployment complete and moves to the next.
+Once the relations are established, Juju considers the deployment complete and moves to the next.
::
- juju-deployer -vW -d -t 7200 -r 5 -c bundles.yaml trusty-liberty
+ juju deploy bundles.yaml
-It will start at the ‘trusty-liberty’ section, which inherits the ‘contrail’ section,
-which inherits the ‘openstack-phase2’ section. it follows the same services and relations
-format as above. We will take a look at another common service configuration next.
+It will start the deployment , which will retry the section,
::
@@ -569,11 +353,11 @@ To find out what other options there are for this particular charm, you can go t
Once the service unit is deployed, you can see the current configuration by running juju get::
- $ juju get nova-cloud-controller
+ $ juju config nova-cloud-controller
-You can change the value with juju set, for example::
+You can change the value with juju config, for example::
- $ juju set nova-cloud-controller network-manager=’FlatManager’
+ $ juju config nova-cloud-controller network-manager=’FlatManager’
Charms encapsulate the operation best practices. The number of options you need to configure should be at the minimum. The Juju Charm Store is a great resource to explore what a charm can offer you. Following the nova-cloud-controller charm example, here is the main page of the recommended charm on the Charm Store: https://jujucharms.com/nova-cloud-controller/trusty/66
@@ -585,7 +369,7 @@ Once juju-deployer is complete, use juju status --format tabular to verify that
Find the Openstack-dashboard IP address from the juju status output, and see if you can login via a web browser. The username and password is admin/openstack.
-Optionally, see if you can log in to the Juju GUI. The Juju GUI is on the Juju bootstrap node, which is the second VM you define in the 02-maasdeploy.sh file. The username and password is admin/admin.
+Optionally, see if you can log in to the Juju GUI. The Juju GUI is on the Juju bootstrap node, which is the second VM you define in the 03-maasdeploy.sh file. The username and password is admin/admin.
If you deploy OpenDaylight, OpenContrail or ONOS, find the IP address of the web UI and login. Please refer to each SDN bundle.yaml for the login username/password.
@@ -607,7 +391,7 @@ Example::
ubuntu@R4N4B1:~$ juju ssh nova-compute/0
Warning: Permanently added '172.16.50.60' (ECDSA) to the list of known hosts.
Warning: Permanently added '3-r4n3b1-compute.maas' (ECDSA) to the list of known hosts.
- Welcome to Ubuntu 14.04.1 LTS (GNU/Linux 3.13.0-77-generic x86_64)
+ Welcome to Ubuntu 16.04.1 LTS (GNU/Linux 3.13.0-77-generic x86_64)
* Documentation: https://help.ubuntu.com/
<skipped>
@@ -626,31 +410,6 @@ Once you resolve the error, go back to the jump host to rerun the charm hook wit
If you would like to start over, run juju destroy-environment <environment name> to release the resources, then you can run deploy.sh again.
-::
-
- $ juju destroy-environment demo-maas
- WARNING! this command will destroy the "demo-maas" environment (type: maas)
- This includes all machines, services, data and other resources.
-
- Continue [y/N]? y
- $
-
-If there is an error destroying the environment, use --force.
-
-::
-
- $ juju destroy-environment demo-maas --force
- $
-
-If the above command hangs, use Ctrl-C to get out of it, and manually remove the environment file in the ~/.juju/environments/ directory.
-
-::
-
- $ ls ~/.juju/environments/
- demo-maas.jenv
- $ sudo rm ~/.juju/environments/demo-maas.jenv
- $
-
The following are the common issues we have collected from the community:
@@ -658,16 +417,16 @@ The following are the common issues we have collected from the community:
::
- ./deploy.sh -o liberty -s odl -t ha -l intelpod5 -f none
+ ./deploy.sh -o newton -s nosdn -t ha -l custom -f none
-- If you have setup maas not with 02-maasdeply.sh then the ./clean.sh command could hang,
- the juju status command may hang because the correct MAAS API keys are not listed in
- environments.yaml, or environments.yaml does not exist in the current working directory.
- Solution: Please make sure you have an environments.yaml file under joid/ci directory
- and the correct MAAS API key has been listed.
+- If you have setup maas not with 03-maasdeploy.sh then the ./clean.sh command could hang,
+ the juju status command may hang because the correct MAAS API keys are not mentioned in
+ cloud listing for MAAS.
+ Solution: Please make sure you have an MAAS cloud listed using juju clouds.
+ and the correct MAAS API key has been added.
- Deployment times out:
use the command juju status --format=tabular and make sure all service containers receive an IP address and they are executing code. Ensure there is no service in the error state.
-- In case the cleanup process hangs,remove the files from the ~/.juju/ directory except environments.yaml and shutdown all nodes manually.
+- In case the cleanup process hangs,run the juju destroy-model command manually.
**Direct console access** via the OpenStack GUI can be quite helpful if you need to login to a VM but cannot get to it over the network.
It can be enabled by setting the ``console-access-protocol`` in the ``nova-cloud-controller`` to ``vnc``. One option is to directly edit the juju-deployer bundle and set it there prior to deploying OpenStack.
@@ -688,13 +447,12 @@ At the end of the deployment, the admin-openrc with OpenStack login credentials
::
- ~/joid/ci/cloud$ cat admin-openrc
+ ~/joid_config$ cat admin-openrc
export OS_USERNAME=admin
export OS_PASSWORD=openstack
export OS_TENANT_NAME=admin
export OS_AUTH_URL=http://172.16.50.114:5000/v2.0
- export OS_REGION_NAME=Canonical
- ~/joid/ci/cloud$
+ export OS_REGION_NAME=RegionOne
We have prepared some scripts to help your configure the OpenStack cloud that you just deployed. In each SDN directory, for example joid/ci/opencontrail, there is a ‘scripts’ folder where you can find the scripts. These scripts are created to help you configure a basic OpenStack Cloud to verify the cloud. For more information on OpenStack Cloud configuration, please refer to the OpenStack Cloud Administrator Guide: http://docs.openstack.org/user-guide-admin/. Similarly, for complete SDN configuration, please refer to the respective SDN administrator guide.
@@ -707,9 +465,8 @@ Let’s take a look at those for the Open vSwitch and briefly go through each sc
::
- ~/joid/ci/nosdn/juju-deployer/scripts$ ls
- cloud-setup.sh glance.sh openstack.sh
- ~/joid/ci/nosdn/juju-deployer/scripts$
+ ~/joid/juju$ ls
+ configure-juju-on-openstack get-cloud-images joid-configure-openstack
openstack.sh
~~~~~~~~~~~~
@@ -717,100 +474,141 @@ Let’s first look at ‘openstack.sh’. First there are 3 functions defined, c
::
- configOpenrc()
- {
+ configOpenrc() {
cat <<-EOF
- export OS_USERNAME=$1
- export OS_PASSWORD=$2
- export OS_TENANT_NAME=$3
- export OS_AUTH_URL=$4
- export OS_REGION_NAME=$5
- EOF
+ export SERVICE_ENDPOINT=$4
+ unset SERVICE_TOKEN
+ unset SERVICE_ENDPOINT
+ export OS_USERNAME=$1
+ export OS_PASSWORD=$2
+ export OS_TENANT_NAME=$3
+ export OS_AUTH_URL=$4
+ export OS_REGION_NAME=$5
+ EOF
}
- unitAddress()
- {
- juju status | python -c "import yaml; import sys; print yaml.load(sys.stdin)[\"services\"][\"$1\"][\"units\"][\"$1/$2\"][\"public-address\"]" 2> /dev/null
+ unitAddress() {
+ if [[ "$jujuver" < "2" ]]; then
+ juju status --format yaml | python -c "import yaml; import sys; print yaml.load(sys.stdin)[\"services\"][\"$1\"][\"units\"][\"$1/$2\"][\"public-address\"]" 2> /dev/null
+ else
+ juju status --format yaml | python -c "import yaml; import sys; print yaml.load(sys.stdin)[\"applications\"][\"$1\"][\"units\"][\"$1/$2\"][\"public-address\"]" 2> /dev/null
+ fi
}
- unitMachine()
- {
- juju status | python -c "import yaml; import sys; print yaml.load(sys.stdin)[\"services\"][\"$1\"][\"units\"][\"$1/$2\"][\"machine\"]" 2> /dev/null
+ unitMachine() {
+ if [[ "$jujuver" < "2" ]]; then
+ juju status --format yaml | python -c "import yaml; import sys; print yaml.load(sys.stdin)[\"services\"][\"$1\"][\"units\"][\"$1/$2\"][\"machine\"]" 2> /dev/null
+ else
+ juju status --format yaml | python -c "import yaml; import sys; print yaml.load(sys.stdin)[\"applications\"][\"$1\"][\"units\"][\"$1/$2\"][\"machine\"]" 2> /dev/null
+ fi
}
The function configOpenrc() creates the OpenStack login credentials, the function unitAddress() finds the IP address of the unit, and the function unitMachine() finds the machine info of the unit.
::
- mkdir -m 0700 -p cloud
- controller_address=$(unitAddress keystone 0)
- configOpenrc admin openstack admin http://$controller_address:5000/v2.0 Canonical > cloud/admin-openrc
- chmod 0600 cloud/admin-openrc
+ create_openrc() {
+ keystoneIp=$(keystoneIp)
+ if [[ "$jujuver" < "2" ]]; then
+ adminPasswd=$(juju get keystone | grep admin-password -A 5 | grep value | awk '{print $2}' 2> /dev/null)
+ else
+ adminPasswd=$(juju config keystone | grep admin-password -A 5 | grep value | awk '{print $2}' 2> /dev/null)
+ fi
+
+ configOpenrc admin $adminPasswd admin http://$keystoneIp:5000/v2.0 RegionOne > ~/joid_config/admin-openrc
+ chmod 0600 ~/joid_config/admin-openrc
+ }
-This creates a folder named ‘cloud’, finds the IP address of the keystone unit 0, feeds in
-the OpenStack admin credentials to a new file name ‘admin-openrc’ in the ‘cloud’ folder
+This finds the IP address of the keystone unit 0, feeds in the OpenStack admin
+credentials to a new file name ‘admin-openrc’ in the ‘~/joid_config/’ folder
and change the permission of the file. It’s important to change the credentials here if
you use a different password in the deployment Juju charm bundle.yaml.
::
- machine=$(unitMachine glance 0)
- juju scp glance.sh cloud/admin-openrc $machine:
- juju run --machine $machine ./glance.sh
-
-This section first finds the machine ID of the glance service unit 0, transfers the
-glance.sh and admin-openrc files over to the glance unit 0, and then run the glance.sh in
-the glance unit 0. We will take a look at the glance.sh in the next section.
+ neutron net-show ext-net > /dev/null 2>&1 || neutron net-create ext-net \
+ --router:external=True \
+ --provider:network_type flat \
+ --provider:physical_network physnet1
::
+ neutron subnet-show ext-subnet > /dev/null 2>&1 || neutron subnet-create ext-net \
+ --name ext-subnet --allocation-pool start=$EXTNET_FIP,end=$EXTNET_LIP \
+ --disable-dhcp --gateway $EXTNET_GW $EXTNET_NET
- machine=$(unitMachine nova-cloud-controller 0)
- juju scp cloud-setup.sh cloud/admin-openrc ~/.ssh/id_rsa.pub $machine:
- juju run --machine $machine ./cloud-setup.sh
-
-This section first finds the the machine ID of the nova-cloud-controller service unit 0,
-transfers 3 files over to the nova-cloud-controller unit 0, and then runs the
-cloud-setup.sh in the nova-cloud-controller unit 0. We will take a look at the
-cloud-setup.sh following glance.sh.
-
-glance.sh
-~~~~~~~~~
+This section will create the ext-net and ext-subnet for defining the for floating ips.
::
- . ~/admin-openrc
+ openstack congress datasource create nova "nova" \
+ --config username=$OS_USERNAME \
+ --config tenant_name=$OS_TENANT_NAME \
+ --config password=$OS_PASSWORD \
+ --config auth_url=http://$keystoneIp:5000/v2.0
-First, this script sources the admin-openrc file.
+This section will create the congress datasource for various services.
+Each service datasource will have entry in the file.
-::
+get-cloud-images
+~~~~~~~~~~~~~~~~
- wget -P /tmp/images http://download.cirros-cloud.net/0.3.3/cirros-0.3.3-x86_64-disk.img
- wget -P /tmp/images http://cloud-images.ubuntu.com/trusty/current/trusty-server-cloudimg-amd64-disk1.img
+::
-Download two images, Cirros and Ubuntu Trusty cloud image to /tmp/images folder.
+ folder=/srv/data/
+ sudo mkdir $folder || true
-::
+ if grep -q 'virt-type: lxd' bundles.yaml; then
+ URLS=" \
+ http://download.cirros-cloud.net/0.3.4/cirros-0.3.4-x86_64-lxc.tar.gz \
+ http://cloud-images.ubuntu.com/xenial/current/xenial-server-cloudimg-amd64-root.tar.gz "
- glance image-create --name "cirros-0.3.3-x86_64" --file /tmp/images/cirros-0.3.3-x86_64-disk.img --disk-format qcow2 --container-format bare --progress
- glance image-create --name "ubuntu-trusty-daily" --file /tmp/images/trusty-server-cloudimg-amd64-disk1.img --disk-format qcow2 --container-format bare --progress
- rm -rf /tmp/images
+ else
+ URLS=" \
+ http://cloud-images.ubuntu.com/precise/current/precise-server-cloudimg-amd64-disk1.img \
+ http://cloud-images.ubuntu.com/trusty/current/trusty-server-cloudimg-amd64-disk1.img \
+ http://cloud-images.ubuntu.com/xenial/current/xenial-server-cloudimg-amd64-disk1.img \
+ http://mirror.catn.com/pub/catn/images/qcow2/centos6.4-x86_64-gold-master.img \
+ http://cloud.centos.org/centos/7/images/CentOS-7-x86_64-GenericCloud.qcow2 \
+ http://download.cirros-cloud.net/0.3.4/cirros-0.3.4-x86_64-disk.img "
+ fi
-Use the glance python client to upload those two images, and finally remove those images from the local file system.
+ for URL in $URLS
+ do
+ FILENAME=${URL##*/}
+ if [ -f $folder/$FILENAME ];
+ then
+ echo "$FILENAME already downloaded."
+ else
+ wget -O $folder/$FILENAME $URL
+ fi
+ done
-If you wish to use different images, please change the image download links and filenames here accordingly.`
+This section of the file will download the images to jumphost if not found to be used with
+openstack VIM.
**NOTE**: The image downloading and uploading might take too long and time out. In this case, use juju ssh glance/0 to log in to the glance unit 0 and run the script again, or manually run the glance commands.
-cloud-setup.sh
-~~~~~~~~~~~~~~
+joid-configure-openstack
+~~~~~~~~~~~~~~~~~~~~~~~~
::
- . ~/admin-openrc
+ source ~/joid_config/admin-openrc
First, source the the admin-openrc file.
::
+ #Upload images to glance
+ glance image-create --name="Xenial LXC x86_64" --visibility=public --container-format=bare --disk-format=root-tar --property architecture="x86_64" < /srv/data/xenial-server-cloudimg-amd64-root.tar.gz
+ glance image-create --name="Cirros LXC 0.3" --visibility=public --container-format=bare --disk-format=root-tar --property architecture="x86_64" < /srv/data/cirros-0.3.4-x86_64-lxc.tar.gz
+ glance image-create --name="Trusty x86_64" --visibility=public --container-format=ovf --disk-format=qcow2 < /srv/data/trusty-server-cloudimg-amd64-disk1.img
+ glance image-create --name="Xenial x86_64" --visibility=public --container-format=ovf --disk-format=qcow2 < /srv/data/xenial-server-cloudimg-amd64-disk1.img
+ glance image-create --name="CentOS 6.4" --visibility=public --container-format=bare --disk-format=qcow2 < /srv/data/centos6.4-x86_64-gold-master.img
+ glance image-create --name="Cirros 0.3" --visibility=public --container-format=bare --disk-format=qcow2 < /srv/data/cirros-0.3.4-x86_64-disk.img
+
+upload the images into glane to be used for creating the VM.
+
+::
# adjust tiny image
nova flavor-delete m1.tiny
@@ -876,48 +674,48 @@ This section creates a router and connects this router to the two networks we ju
Finally, the script will request 10 floating IPs.
+configure-juju-on-openstack
+~~~~~~~~~~~~~~~~~~~~~~~~~~~
+
+This script can be used to do juju bootstrap on openstack so that Juju can be used as model tool to deploy the services and VNF on top of openstack using the JOID.
+
+
Appendix A: Single Node Deployment
==================================
-By default, running the script ./02-maasdeploy.sh will automatically create the KVM VMs on a single machine and configure everything for you.
+By default, running the script ./03-maasdeploy.sh will automatically create the KVM VMs on a single machine and configure everything for you.
::
- * )
- virtinstall=1
- ./cleanvm.sh
- cp maas/default/deployment.yaml ./deployment.yaml
- ;;
+ if [ ! -e ./labconfig.yaml ]; then
+ virtinstall=1
+ labname="default"
+ cp ../labconfig/default/labconfig.yaml ./
+ cp ../labconfig/default/deployconfig.yaml ./
-Please change ~/joid/ci/maas/default/deployment.yaml accordingly. The MAAS-deployer will do the following:
-1. Create 2 VMs (KVM).
-2. Install MAAS in one of the VMs.
-3. Configure MAAS to enlist and commission a VM for Juju bootstrap node.
+Please change joid/ci/labconfig/default/labconfig.yaml accordingly. The MAAS deployment script will do the following:
+1. Create bootstrap VM.
+2. Install MAAS on the jumphost.
+3. Configure MAAS to enlist and commission VM for Juju bootstrap node.
-Later, the 02-massdeploy.sh script will create two additional VMs and register them into the MAAS Server:
+Later, the 03-massdeploy.sh script will create three additional VMs and register them into the MAAS Server:
::
if [ "$virtinstall" -eq 1 ]; then
- # create two more VMs to do the deployment.
- sudo virt-install --connect qemu:///system --name node1-control --ram 8192 --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node1-control
- sudo virt-install --connect qemu:///system --name node2-compute --ram 8192 --vcpus 4 --disk size=120,format=qcow2,bus=virtio,io=native,pool=default --network bridge=virbr0,model=virtio --network bridge=virbr0,model=virtio --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee node2-compute
-
- node1controlmac=`grep "mac address" node1-control | head -1 | cut -d "'" -f 2`
- node2computemac=`grep "mac address" node2-compute | head -1 | cut -d "'" -f 2`
-
- sudo virsh -c qemu:///system define --file node1-control
- sudo virsh -c qemu:///system define --file node2-compute
-
- maas maas tags new name='control'
- maas maas tags new name='compute'
-
- controlnodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node1-control' tags='control' hostname='node1-control' power_type='virsh' mac_addresses=$node1controlmac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node1-control' | grep system_id | cut -d '"' -f 4 `
-
- maas maas tag update-nodes control add=$controlnodeid
-
- computenodeid=`maas maas nodes new autodetect_nodegroup='yes' name='node2-compute' tags='compute' hostname='node2-compute' power_type='virsh' mac_addresses=$node2computemac power_parameters_power_address='qemu+ssh://'$USER'@192.168.122.1/system' architecture='amd64/generic' power_parameters_power_id='node2-compute' | grep system_id | cut -d '"' -f 4 `
-
- maas maas tag update-nodes compute add=$computenodeid
+ sudo virt-install --connect qemu:///system --name $NODE_NAME --ram 8192 --cpu host --vcpus 4 \
+ --disk size=120,format=qcow2,bus=virtio,io=native,pool=default \
+ $netw $netw --boot network,hd,menu=off --noautoconsole --vnc --print-xml | tee $NODE_NAME
+
+ nodemac=`grep "mac address" $NODE_NAME | head -1 | cut -d '"' -f 2`
+ sudo virsh -c qemu:///system define --file $NODE_NAME
+ rm -f $NODE_NAME
+ maas $PROFILE machines create autodetect_nodegroup='yes' name=$NODE_NAME \
+ tags='control compute' hostname=$NODE_NAME power_type='virsh' mac_addresses=$nodemac \
+ power_parameters_power_address='qemu+ssh://'$USER'@'$MAAS_IP'/system' \
+ architecture='amd64/generic' power_parameters_power_id=$NODE_NAME
+ nodeid=$(maas $PROFILE machines read | jq -r '.[] | select(.hostname == '\"$NODE_NAME\"').system_id')
+ maas $PROFILE tag update-nodes control add=$nodeid || true
+ maas $PROFILE tag update-nodes compute add=$nodeid || true
fi