diff options
author | Cristina Pauna <cristina.pauna@enea.com> | 2018-02-01 15:30:01 +0200 |
---|---|---|
committer | Cristina Pauna <cristina.pauna@enea.com> | 2018-02-05 18:25:21 +0200 |
commit | 32bba0e639eaca254e18b5e4d36969f21f9e3d0e (patch) | |
tree | 847969c5c4a18eedc66aef74060647d2b9b1dad5 /docs/release/userguide/userguide.rst | |
parent | f7a780a85429d1975e3109e67760865d2ee2226d (diff) |
[docs] Remove redundant information
Armband and fuel documentation have been the same for E release.
Instead of duplicating everyhting, make references to Fuel from armband
docs. I removed the scenario folder as it should not have been here in the
first place, otherwise the folder structure is kept as is.
JIRA: ARMBAND-357
Change-Id: I060f22aee60713cabfd09ccf2fc0201e68a03c2a
Signed-off-by: Cristina Pauna <cristina.pauna@enea.com>
Diffstat (limited to 'docs/release/userguide/userguide.rst')
-rw-r--r-- | docs/release/userguide/userguide.rst | 326 |
1 files changed, 10 insertions, 316 deletions
diff --git a/docs/release/userguide/userguide.rst b/docs/release/userguide/userguide.rst index 2b46a84a..7b86c8ed 100644 --- a/docs/release/userguide/userguide.rst +++ b/docs/release/userguide/userguide.rst @@ -6,319 +6,13 @@ Abstract ======== -This document contains details about how to use OPNFV Fuel - Euphrates -release - after it was deployed. For details on how to deploy check the -installation instructions in the :ref:`references` section. - -This is an unified documentation for both x86_64 and aarch64 -architectures. All information is common for both architectures -except when explicitly stated. - - - -================ -Network Overview -================ - -Fuel uses several networks to deploy and administer the cloud: - -+------------------+-------------------+---------------------------------------------------------+ -| Network name | Deploy Type | Description | -| | | | -+==================+===================+=========================================================+ -| **PXE/ADMIN** | baremetal only | Used for booting the nodes via PXE | -+------------------+-------------------+---------------------------------------------------------+ -| **MCPCONTROL** | baremetal & | Used to provision the infrastructure VMs (Salt & MaaS). | -| | virtual | On virtual deploys, it is used for Admin too (on target | -| | | VMs) leaving the PXE/Admin bridge unused | -+------------------+-------------------+---------------------------------------------------------+ -| **Mgmt** | baremetal & | Used for internal communication between | -| | virtual | OpenStack components | -+------------------+-------------------+---------------------------------------------------------+ -| **Internal** | baremetal & | Used for VM data communication within the | -| | virtual | cloud deployment | -+------------------+-------------------+---------------------------------------------------------+ -| **Public** | baremetal & | Used to provide Virtual IPs for public endpoints | -| | virtual | that are used to connect to OpenStack services APIs. | -| | | Used by Virtual machines to access the Internet | -+------------------+-------------------+---------------------------------------------------------+ - - -These networks - except mcpcontrol - can be linux bridges configured before the deploy on the -Jumpserver. If they don't exists at deploy time, they will be created by the scripts as virsh -networks. - -Mcpcontrol exists only on the Jumpserver and needs to be virtual because a DHCP server runs -on this network and associates static host entry IPs for Salt and Maas VMs. - - - -=================== -Accessing the Cloud -=================== - -Access to any component of the deployed cloud is done from Jumpserver to user *ubuntu* with -ssh key */var/lib/opnfv/mcp.rsa*. The example below is a connection to Salt master. - - .. code-block:: bash - - $ ssh -o StrictHostKeyChecking=no -i /var/lib/opnfv/mcp.rsa -l ubuntu 10.20.0.2 - -**Note**: The Salt master IP is not hard set, it is configurable via INSTALLER_IP during deployment - - -The Fuel baremetal deploy has a Virtualized Control Plane (VCP) which means that the controller -services are installed in VMs on the baremetal targets (kvm servers). These VMs can also be -accessed with virsh console: user *opnfv*, password *opnfv_secret*. This method does not apply -to infrastructure VMs (Salt master and MaaS). - -The example below is a connection to a controller VM. The connection is made from the baremetal -server kvm01. - - .. code-block:: bash - - $ ssh -o StrictHostKeyChecking=no -i /var/lib/opnfv/mcp.rsa -l ubuntu x.y.z.141 - ubuntu@kvm01:~$ virsh console ctl01 - -User *ubuntu* has sudo rights. User *opnfv* has sudo rights only on aarch64 deploys. - - -============================= -Exploring the Cloud with Salt -============================= - -To gather information about the cloud, the salt commands can be used. It is based -around a master-minion idea where the salt-master pushes config to the minions to -execute actions. - -For example tell salt to execute a ping to 8.8.8.8 on all the nodes. - -.. figure:: img/saltstack.png - -Complex filters can be done to the target like compound queries or node roles. -For more information about Salt see the :ref:`references` section. - -Some examples are listed below. Note that these commands are issued from Salt master -with *root* user. - - -#. View the IPs of all the components - - .. code-block:: bash - - root@cfg01:~$ salt "*" network.ip_addrs - cfg01.baremetal-mcp-ocata-odl-ha.local: - - 10.20.0.2 - - 172.16.10.100 - mas01.baremetal-mcp-ocata-odl-ha.local: - - 10.20.0.3 - - 172.16.10.3 - - 192.168.11.3 - ......................... - - -#. View the interfaces of all the components and put the output in a file with yaml format - - .. code-block:: bash - - root@cfg01:~$ salt "*" network.interfaces --out yaml --output-file interfaces.yaml - root@cfg01:~# cat interfaces.yaml - cfg01.baremetal-mcp-ocata-odl-ha.local: - enp1s0: - hwaddr: 52:54:00:72:77:12 - inet: - - address: 10.20.0.2 - broadcast: 10.20.0.255 - label: enp1s0 - netmask: 255.255.255.0 - inet6: - - address: fe80::5054:ff:fe72:7712 - prefixlen: '64' - scope: link - up: true - ......................... - - -#. View installed packages in MaaS node - - .. code-block:: bash - - root@cfg01:~# salt "mas*" pkg.list_pkgs - mas01.baremetal-mcp-ocata-odl-ha.local: - ---------- - accountsservice: - 0.6.40-2ubuntu11.3 - acl: - 2.2.52-3 - acpid: - 1:2.0.26-1ubuntu2 - adduser: - 3.113+nmu3ubuntu4 - anerd: - 1 - ......................... - - -#. Execute any linux command on all nodes (list the content of */var/log* in this example) - - .. code-block:: bash - - root@cfg01:~# salt "*" cmd.run 'ls /var/log' - cfg01.baremetal-mcp-ocata-odl-ha.local: - alternatives.log - apt - auth.log - boot.log - btmp - cloud-init-output.log - cloud-init.log - ......................... - - -#. Execute any linux command on nodes using compound queries filter - - .. code-block:: bash - - root@cfg01:~# salt -C '* and cfg01*' cmd.run 'ls /var/log' - cfg01.baremetal-mcp-ocata-odl-ha.local: - alternatives.log - apt - auth.log - boot.log - btmp - cloud-init-output.log - cloud-init.log - ......................... - - -#. Execute any linux command on nodes using role filter - - .. code-block:: bash - - root@cfg01:~# salt -I 'nova:compute' cmd.run 'ls /var/log' - cmp001.baremetal-mcp-ocata-odl-ha.local: - alternatives.log - apache2 - apt - auth.log - btmp - ceilometer - cinder - cloud-init-output.log - cloud-init.log - ......................... - - - -=================== -Accessing Openstack -=================== - -Once the deployment is complete, Openstack CLI is accessible from controller VMs (ctl01..03). -Openstack credentials are at */root/keystonercv3*. - - .. code-block:: bash - - root@ctl01:~# source keystonercv3 - root@ctl01:~# openstack image list - +--------------------------------------+-----------------------------------------------+--------+ - | ID | Name | Status | - +======================================+===============================================+========+ - | 152930bf-5fd5-49c2-b3a1-cae14973f35f | CirrosImage | active | - | 7b99a779-78e4-45f3-9905-64ae453e3dcb | Ubuntu16.04 | active | - +--------------------------------------+-----------------------------------------------+--------+ - - -The OpenStack Dashboard, Horizon is available at http://<controller VIP>:8078, e.g. http://10.16.0.101:8078. -The administrator credentials are *admin*/*opnfv_secret*. - -.. figure:: img/horizon_login.png - - -A full list of IPs/services is available at <proxy public VIP>:8090 for baremetal deploys. - -.. figure:: img/salt_services_ip.png - -For Virtual deploys, the most commonly used IPs are in the table below. - -+-----------+--------------+---------------+ -| Component | IP | Default value | -+===========+==============+===============+ -| gtw01 | x.y.z.110 | 172.16.10.110 | -+-----------+--------------+---------------+ -| ctl01 | x.y.z.100 | 172.16.10.100 | -+-----------+--------------+---------------+ -| cmp001 | x.y.z.105 | 172.16.10.105 | -+-----------+--------------+---------------+ -| cmp002 | x.y.z.106 | 172.16.10.106 | -+-----------+--------------+---------------+ - - -============================= -Reclass model viewer tutorial -============================= - - -In order to get a better understanding on the reclass model Fuel uses, the `reclass-doc -<https://github.com/jirihybek/reclass-doc>`_ can be used to visualise the reclass model. -A simplified installation can be done with the use of a docker ubuntu container. This -approach will avoid installing packages on the host, which might collide with other packages. -After the installation is done, a webbrowser on the host can be used to view the results. - -**NOTE**: The host can be any device with Docker package already installed. - The user which runs the docker needs to have root priviledges. - - -**Instructions** - - -#. Create a new directory at any location - - .. code-block:: bash - - $ mkdir -p modeler - - -#. Place fuel repo in the above directory - - .. code-block:: bash - - $ cd modeler - $ git clone https://gerrit.opnfv.org/gerrit/fuel && cd fuel - - -#. Create a container and mount the above host directory - - .. code-block:: bash - - $ docker run --privileged -it -v <absolute_path>/modeler:/host ubuntu bash - - -#. Install all the required packages inside the container. - - .. code-block:: bash - - $ apt-get update - $ apt-get install -y npm nodejs - $ npm install -g reclass-doc - $ cd /host/fuel/mcp/reclass - $ ln -s /usr/bin/nodejs /usr/bin/node - $ reclass-doc --output /host /host/fuel/mcp/reclass - - -#. View the results from the host by using a browser. The file to open should be now at modeler/index.html - - .. figure:: img/reclass_doc.png - - -.. _references: - -========== -References -========== - -1) `Installation instructions <http://docs.opnfv.org/en/stable-euphrates/submodules/fuel/docs/release/installation/installation.instruction.html>`_ -2) `Saltstack Documentation <https://docs.saltstack.com/en/latest/topics>`_ -3) `Saltstack Formulas <http://salt-formulas.readthedocs.io/en/latest/develop/overview-reclass.html>`_ - - +Armband project aims to integrate and test all aspects of OPNFV releases +on ARM-based servers. The goal is to replicate all OPNFV software build, +continuous integration, lab provisioning, and testing processes of each +standard release OPNFV, such that the release can be available on both +Intel Architecture-based and ARM Architecture-based servers. + +The armband repo contains the patches necessary for Fuel installer to run on +aarch64 hardware. For more information on how to use Fuel@OPNFV - Euphrates +release - after it was deployed check +:ref:`fuel-release-userguide-label` |