summaryrefslogtreecommitdiffstats
path: root/docs/installation-instructions
diff options
context:
space:
mode:
authorSofia Wallin <sofia.wallin@ericsson.com>2016-08-19 16:39:06 +0200
committerSofia Wallin <sofia.wallin@ericsson.com>2016-08-19 16:39:06 +0200
commit1e4742787c6c45b44f5e49de03ec8b1b1075c04e (patch)
treea58b62aaf2637676066e26517ecda0e1b07f6f23 /docs/installation-instructions
parent4862f3b57479d64c5408268cc2086b53e4e97b81 (diff)
Renamed install instructions folder according to docs directive
Change-Id: I8970be0bcbed501e43d6fb76af1278881274f69a Signed-off-by: Sofia Wallin <sofia.wallin@ericsson.com>
Diffstat (limited to 'docs/installation-instructions')
-rw-r--r--docs/installation-instructions/abstract.rst16
-rw-r--r--docs/installation-instructions/architecture.rst77
-rw-r--r--docs/installation-instructions/baremetal.rst242
-rw-r--r--docs/installation-instructions/index.rst27
-rw-r--r--docs/installation-instructions/introduction.rst42
-rw-r--r--docs/installation-instructions/references.rst38
-rw-r--r--docs/installation-instructions/requirements.rst79
-rw-r--r--docs/installation-instructions/verification.rst89
-rw-r--r--docs/installation-instructions/virtualinstall.rst69
9 files changed, 0 insertions, 679 deletions
diff --git a/docs/installation-instructions/abstract.rst b/docs/installation-instructions/abstract.rst
deleted file mode 100644
index 70814cea..00000000
--- a/docs/installation-instructions/abstract.rst
+++ /dev/null
@@ -1,16 +0,0 @@
-Abstract
-========
-
-This document describes how to install the Colorado release of OPNFV when
-using Apex as a deployment tool covering it's limitations, dependencies
-and required system resources.
-
-License
-=======
-Colorado release of OPNFV when using Apex as a deployment tool Docs
-(c) by Tim Rozet (Red Hat) and Dan Radez (Red Hat)
-
-Colorado release of OPNFV when using Apex as a deployment tool Docs
-are licensed under a Creative Commons Attribution 4.0 International License.
-You should have received a copy of the license along with this.
-If not, see <http://creativecommons.org/licenses/by/4.0/>.
diff --git a/docs/installation-instructions/architecture.rst b/docs/installation-instructions/architecture.rst
deleted file mode 100644
index c63da27c..00000000
--- a/docs/installation-instructions/architecture.rst
+++ /dev/null
@@ -1,77 +0,0 @@
-Triple-O Deployment Architecture
-================================
-
-Apex is based on the OpenStack Triple-O project as distributed by
-the RDO Project. It is important to understand the basics
-of a Triple-O deployment to help make decisions that will assist in
-successfully deploying OPNFV.
-
-Triple-O stands for OpenStack On OpenStack. This means that OpenStack
-will be used to install OpenStack. The target OPNFV deployment is an
-OpenStack cloud with NFV features built-in that will be deployed by a
-smaller all-in-one deployment of OpenStack. In this deployment
-methodology there are two OpenStack installations. They are referred
-to as the undercloud and the overcloud. The undercloud is used to
-deploy the overcloud.
-
-The undercloud is the all-in-one installation of OpenStack that includes
-baremetal provisioning capability. The undercloud will be deployed as a
-virtual machine on a jumphost. This VM is pre-built and distributed as part
-of the Apex RPM.
-
-The overcloud is OPNFV. Configuration will be passed into undercloud and
-the undercloud will use OpenStack's orchestration component, named Heat, to
-execute a deployment that will provision the target OPNFV nodes.
-
-OPNFV Scenario Architecture
-===========================
-
-OPNFV distinguishes different types of SDN controllers, deployment options, and
-features into "scenarios". These scenarios are universal across all OPNFV
-installers, although some may or may not be supported by each installer.
-
-The standard naming convention for a scenario is:
-<VIM platform>-<SDN type>-<feature>-<ha/noha>
-
-The only supported VIM type is "OS" (OpenStack), while SDN types can be any
-supported SDN controller. "feature" includes things like ovs_dpdk, sfc, etc.
-"ha" or "noha" determines if the deployment will be highly available. If "ha"
-is used at least 3 control nodes are required.
-
-OPNFV Scenarios in Apex
-=======================
-
-Apex provides pre-built scenario files in /etc/opnfv-apex which a user can
-select from to deploy the desired scenario. Simply pass the desired file to
-the installer as a (-d) deploy setting. Read further in the Apex documentation
-to learn more about invoking the deploy command. Below is quick reference
-matrix for OPNFV scenarios supported in Apex. Please refer to the respective
-OPNFV Docs documentation for each scenario in order to see a full scenario
-description. The following scenarios correspond to a supported <Scenario>.yaml
-deploy settings file:
-
-+-------------------------+------------+-----------------+
-| **Scenario** | **Owner** | **Known Issues**|
-+-------------------------+------------+-----------------+
-| os-nosdn-nofeature-ha | Apex | |
-+-------------------------+------------+-----------------+
-| os-nosdn-nofeature-noha | Apex | |
-+-------------------------+------------+-----------------+
-| os-nosdn-ovs-noha | OVS for NFV| |
-+-------------------------+------------+-----------------+
-| os-nosdn-fdio-noha | FDS | |
-+-------------------------+------------+-----------------+
-| os-odl_l2-nofeature-ha | Apex | |
-+-------------------------+------------+-----------------+
-| os-odl_l3-nofeature-ha | Apex | APEX-112 |
-+-------------------------+------------+-----------------+
-| os-odl_l2-sfc-noha | SFC | |
-+-------------------------+------------+-----------------+
-| os-odl_l2-bgpvpn-noha | SDNVPN | |
-+-------------------------+------------+-----------------+
-| os-odl_l2-fdio-noha | FDS | |
-+-------------------------+------------+-----------------+
-| os-onos-nofeature-ha | ONOSFW | |
-+-------------------------+------------+-----------------+
-| os-onos-sfc-ha | ONOSFW | |
-+-------------------------+------------+-----------------+
diff --git a/docs/installation-instructions/baremetal.rst b/docs/installation-instructions/baremetal.rst
deleted file mode 100644
index 1b22a74c..00000000
--- a/docs/installation-instructions/baremetal.rst
+++ /dev/null
@@ -1,242 +0,0 @@
-Installation High-Level Overview - Bare Metal Deployment
-========================================================
-
-The setup presumes that you have 6 or more bare metal servers already setup
-with network connectivity on at least 1 or more network interfaces for all
-servers via a TOR switch or other network implementation.
-
-The physical TOR switches are **not** automatically configured from the OPNFV
-reference platform. All the networks involved in the OPNFV infrastructure as
-well as the provider networks and the private tenant VLANs needs to be manually
-configured.
-
-The Jumphost can be installed using the bootable ISO or by using the
-(``opnfv-apex*.rpm``) RPMs and their dependencies. The Jumphost should then be
-configured with an IP gateway on its admin or public interface and configured
-with a working DNS server. The Jumphost should also have routable access
-to the lights out network for the overcloud nodes.
-
-``opnfv-deploy`` is then executed in order to deploy the undercloud VM and to
-provision the overcloud nodes. ``opnfv-deploy`` uses three configuration files
-in order to know how to install and provision the OPNFV target system.
-The information gathered under section
-`Execution Requirements (Bare Metal Only)`_ is put into the YAML file
-``/etc/opnfv-apex/inventory.yaml`` configuration file. Deployment options are
-put into the YAML file ``/etc/opnfv-apex/deploy_settings.yaml``. Alternatively
-there are pre-baked deploy_settings files available in ``/etc/opnfv-apex/``.
-These files are named with the naming convention
-os-sdn_controller-enabled_feature-[no]ha.yaml. These files can be used in place
-of the ``/etc/opnfv-apex/deploy_settings.yaml`` file if one suites your
-deployment needs. Networking definitions gathered under section
-`Network Requirements`_ are put into the YAML file
-``/etc/opnfv-apex/network_settings.yaml``. ``opnfv-deploy`` will boot the
-undercloud VM and load the target deployment configuration into the
-provisioning toolchain. This information includes MAC address, IPMI,
-Networking Environment and OPNFV deployment options.
-
-Once configuration is loaded and the undercloud is configured it will then
-reboot the overcloud nodes via IPMI. The nodes should already be set to PXE
-boot first off the admin interface. The nodes will first PXE off of the
-undercloud PXE server and go through a discovery/introspection process.
-
-Introspection boots off of custom introspection PXE images. These images are
-designed to look at the properties of the hardware that is being booted
-and report the properties of it back to the undercloud node.
-
-After introspection the undercloud will execute a Heat Stack Deployment to
-continue node provisioning and configuration. The nodes will reboot and PXE
-from the undercloud PXE server again to provision each node using Glance disk
-images provided by the undercloud. These disk images include all the necessary
-packages and configuration for an OPNFV deployment to execute. Once the disk
-images have been written to node's disks the nodes will boot locally and
-execute cloud-init which will execute the final node configuration. This
-configuration is largly completed by executing a puppet apply on each node.
-
-Installation High-Level Overview - VM Deployment
-================================================
-
-The VM nodes deployment operates almost the same way as the bare metal
-deployment with a few differences mainly related to power management.
-``opnfv-deploy`` still deploys an undercloud VM. In addition to the undercloud
-VM a collection of VMs (3 control nodes + 2 compute for an HA deployment or 1
-control node and 1 or more compute nodes for a Non-HA Deployment) will be
-defined for the target OPNFV deployment. The part of the toolchain that
-executes IPMI power instructions calls into libvirt instead of the IPMI
-interfaces on baremetal servers to operate the power managment. These VMs are
-then provisioned with the same disk images and configuration that baremetal
-would be.
-
-To Triple-O these nodes look like they have just built and registered the same
-way as bare metal nodes, the main difference is the use of a libvirt driver for
-the power management.
-
-Installation Guide - Bare Metal Deployment
-==========================================
-
-This section goes step-by-step on how to correctly install and provision the
-OPNFV target system to bare metal nodes.
-
-Install Bare Metal Jumphost
----------------------------
-
-1a. If your Jumphost does not have CentOS 7 already on it, or you would like to
- do a fresh install, then download the Apex bootable ISO from the OPNFV
- artifacts site <http://artifacts.opnfv.org/apex.html>. There have been
- isolated reports of problems with the ISO having trouble completing
- installation successfully. In the unexpected event the ISO does not work
- please workaround this by downloading the CentOS 7 DVD and performing a
- "Virtualization Host" install. If you perform a "Minimal Install" or
- install type other than "Virtualization Host" simply run
- ``sudo yum groupinstall "Virtualization Host"``
- ``chkconfig libvirtd on && reboot``
- to install virtualzation support and enable libvirt on boot. If you use the
- CentOS 7 DVD proceed to step 1b once the CentOS 7 with "Virtualzation Host"
- support is completed.
-
-1b. If your Jump host already has CentOS 7 with libvirt running on it then
- install the install the RDO Release RPM:
-
- ``sudo yum install -y https://www.rdoproject.org/repos/rdo-release.rpm``
-
- The RDO Project release repository is needed to install OpenVSwitch, which
- is a dependency of opnfv-apex. If you do not have external connectivity to
- use this repository you need to download the OpenVSwitch RPM from the RDO
- Project repositories and install it with the opnfv-apex RPM.
-
-2a. Boot the ISO off of a USB or other installation media and walk through
- installing OPNFV CentOS 7. The ISO comes prepared to be written directly
- to a USB drive with dd as such:
-
- ``dd if=opnfv-apex.iso of=/dev/sdX bs=4M``
-
- Replace /dev/sdX with the device assigned to your usb drive. Then select
- the USB device as the boot media on your Jumphost
-
-2b. If your Jump host already has CentOS 7 with libvirt running on it then
- install the opnfv-apex RPMs from the OPNFV artifacts site
- <http://artifacts.opnfv.org/apex.html>. The following RPMS are available
- for installation:
-
- - opnfv-apex - OpenDaylight L2 / L3 and ONOS support *
- - opnfv-apex-onos - ONOS support *
- - opnfv-apex-opendaylight-sfc - OpenDaylight SFC support *
- - opnfv-apex-undercloud - (reqed) Undercloud Image
- - opnfv-apex-common - (reqed) Supporting config files and scripts
- - python34-markupsafe - (reqed) Dependency of opnfv-apex-common **
- - python3-jinja2 - (reqed) Dependency of opnfv-apex-common **
-
- \* One or more of these RPMs is required
- Only one of opnfv-apex, opnfv-apex-onos and opnfv-apex-opendaylight-sfc is
- required. It is safe to leave the unneeded SDN controller's RPMs
- uninstalled if you do not intend to use them.
-
- ** These RPMs are not yet distributed by CentOS or EPEL.
- Apex has built these for distribution with Apex while CentOS and EPEL do
- not distribute them. Once they are carried in an upstream channel Apex will
- no longer carry them and they will not need special handling for
- installation.
-
- To install these RPMs download them to the local disk on your CentOS 7
- install and pass the file names directly to yum:
- ``sudo yum install python34-markupsafe-<version>.rpm
- python3-jinja2-<version>.rpm``
- ``sudo yum install opnfv-apex-<version>.rpm
- opnfv-apex-undercloud-<version>.rpm opnfv-apex-common-<version>.rpm``
-
-
-3. After the operating system and the opnfv-apex RPMs are installed, login to
- your Jumphost as root.
-
-4. Configure IP addresses on the interfaces that you have selected as your
- networks.
-
-5. Configure the IP gateway to the Internet either, preferably on the public
- interface.
-
-6. Configure your ``/etc/resolv.conf`` to point to a DNS server
- (8.8.8.8 is provided by Google).
-
-Creating a Node Inventory File
-------------------------------
-
-IPMI configuration information gathered in section
-`Execution Requirements (Bare Metal Only)`_ needs to be added to the
-``inventory.yaml`` file.
-
-1. Copy ``/usr/share/doc/opnfv/inventory.yaml.example`` as your inventory file
- template to ``/etc/opnfv-apex/inventory.yaml``.
-
-2. The nodes dictionary contains a definition block for each baremetal host
- that will be deployed. 1 or more compute nodes and 3 controller nodes are
- required. (The example file contains blocks for each of these already).
- It is optional at this point to add more compute nodes into the node list.
-
-3. Edit the following values for each node:
-
- - ``mac_address``: MAC of the interface that will PXE boot from undercloud
- - ``ipmi_ip``: IPMI IP Address
- - ``ipmi_user``: IPMI username
- - ``ipmi_password``: IPMI password
- - ``pm_type``: Power Management driver to use for the node
- - ``cpus``: (Introspected*) CPU cores available
- - ``memory``: (Introspected*) Memory available in Mib
- - ``disk``: (Introspected*) Disk space available in Gb
- - ``arch``: (Introspected*) System architecture
- - ``capabilities``: (Opt**) Node role (profile:control or profile:compute)
-
-\* *Introspection looks up the overcloud node's resources and overrides these
- value. You can leave default values and Apex will get the correct values when
- it runs introspection on the nodes.*
-
-** *If capabilities profile is not specified then Apex will select node's roles
- in the OPNFV cluster in a non-deterministic fashion.*
-
-Creating the Settings Files
----------------------------
-
-Edit the 2 settings files in /etc/opnfv-apex/. These files have comments to
-help you customize them.
-
-1. deploy_settings.yaml
- This file includes basic configuration options deployment.
- Alternatively, there are pre-built deploy_settings files available in
- (``/etc/opnfv-apex/``). These files are named with the naming convention
- os-sdn_controller-enabled_feature-[no]ha.yaml. These files can be used in
- place of the (``/etc/opnfv-apex/deploy_settings.yaml``) file if one suites
- your deployment needs. If a pre-built deploy_settings file is choosen there
- is no need to customize (``/etc/opnfv-apex/deploy_settings.yaml``). The
- pre-built file can be used in place of the
- (``/etc/opnfv-apex/deploy_settings.yaml``) file.
-
-2. network_settings.yaml
- This file provides Apex with the networking information that satisfies the
- prerequisite `Network Requirements`_. These are specific to your
- environment.
-
-Running ``opnfv-deploy``
-------------------------
-
-You are now ready to deploy OPNFV using Apex!
-``opnfv-deploy`` will use the inventory and settings files to deploy OPNFV.
-
-Follow the steps below to execute:
-
-1. Execute opnfv-deploy
- ``sudo opnfv-deploy [ --flat ] -n network_settings.yaml
- -i inventory.yaml -d deploy_settings.yaml``
- If you need more information about the options that can be passed to
- opnfv-deploy use ``opnfv-deploy --help`` --flat collapses all networks to a
- single nic, only uses the admin network from the network settings file. -n
- network_settings.yaml allows you to customize your networking topology.
-
-2. Wait while deployment is executed.
- If something goes wrong during this part of the process, start by reviewing
- your network or the information in your configuration files. It's not
- uncommon for something small to be overlooked or mis-typed.
- You will also notice outputs in your shell as the deployment progresses.
-
-3. When the deployment is complete the undercloud IP and ovecloud dashboard
- url will be printed. OPNFV has now been deployed using Apex.
-
-.. _`Execution Requirements (Bare Metal Only)`: requirements.html#execution-requirements-bare-metal-only
-.. _`Network Requirements`: requirements.html#network-requirements
diff --git a/docs/installation-instructions/index.rst b/docs/installation-instructions/index.rst
deleted file mode 100644
index 0dd3d87b..00000000
--- a/docs/installation-instructions/index.rst
+++ /dev/null
@@ -1,27 +0,0 @@
-**************************************
-OPNFV Installation instructions (Apex)
-**************************************
-
-Contents:
-
-.. toctree::
- :numbered:
- :maxdepth: 4
-
- abstract.rst
- introduction.rst
- architecture.rst
- requirements.rst
- baremetal.rst
- virtualinstall.rst
- verification.rst
- references.rst
-
-:Authors: Tim Rozet (trozet@redhat.com)
-:Authors: Dan Radez (dradez@redhat.com)
-:Version: 3.0
-
-Indices and tables
-==================
-
-* :ref:`search`
diff --git a/docs/installation-instructions/introduction.rst b/docs/installation-instructions/introduction.rst
deleted file mode 100644
index cc489917..00000000
--- a/docs/installation-instructions/introduction.rst
+++ /dev/null
@@ -1,42 +0,0 @@
-Introduction
-============
-
-This document describes the steps to install an OPNFV Colorado reference
-platform, as defined by the Genesis Project using the Apex installer.
-
-The audience is assumed to have a good background in networking
-and Linux administration.
-
-Preface
-=======
-
-Apex uses Triple-O from the RDO Project OpenStack distribution as a
-provisioning tool. The Triple-O image based life cycle installation
-tool provisions an OPNFV Target System (3 controllers, 2 or more
-compute nodes) with OPNFV specific configuration provided by the Apex
-deployment tool chain.
-
-The Apex deployment artifacts contain the necessary tools to deploy and
-configure an OPNFV target system using the Apex deployment toolchain.
-These artifacts offer the choice of using the Apex bootable ISO
-(``opnfv-apex-colorado.iso``) to both install CentOS 7 and the
-necessary materials to deploy or the Apex RPMs (``opnfv-apex*.rpm``),
-and their associated dependencies, which expects installation to a
-CentOS 7 libvirt enabled host. The RPM contains a collection of
-configuration files, prebuilt disk images, and the automatic deployment
-script (``opnfv-deploy``).
-
-An OPNFV install requires a "Jumphost" in order to operate. The bootable
-ISO will allow you to install a customized CentOS 7 release to the Jumphost,
-which includes the required packages needed to run ``opnfv-deploy``.
-If you already have a Jumphost with CentOS 7 installed, you may choose to
-skip the ISO step and simply install the (``opnfv-apex*.rpm``) RPMs. The RPMs
-are the same RPMs included in the ISO and include all the necessary disk
-images and configuration files to execute an OPNFV deployment. Either method
-will prepare a host to the same ready state for OPNFV deployment.
-
-``opnfv-deploy`` instantiates a Triple-O Undercloud VM server using libvirt
-as its provider. This VM is then configured and used to provision the
-OPNFV target deployment (3 controllers, n compute nodes). These nodes can
-be either virtual or bare metal. This guide contains instructions for
-installing either method.
diff --git a/docs/installation-instructions/references.rst b/docs/installation-instructions/references.rst
deleted file mode 100644
index 5ff2a542..00000000
--- a/docs/installation-instructions/references.rst
+++ /dev/null
@@ -1,38 +0,0 @@
-Frequently Asked Questions
-==========================
-
-License
-=======
-
-All Apex and "common" entities are protected by the `Apache 2.0 License <http://www.apache.org/licenses/>`_.
-
-References
-==========
-
-OPNFV
------
-
-`OPNFV Home Page <www.opnfv.org>`_
-
-`OPNFV Genesis project page <https://wiki.opnfv.org/get_started>`_
-
-`OPNFV Apex project page <https://wiki.opnfv.org/apex>`_
-
-OpenStack
----------
-
-`OpenStack Mitaka Release artifacts <http://www.openstack.org/software/mitaka>`_
-
-`OpenStack documentation <http://docs.openstack.org>`_
-
-OpenDaylight
-------------
-
-Upstream OpenDaylight provides `a number of packaging and deployment options <https://wiki.opendaylight.org/view/Deployment>`_ meant for consumption by downstream projects like OPNFV.
-
-Currently, OPNFV Apex uses `OpenDaylight's Puppet module <https://github.com/dfarrell07/puppet-opendaylight>`_, which in turn depends on `OpenDaylight's RPM <http://cbs.centos.org/repos/nfv7-opendaylight-4-release/>`_.
-
-RDO Project
------------
-
-`RDO Project website <https://www.rdoproject.org/>`_
diff --git a/docs/installation-instructions/requirements.rst b/docs/installation-instructions/requirements.rst
deleted file mode 100644
index d54d584b..00000000
--- a/docs/installation-instructions/requirements.rst
+++ /dev/null
@@ -1,79 +0,0 @@
-Setup Requirements
-==================
-
-Jumphost Requirements
----------------------
-
-The Jumphost requirements are outlined below:
-
-1. CentOS 7 (from ISO or self-installed).
-
-2. Root access.
-
-3. libvirt virtualization support.
-
-4. minimum 1 networks and maximum 5 networks, multiple NIC and/or VLAN
- combinations are supported. This is virtualized for a VM deployment.
-
-5. The Colorado Apex RPMs and their dependencies.
-
-6. 16 GB of RAM for a bare metal deployment, 64 GB of RAM for a VM
- deployment.
-
-Network Requirements
---------------------
-
-Network requirements include:
-
-1. No DHCP or TFTP server running on networks used by OPNFV.
-
-2. 1-5 separate networks with connectivity between Jumphost and nodes.
-
- - Control Plane (Provisioning)
-
- - Private Tenant-Networking Network*
-
- - External Network
-
- - Storage Network*
-
- - Internal API Network* (required for IPv6 \*\*)
-
-3. Lights out OOB network access from Jumphost with IPMI node enabled
- (bare metal deployment only).
-
-4. External network is a routable network from outside the cloud,
- deployment. The External network is where public internet access would
- reside if available.
-
-\* *These networks can be combined with each other or all combined on the
- Control Plane network.*
-\* *Non-External networks will be consolidated to the Control Plane network
- if not specifically configured.*
-\*\* *Internal API network, by default, is collapsed with provisioning in IPv4
- deployments, this is not possible with the current lack of PXE boot
- support and therefore the API network is required to be its own
- network in an IPv6 deployment.*
-
-Bare Metal Node Requirements
-----------------------------
-
-Bare metal nodes require:
-
-1. IPMI enabled on OOB interface for power control.
-
-2. BIOS boot priority should be PXE first then local hard disk.
-
-3. BIOS PXE interface should include Control Plane network mentioned above.
-
-Execution Requirements (Bare Metal Only)
-----------------------------------------
-
-In order to execute a deployment, one must gather the following information:
-
-1. IPMI IP addresses for the nodes.
-
-2. IPMI login information for the nodes (user/pass).
-
-3. MAC address of Control Plane / Provisioning interfaces of the overcloud
- nodes.
diff --git a/docs/installation-instructions/verification.rst b/docs/installation-instructions/verification.rst
deleted file mode 100644
index 81e4c8e4..00000000
--- a/docs/installation-instructions/verification.rst
+++ /dev/null
@@ -1,89 +0,0 @@
-Verifying the Setup
--------------------
-
-Once the deployment has finished, the OPNFV deployment can be accessed via the
-undercloud node. From the jump host ssh to the undercloud host and become the
-stack user. Alternativly ssh keys have been setup such that the root user on
-the jump host can ssh to undercloud directly as the stack user. For
-convenience a utility script has been provided to look up the undercloud's ip
-address and ssh to the undercloud all in one command. An optional user name can
-be passed to indicate whether to connect as the stack or root user. The stack
-user is default if a username is not specified.
-
-| ``opnfv-util undercloud root``
-| ``su - stack``
-
-Once connected to undercloud as the stack user look for two keystone files that
-can be used to interact with the undercloud and the overcloud. Source the
-appropriate RC file to interact with the respective OpenStack deployment.
-
-| ``source stackrc`` (undercloud)
-| ``source overcloudrc`` (overcloud / OPNFV)
-
-The contents of these files include the credentials for the administrative user
-for undercloud and OPNFV respectivly. At this point both undercloud and OPNFV
-can be interacted with just as any OpenStack installation can be. Start by
-listing the nodes in the undercloud that were used to deploy the overcloud.
-
-| ``source stackrc``
-| ``openstack server list``
-
-The control and compute nodes will be listed in the output of this server list
-command. The IP addresses that are listed are the control plane addresses that
-were used to provision the nodes. Use these IP addresses to connect to these
-nodes. Initial authentication requires using the user heat-admin.
-
-| ``ssh heat-admin@192.0.2.7``
-
-To begin creating users, images, networks, servers, etc in OPNFV source the
-overcloudrc file or retrieve the admin user's credentials from the overcloudrc
-file and connect to the web Dashboard.
-
-
-You are now able to follow the `OpenStack Verification`_ section.
-
-OpenStack Verification
-----------------------
-
-Once connected to the OPNFV Dashboard make sure the OPNFV target system is
-working correctly:
-
-1. In the left pane, click Compute -> Images, click Create Image.
-
-2. Insert a name "cirros", Insert an Image Location
- ``http://download.cirros-cloud.net/0.3.4/cirros-0.3.4-x86_64-disk.img``.
-
-3. Select format "QCOW2", select Public, then click Create Image.
-
-4. Now click Project -> Network -> Networks, click Create Network.
-
-5. Enter a name "internal", click Next.
-
-6. Enter a subnet name "internal_subnet", and enter Network Address
- ``172.16.1.0/24``, click Next.
-
-7. Now go to Project -> Compute -> Instances, click Launch Instance.
-
-8. Enter Instance Name "first_instance", select Instance Boot Source
- "Boot from image", and then select Image Name "cirros".
-
-9. Click Launch, status will cycle though a couple states before becoming
- "Active".
-
-10. Steps 7 though 9 can be repeated to launch more instances.
-
-11. Once an instance becomes "Active" their IP addresses will display on the
- Instances page.
-
-12. Click the name of an instance, then the "Console" tab and login as
- "cirros"/"cubswin:)"
-
-13. To verify storage is working,
- click Project -> Compute -> Volumes, Create Volume
-
-14. Give the volume a name and a size of 1 GB
-
-15. Once the volume becomes "Available" click the dropdown arrow and attach it
- to an instance.
-
-Congratulations you have successfully installed OPNFV!
diff --git a/docs/installation-instructions/virtualinstall.rst b/docs/installation-instructions/virtualinstall.rst
deleted file mode 100644
index 01971893..00000000
--- a/docs/installation-instructions/virtualinstall.rst
+++ /dev/null
@@ -1,69 +0,0 @@
-Installation High-Level Overview - Virtual Deployment
-=====================================================
-
-The VM nodes deployment operates almost the same way as the bare metal
-deployment with a few differences. ``opnfv-deploy`` still deploys an
-undercloud VM. In addition to the undercloud VM a collection of VMs
-(3 control nodes + 2 compute for an HA deployment or 1 control node and 1
-or more compute nodes for a non-HA Deployment) will be defined for the target
-OPNFV deployment. The part of the toolchain that executes IPMI power
-instructions calls into libvirt instead of the IPMI interfaces on baremetal
-servers to operate the power managment. These VMs are then provisioned with
-the same disk images and configuration that baremetal would be. To Triple-O
-these nodes look like they have just built and registered the same way as bare
-metal nodes, the main difference is the use of a libvirt driver for the power
-management. Finally, the default network_settings file will deploy without
-modification. Customizations are welcome but not needed if a generic set of
-network_settings are acceptable.
-
-Installation Guide - Virtual Deployment
-=======================================
-
-This section goes step-by-step on how to correctly install and provision the
-OPNFV target system to VM nodes.
-
-Install Jumphost
-----------------
-
-Follow the instructions in the `Install Bare Metal Jumphost`_ section.
-
-Running ``opnfv-deploy``
-------------------------
-
-You are now ready to deploy OPNFV!
-``opnfv-deploy`` has virtual deployment capability that includes all of
-the configuration nessesary to deploy OPNFV with no modifications.
-
-If no modifications are made to the included configurations the target
-environment will deploy with the following architecture:
-
- - 1 undercloud VM
-
- - The option of 3 control and 2 or more compute VMs (HA Deploy / default)
- or 1 control and 1 or more compute VM (Non-HA deploy / pass -n)
-
- - 1-5 networks: provisioning, private tenant networking, external, storage
- and internal API. The API, storage and tenant networking networks can be
- collapsed onto the provisioning network.
-
-Follow the steps below to execute:
-
-1. ``sudo opnfv-deploy -v [ --virtual-computes n ]
- [ --virtual-cpus n ] [ --virtual-ram n ] [ --flat ]
- -n network_settings.yaml -i inventory.yaml -d deploy_settings.yaml``
-
-2. It will take approximately 45 minutes to an hour to stand up undercloud,
- define the target virtual machines, configure the deployment and execute
- the deployment. You will notice different outputs in your shell.
-
-3. When the deployment is complete the IP for the undercloud and a url for the
- OpenStack dashboard will be displayed
-
-Verifying the Setup - VMs
--------------------------
-
-To verify the set you can follow the instructions in the `Verifying the Setup`_
-section.
-
-.. _`Install Bare Metal Jumphost`: baremetal.html
-.. _`Verifying the Setup`: verification.html