summaryrefslogtreecommitdiffstats
path: root/docs
diff options
context:
space:
mode:
Diffstat (limited to 'docs')
-rw-r--r--docs/configguide/configguide.rst5
-rw-r--r--docs/specification/hardwarespec.rst44
-rw-r--r--docs/specification/index.rst19
-rw-r--r--docs/specification/jumpserverinstall.rst83
-rw-r--r--docs/specification/networkconfig.rst74
-rw-r--r--docs/specification/objectives.rst20
-rw-r--r--docs/specification/remoteaccess.rst58
7 files changed, 303 insertions, 0 deletions
diff --git a/docs/configguide/configguide.rst b/docs/configguide/configguide.rst
new file mode 100644
index 0000000..92d23b7
--- /dev/null
+++ b/docs/configguide/configguide.rst
@@ -0,0 +1,5 @@
+Pharos Lab Configuration requirements
+=====================================
+
+Add an overview for the Pharos configuration guide here. This should be informative regarding the tasks
+and expectations when configuring a lab to be pharos compliant and refer to the relevant pharos docs published from the project.
diff --git a/docs/specification/hardwarespec.rst b/docs/specification/hardwarespec.rst
new file mode 100644
index 0000000..a861b7d
--- /dev/null
+++ b/docs/specification/hardwarespec.rst
@@ -0,0 +1,44 @@
+Pharos compliant environment
+----------------------------
+
+A pharos compliant OPNFV test-bed provides:
+
+- One CentOS 7 jump server on which the virtualized Openstack/OPNFV installer runs
+- In the Brahmaputra release you may select a variety of deployment toolchains to deploy from the jump server.
+- 5 compute / controller nodes (`BGS <https://wiki.opnfv.org/get_started/get_started_work_environment>`_ requires 5 nodes)
+- A configured network topology allowing for LOM, Admin, Public, Private, and Storage Networks
+- Remote access as defined by the Jenkins slave configuration guide
+http://artifacts.opnfv.org/arno.2015.1.0/docs/opnfv-jenkins-slave-connection.arno.2015.1.0.pdf
+
+Hardware requirements
+---------------------
+
+**Servers**
+
+CPU:
+
+* Intel Xeon E5-2600v2 Series
+(Ivy Bridge and newer, or similar)
+
+Local Storage Configuration:
+
+Below describes the minimum for the Pharos spec,
+which is designed to provide enough capacity for
+a reasonably functional environment. Additional
+and/or faster disks are nice to have and may
+produce a better result.
+
+* Disks: 2 x 1TB + 1 x 100GB SSD
+* The first 1TB HDD should be used for OS & additional software/tool installation
+* The second 1TB HDD configured for CEPH object storage
+* Finally, the 100GB SSD should be used as the CEPH journal
+* Performance testing requires a mix of compute nodes that have CEPH(swift+Cinder) and without CEPH storage
+* Virtual ISO boot capabilities or a separate PXE boot server (DHCP/tftp or Cobbler)
+
+Memory:
+
+* 32G RAM Minimum
+
+Power Supply Single
+
+* Single power supply acceptable (redundant power not required/nice to have)
diff --git a/docs/specification/index.rst b/docs/specification/index.rst
new file mode 100644
index 0000000..39764ef
--- /dev/null
+++ b/docs/specification/index.rst
@@ -0,0 +1,19 @@
+Pharos Specification
+====================
+
+.. toctree::
+ :maxdepth: 2
+
+ ./objectives.rst
+ ./hardwarespec.rst
+ ./networkconfig.rst
+ ./jumpserverinstall.rst
+ ./remoteaccess.rst
+
+:Authors: Trevor Cooper (Intel)
+:Version: 1.0
+
+Indices and tables
+==================
+
+* :ref:`search`
diff --git a/docs/specification/jumpserverinstall.rst b/docs/specification/jumpserverinstall.rst
new file mode 100644
index 0000000..a82c335
--- /dev/null
+++ b/docs/specification/jumpserverinstall.rst
@@ -0,0 +1,83 @@
+**Jump Server Configuration:**
+
+(Rough Placeholder, edit me)
+
+**Fuel**
+
+1. Obtain CentOS 7 Minimal ISO and install
+
+ ``wget http://mirrors.kernel.org/centos/7/isos/x86_64/CentOS-7-x86_64-Minimal-1503-01.iso``
+
+2. Set parameters appropriate for your environment during installation
+
+3. Disable NetworkManager
+
+ ``systemctl disable NetworkManager``
+
+4. Configure your /etc/sysconfig/network-scripts/ifcfg-* files for your network
+
+5. Restart networking
+
+ ``service network restart``
+
+6. Edit /etc/resolv.conf and add a nameserver
+
+ ``vi /etc/resolv.conf``
+
+7. Install libvirt & kvm
+
+ ``yum -y update``
+ ``yum -y install kvm qemu-kvm libvirt``
+ ``systemctl enable libvirtd``
+
+8. Reboot:
+
+ ``shutdown -r now``
+
+9. If you wish to avoid annoying delay when use ssh to log in, disable DNS lookups:
+
+ ``vi /etc/ssh/sshd_config``
+
+ Uncomment "UseDNS yes", change 'yes' to 'no'.
+
+ Save
+
+10. Restart sshd
+
+ ``systemctl restart sshd``
+
+11. Install virt-install
+
+ ``yum -y install virt-install``
+
+12. Visit artifacts.opnfv.org and D/L the OPNFV Fuel ISO
+
+13. Create a bridge using the interface on the PXE network, for example: br0
+
+14. Make a directory owned by qemu:
+
+ ``mkdir /home/qemu; mkdir -p /home/qemu/VMs/fuel-6.0/disk``
+
+ ``chown -R qemu:qemu /home/qemu``
+
+15. Copy the ISO to /home/qemu
+
+ ``cd /home/qemu``
+
+ ``virt-install -n opnfv-2015-05-22_18-34-07-fuel -r 4096 --vcpus=4
+ --cpuset=0-3 -c opnfv-2015-05-22_18-34-07.iso --os-type=linux
+ --os-variant=rhel6 --boot hd,cdrom --disk
+ path=/home/qemu/VMs/mirantis-fuel-6.0/disk/fuel-vhd0.qcow2,bus=virtio,size=50,format=qcow2
+ -w bridge=br0,model=virtio --graphics vnc,listen=0.0.0.0``
+
+16. Temporarily flush the firewall rules to make things easier:
+
+ ``iptables -F``
+
+17. Connect to the console of the installing VM with your favorite VNC client.
+
+18. Change the IP settings to match the pod, use an IP in the PXE/Admin network for the Fuel Master
+
+**Foreman**
+
+TBA
diff --git a/docs/specification/networkconfig.rst b/docs/specification/networkconfig.rst
new file mode 100644
index 0000000..510ea80
--- /dev/null
+++ b/docs/specification/networkconfig.rst
@@ -0,0 +1,74 @@
+Networking
+----------
+
+Test-bed network
+
+* 24 or 48 Port TOR Switch
+* NICS - 1GE, 10GE - per server can be on-board or PCI-e
+* Connectivity for each data/control network is through a separate NIC.
+ *This simplifies Switch Management howeverrequires more NICs on the server and also more switch ports
+* Lights-out network can share with Admin/Management
+
+Network Interfaces
+
+* Option I: 4x1G Control, 2x40G Data, 48 Port Switch
+
+ * 1 x 1G for ILMI (Lights out Management )
+ * 1 x 1G for Admin/PXE boot
+ * 1 x 1G for control Plane connectivity
+ * 1 x 1G for storage
+ * 2 x 40G (or 10G) for data network (redundancy, NIC bonding, High bandwidth testing)
+
+* Option II: 1x1G Control, 2x 40G (or 10G) Data, 24 Port Switch
+
+ * Connectivity to networks is through VLANs on the Control NIC.
+ * Data NIC used for VNF traffic and storage traffic segmented through VLANs
+
+* Option III: 2x1G Control, 2x10G Data, 2x40G Storage, 24 Port Switch
+
+ * Data NIC used for VNF traffic
+ * Storage NIC used for control plane and Storage segmented through VLANs (separate host traffic from VNF)
+ * 1 x 1G for IPMI
+ * 1 x 1G for Admin/PXE boot
+ * 2 x 10G for control plane connectivity/Storage
+ * 2 x 40G (or 10G) for data network
+
+Documented configuration to include:
+- Subnet, VLANs (may be constrained by existing lab setups or rules)
+- IPs
+- Types of NW - lights-out, public, private, admin, storage
+- May be special NW requirements for performance related projects
+- Default gateways
+
+ontroller node bridge topology overview
+
+.. image:: ../images/bridge1.png
+
+compute node bridge topology overview
+
+.. image:: ../images/bridge2.png
+
+Architecture
+-------------
+
+** Network Diagram **
+
+The Pharos architecture may be described as follow:
+Figure 1: Standard Deployment Environment
+
+.. image:: ../images/opnfv-pharos-diagram-v01.jpg
+
+Figure 1: Standard Deployment Environment
+
+Sample Network Drawings
+-----------------------
+
+Files for documenting lab network layout.
+These were contributed as Visio VSDX format compressed as a ZIP
+file. Here is a sample of what the visio looks like.
+
+Download the visio zip file here:
+`opnfv-example-lab-diagram.vsdx.zip
+<https://wiki.opnfv.org/_media/opnfv-example-lab-diagram.vsdx.zip>`_
+
+.. image:: ../images/opnfv-example-lab-diagram.png
diff --git a/docs/specification/objectives.rst b/docs/specification/objectives.rst
new file mode 100644
index 0000000..5342388
--- /dev/null
+++ b/docs/specification/objectives.rst
@@ -0,0 +1,20 @@
+Objectives / Scope
+-------------------
+
+The Pharos specification defines the OPNFV hardware environment
+upon which the OPNFV Brahmaputra platform release can be deployed
+on and tested. This specification defines:
+
+- A secure, scalable, standard and HA environment
+- Supports the full Brahmaputra deployment lifecycle (this requires a bare metal environment)
+- Supports functional and performance testing of the Brahmaputra release
+- Provides mechanisms and procedures for secure remote access to the test environment
+
+Deploying Brahmaputra in a Virtualized environment is possible
+and will be useful, however it does not provide a fully
+featured deployment and test environment for the Brahmaputra
+release of OPNFV.
+
+The high level architecture is outlined in the following diagram:
+
+.. image:: ../images/pharos-archi1.jpg
diff --git a/docs/specification/remoteaccess.rst b/docs/specification/remoteaccess.rst
new file mode 100644
index 0000000..e91d55e
--- /dev/null
+++ b/docs/specification/remoteaccess.rst
@@ -0,0 +1,58 @@
+Remote management
+------------------
+
+**Remote access**
+
+- Remote access is required for …
+
+ 1. Developers to access deploy/test environments (credentials to be issued per POD / user)
+ 2. Connection of each environment to Jenkins master hosted by Linux Foundation for automated deployment and test
+
+- OpenVPN is generally used for remote however community hosted labs may vary due to company security rules
+- POD access rules / restrictions …
+
+ - Refer to individual test-bed as each company may have different access rules and acceptable usage policies
+
+- Basic requirement is for SSH sessions to be established (initially on jump server)
+- Majority of packages installed on a system (tools or applications) will be pulled from an external repo.
+
+Firewall rules should include
+
+- SSH sessions
+- Jenkins sessions
+
+Lights-out Management:
+
+- Out-of-band management for power on/off/reset and bare-metal provisioning
+- Access to server is through lights-out-management tool and/or a serial console
+- Intel lights-out ⇒ RMM http://www.intel.com/content/www/us/en/server-management/intel-remote-management-module.html
+- HP lights-out ⇒ ILO http://www8.hp.com/us/en/products/servers/ilo/index.html
+- CISCO lights-out ⇒ UCS https://developer.cisco.com/site/ucs-dev-center/index.gsp
+
+Linux Foundation - VPN service for accessing Lights-Out
+Management (LOM) infrastructure for the UCS-M hardware
+
+- People with admin access to LF infrastructure:
+
+1. amaged@cisco.com
+2. cogibbs@cisco.com
+3. daniel.smith@ericsson.com
+4. dradez@redhat.com
+5. fatih.degirmenci@ericsson.com
+6. fbrockne@cisco.com
+7. jonas.bjurel@ericsson.com
+8. jose.lausuch@ericsson.com
+9. joseph.gasparakis@intel.com
+10. morgan.richomme@orange.com
+11. pbandzi@cisco.com
+12. phladky@cisco.com
+13. stefan.k.berg@ericsson.com
+14. szilard.cserey@ericsson.com
+15. trozet@redhat.com
+
+- The people who require VPN access must have a valid
+PGP key bearing a valid signature from one of these
+three people. When issuing OpenVPN credentials, LF
+will be sending TLS certificates and 2-factor
+authentication tokens, encrypted to each recipient's PGP key.
+