From e68b83768784f28acf28ef1606bc744784672bae Mon Sep 17 00:00:00 2001 From: "juraj.linkes" Date: Wed, 23 Aug 2017 10:52:15 +0200 Subject: Update of existing scenarios documentation We're not releasing l2 scenarios and l3 scenarios have been renamed. Change-Id: Ia6862ab937bf52beb534c46e95747d57be335a25 Signed-off-by: juraj.linkes --- .../os-odl-fdio-ha/FDS-L3-noha-sample-setup.png | Bin 0 -> 168854 bytes .../os-odl-fdio-ha/FDS-basic-components.jpg | Bin 0 -> 184742 bytes .../os-odl-fdio-ha/FDS-odl_l3-noha-overview.png | Bin 0 -> 103223 bytes .../os-odl-fdio-ha/FDS-simple-callflow.png | Bin 0 -> 295451 bytes docs/scenarios/os-odl-fdio-ha/index.rst | 20 ++ .../os-odl-fdio-ha/scenario.description.rst | 278 +++++++++++++++++++ .../os-odl-fdio-noha/FDS-L3-noha-sample-setup.png | Bin 0 -> 168854 bytes .../os-odl-fdio-noha/FDS-basic-components.jpg | Bin 0 -> 184742 bytes .../os-odl-fdio-noha/FDS-odl_l3-noha-overview.png | Bin 0 -> 103223 bytes .../os-odl-fdio-noha/FDS-simple-callflow.png | Bin 0 -> 295451 bytes docs/scenarios/os-odl-fdio-noha/index.rst | 20 ++ .../os-odl-fdio-noha/scenario.description.rst | 272 +++++++++++++++++++ .../FDS-L3-tenant-connectivity.png | Bin 50947 -> 0 bytes .../os-odl_l2-fdio-ha/FDS-basic-callflow.jpg | Bin 148454 -> 0 bytes .../os-odl_l2-fdio-ha/FDS-basic-components.jpg | Bin 184742 -> 0 bytes .../os-odl_l2-fdio-ha/FDS-odl_l2-ha-overview.png | Bin 92329 -> 0 bytes .../os-odl_l2-fdio-ha/FDS-odl_l2-overview.png | Bin 90635 -> 0 bytes .../os-odl_l2-fdio-ha/FDS-simple-callflow.png | Bin 295451 -> 0 bytes docs/scenarios/os-odl_l2-fdio-ha/index.rst | 20 -- .../os-odl_l2-fdio-ha-colorado2_1.png | Bin 174442 -> 0 bytes .../os-odl_l2-fdio-ha/scenario.description.rst | 299 --------------------- .../FDS-L3-tenant-connectivity.png | Bin 50947 -> 0 bytes .../os-odl_l2-fdio-noha/FDS-basic-callflow.jpg | Bin 148454 -> 0 bytes .../os-odl_l2-fdio-noha/FDS-basic-components.jpg | Bin 184742 -> 0 bytes .../os-odl_l2-fdio-noha/FDS-odl_l2-overview.png | Bin 90635 -> 0 bytes .../os-odl_l2-fdio-noha/FDS-simple-callflow.png | Bin 295451 -> 0 bytes docs/scenarios/os-odl_l2-fdio-noha/index.rst | 20 -- .../os-odl_l2-fdio-noha/scenario.description.rst | 281 ------------------- .../os-odl_l3-fdio-ha/FDS-L3-DVR-example.png | Bin 299709 -> 0 bytes .../os-odl_l3-fdio-ha/FDS-L3-noha-sample-setup.png | Bin 168854 -> 0 bytes .../os-odl_l3-fdio-ha/FDS-basic-callflow.jpg | Bin 148454 -> 0 bytes .../os-odl_l3-fdio-ha/FDS-basic-components.jpg | Bin 184742 -> 0 bytes .../os-odl_l3-fdio-ha/FDS-odl_l3-noha-overview.png | Bin 103223 -> 0 bytes .../os-odl_l3-fdio-ha/FDS-odl_l3-overview.png | Bin 89119 -> 0 bytes .../os-odl_l3-fdio-ha/FDS-simple-callflow.png | Bin 295451 -> 0 bytes docs/scenarios/os-odl_l3-fdio-ha/index.rst | 20 -- .../os-odl_l3-fdio-ha/scenario.description.rst | 280 ------------------- .../os-odl_l3-fdio-noha/FDS-L3-DVR-example.png | Bin 299709 -> 0 bytes .../FDS-L3-noha-sample-setup.png | Bin 168854 -> 0 bytes .../os-odl_l3-fdio-noha/FDS-basic-callflow.jpg | Bin 148454 -> 0 bytes .../os-odl_l3-fdio-noha/FDS-basic-components.jpg | Bin 184742 -> 0 bytes .../FDS-odl_l3-noha-overview.png | Bin 103223 -> 0 bytes .../os-odl_l3-fdio-noha/FDS-odl_l3-overview.png | Bin 89119 -> 0 bytes .../os-odl_l3-fdio-noha/FDS-simple-callflow.png | Bin 295451 -> 0 bytes docs/scenarios/os-odl_l3-fdio-noha/index.rst | 20 -- .../os-odl_l3-fdio-noha/scenario.description.rst | 274 ------------------- 46 files changed, 590 insertions(+), 1214 deletions(-) create mode 100755 docs/scenarios/os-odl-fdio-ha/FDS-L3-noha-sample-setup.png create mode 100755 docs/scenarios/os-odl-fdio-ha/FDS-basic-components.jpg create mode 100755 docs/scenarios/os-odl-fdio-ha/FDS-odl_l3-noha-overview.png create mode 100755 docs/scenarios/os-odl-fdio-ha/FDS-simple-callflow.png create mode 100644 docs/scenarios/os-odl-fdio-ha/index.rst create mode 100755 docs/scenarios/os-odl-fdio-ha/scenario.description.rst create mode 100755 docs/scenarios/os-odl-fdio-noha/FDS-L3-noha-sample-setup.png create mode 100755 docs/scenarios/os-odl-fdio-noha/FDS-basic-components.jpg create mode 100755 docs/scenarios/os-odl-fdio-noha/FDS-odl_l3-noha-overview.png create mode 100755 docs/scenarios/os-odl-fdio-noha/FDS-simple-callflow.png create mode 100644 docs/scenarios/os-odl-fdio-noha/index.rst create mode 100755 docs/scenarios/os-odl-fdio-noha/scenario.description.rst delete mode 100755 docs/scenarios/os-odl_l2-fdio-ha/FDS-L3-tenant-connectivity.png delete mode 100755 docs/scenarios/os-odl_l2-fdio-ha/FDS-basic-callflow.jpg delete mode 100755 docs/scenarios/os-odl_l2-fdio-ha/FDS-basic-components.jpg delete mode 100755 docs/scenarios/os-odl_l2-fdio-ha/FDS-odl_l2-ha-overview.png delete mode 100755 docs/scenarios/os-odl_l2-fdio-ha/FDS-odl_l2-overview.png delete mode 100755 docs/scenarios/os-odl_l2-fdio-ha/FDS-simple-callflow.png delete mode 100644 docs/scenarios/os-odl_l2-fdio-ha/index.rst delete mode 100755 docs/scenarios/os-odl_l2-fdio-ha/os-odl_l2-fdio-ha-colorado2_1.png delete mode 100755 docs/scenarios/os-odl_l2-fdio-ha/scenario.description.rst delete mode 100755 docs/scenarios/os-odl_l2-fdio-noha/FDS-L3-tenant-connectivity.png delete mode 100755 docs/scenarios/os-odl_l2-fdio-noha/FDS-basic-callflow.jpg delete mode 100755 docs/scenarios/os-odl_l2-fdio-noha/FDS-basic-components.jpg delete mode 100755 docs/scenarios/os-odl_l2-fdio-noha/FDS-odl_l2-overview.png delete mode 100755 docs/scenarios/os-odl_l2-fdio-noha/FDS-simple-callflow.png delete mode 100644 docs/scenarios/os-odl_l2-fdio-noha/index.rst delete mode 100755 docs/scenarios/os-odl_l2-fdio-noha/scenario.description.rst delete mode 100755 docs/scenarios/os-odl_l3-fdio-ha/FDS-L3-DVR-example.png delete mode 100755 docs/scenarios/os-odl_l3-fdio-ha/FDS-L3-noha-sample-setup.png delete mode 100755 docs/scenarios/os-odl_l3-fdio-ha/FDS-basic-callflow.jpg delete mode 100755 docs/scenarios/os-odl_l3-fdio-ha/FDS-basic-components.jpg delete mode 100755 docs/scenarios/os-odl_l3-fdio-ha/FDS-odl_l3-noha-overview.png delete mode 100755 docs/scenarios/os-odl_l3-fdio-ha/FDS-odl_l3-overview.png delete mode 100755 docs/scenarios/os-odl_l3-fdio-ha/FDS-simple-callflow.png delete mode 100644 docs/scenarios/os-odl_l3-fdio-ha/index.rst delete mode 100755 docs/scenarios/os-odl_l3-fdio-ha/scenario.description.rst delete mode 100755 docs/scenarios/os-odl_l3-fdio-noha/FDS-L3-DVR-example.png delete mode 100755 docs/scenarios/os-odl_l3-fdio-noha/FDS-L3-noha-sample-setup.png delete mode 100755 docs/scenarios/os-odl_l3-fdio-noha/FDS-basic-callflow.jpg delete mode 100755 docs/scenarios/os-odl_l3-fdio-noha/FDS-basic-components.jpg delete mode 100755 docs/scenarios/os-odl_l3-fdio-noha/FDS-odl_l3-noha-overview.png delete mode 100755 docs/scenarios/os-odl_l3-fdio-noha/FDS-odl_l3-overview.png delete mode 100755 docs/scenarios/os-odl_l3-fdio-noha/FDS-simple-callflow.png delete mode 100644 docs/scenarios/os-odl_l3-fdio-noha/index.rst delete mode 100755 docs/scenarios/os-odl_l3-fdio-noha/scenario.description.rst (limited to 'docs') diff --git a/docs/scenarios/os-odl-fdio-ha/FDS-L3-noha-sample-setup.png b/docs/scenarios/os-odl-fdio-ha/FDS-L3-noha-sample-setup.png new file mode 100755 index 0000000..27c8335 Binary files /dev/null and b/docs/scenarios/os-odl-fdio-ha/FDS-L3-noha-sample-setup.png differ diff --git a/docs/scenarios/os-odl-fdio-ha/FDS-basic-components.jpg b/docs/scenarios/os-odl-fdio-ha/FDS-basic-components.jpg new file mode 100755 index 0000000..e92851f Binary files /dev/null and b/docs/scenarios/os-odl-fdio-ha/FDS-basic-components.jpg differ diff --git a/docs/scenarios/os-odl-fdio-ha/FDS-odl_l3-noha-overview.png b/docs/scenarios/os-odl-fdio-ha/FDS-odl_l3-noha-overview.png new file mode 100755 index 0000000..1193ea4 Binary files /dev/null and b/docs/scenarios/os-odl-fdio-ha/FDS-odl_l3-noha-overview.png differ diff --git a/docs/scenarios/os-odl-fdio-ha/FDS-simple-callflow.png b/docs/scenarios/os-odl-fdio-ha/FDS-simple-callflow.png new file mode 100755 index 0000000..04546aa Binary files /dev/null and b/docs/scenarios/os-odl-fdio-ha/FDS-simple-callflow.png differ diff --git a/docs/scenarios/os-odl-fdio-ha/index.rst b/docs/scenarios/os-odl-fdio-ha/index.rst new file mode 100644 index 0000000..f02c451 --- /dev/null +++ b/docs/scenarios/os-odl-fdio-ha/index.rst @@ -0,0 +1,20 @@ +.. _os-odl-fdio-ha: + +.. OPNFV - Open Platform for Network Function Virtualization +.. This work is licensed under a Creative Commons Attribution 4.0 International License. +.. http://creativecommons.org/licenses/by/4.0 + + +********************************************************************* +Fast Data Stacks Scenario: os-odl-fdio-ha Overview and Description +********************************************************************* + +Scenario: "OpenStack - Opendaylight - FD.io" (apex-os-odl-fdio-ha) +is a scenario developed as part of the FastDataStacks +OPNFV project. + +.. toctree:: + :numbered: + :maxdepth: 2 + + scenario.description.rst diff --git a/docs/scenarios/os-odl-fdio-ha/scenario.description.rst b/docs/scenarios/os-odl-fdio-ha/scenario.description.rst new file mode 100755 index 0000000..9377879 --- /dev/null +++ b/docs/scenarios/os-odl-fdio-ha/scenario.description.rst @@ -0,0 +1,278 @@ +.. OPNFV - Open Platform for Network Function Virtualization +.. This work is licensed under a Creative Commons Attribution 4.0 International License. +.. http://creativecommons.org/licenses/by/4.0 + +Scenario: "OpenStack - OpenDaylight - FD.io" +====================================================== + +Scenario: apex-os-odl-fdio-ha + +"apex-os-odl-fdio-ha" is a scenario developed as part of the +FastDataStacks OPNFV project. The main components of the +"apex-os-odl-fdio-ha" scenario are: + + - APEX (TripleO) installer (please also see APEX installer documentation) + - Openstack (in HA configuration) + - OpenDaylight controller (clustered) + controlling layer 2 and layer 3 networking + - FD.io/VPP virtual forwarder for tenant networking + +Introduction +============ + +NFV and virtualized high performance applications, such as video processing, +require a "fast data stack" solution that provides both carrier grade +forwarding performance, scalability and open extensibility, along with +functionality for realizing application policies and controlling a complex +network topology. + +A solution stack is only as good as its foundation. Key foundational assets for +NFV infrastructure are + + * The virtual forwarder: The virtual forwarder needs to be a feature rich, + high performance, highly scale virtual switch-router. It needs to leverage + hardware accelerators when available and run in user space. + In addition, it should be modular and easily extensible. + * Forwarder diversity: A solution stack should support a variety of + forwarders, hardware forwarders (physical switches and routers) + as well as software forwarders. This way virtual and physical + forwarding domains can be seamlessly glued together. + * Policy driven connectivity: Connectivity should respect and + reflect different business + +In order to meet the desired qualities of an NFV infrastructure, the +following components were chosen for the "Openstack - OpenDaylight - FD.io" +scenario: + + * FD.io Vector Packet Processor (VPP) - a highly scalable, + high performance, extensible virtual forwarder + * OpenDaylight Controller - an extensible controller platform which + offers the ability to separate business logic from networking + constructs, supports a diverse set of network devices + (virtual and physical) via the "group based policy (GBP)" + component, and can be clustered to achieve a highly available + deployment. + + +The "Openstack - OpenDaylight - FD.io" scenario provides the capability to +realize a set of use-cases relevant to the deployment of NFV nodes instantiated +by means of an Openstack orchestration system on FD.io/VPP enabled compute +nodes. The role of the Opendaylight network controller in this integration is +twofold. It provides a network device configuration and topology abstraction +via the Openstack Neutron interface, while providing the capability to realize +more complex network policies by means of Group Based Policies. Furthermore it +also provides the capabilities to monitor as well as visualize the operation of +the virtual network devices and their topologies. In supporting the general +use-case of instantiatiting an NFV instance, two specific types of network +transport use cases are realized: + + * NFV instances with VPP data-plane forwarding using a VLAN provider network + * NFV instances with VPP data-plane forwarding using a VXLAN overlay + transport network + +A deployment of the "apex-os-odl-fdio-ha" scenario consists of 6 or more +servers: + + * 1 Jumphost hosting the APEX installer - running the Undercloud + * 3 Controlhosts, which run the Overcloud as well as + 3 instances of OpenDaylight as a network controller + * 2 or more Computehosts. One of the Computehosts also serves as + layer 3 gateway for the tenant networks. + +The following picture depicts a non-HA setup of the L3 scenario. + +.. image:: FDS-odl_l3-noha-overview.png + +Tenant and public networking leverages FD.io/VPP. On one of the compute nodes, +VPP binds to both, the tenant networking interface as well as to the public +networking interface. This means that VPP is used for communication within +a tenant network, between tenant networks, as well as between a tenant network +and the Internet. Note that this setup slightly differs from the usual +centralized L3 setup with qrouter on the control node. This setup was chosen +to limit the configuration changes for the introduction of FD.io/VPP. The +OpenDaylight network controller is used to setup and manage layer 2 and +layer 3 networking for the scenario - with Group Based Policy (GBP) being the +key component. Tenant networking can either leverage VXLAN (in which case a +full mesh of VXLAN tunnels is created) or VLANs. + +The picture below shows an example setup with two compute and one control +node. Note that the external network is connected via compute node 0 through +VPP. VPP provides all layer 3 services which are provided in a "vanilla" +OpenStack deployment, including SNAT and DNAT, as well as north-south +and east-west traffic filtering for security purposes ("security groups"). + +.. image:: FDS-L3-noha-sample-setup.png + +Features of the scenario +------------------------ + +Main features of the "apex-os-odl-fdio-ha" scenario: + + * Automated installation using the APEX installer + * Fast and scalable tenant networking using FD.io/VPP as forwarder + * Layer 2 networking using VLANs or VXLAN, managed + and controlled through OpenDaylight + * Layer 3 connectivitiy for tenant networks supplied + through FD.io/VPP. Layer 3 features, including security groups as well as + floating IP addresses (i.e. NAT) are implemented by the FD.io/VPP forwarder + * Manual and automatic (via DHCP) addressing on tenant networks + * OpenStack high-availability and OpenDaylight clustered mode + +Scenario components and composition +=================================== + +The apex-os-odl-fdio-ha scenario combines components from three key open +source projects: OpenStack, OpenDaylight, and Fast Data (FD.io). The key +components that realize the apex-os-odl-fdio-ha scenario and which differ +from a regular, OVS-based scenario, are the OpenStack ML2 OpenDaylight plugin, +OpenDaylight Neutron Northbound, OpenDaylight Group Based Policy, OpenDaylight +Virtual Bridge Domain Manager, FD.io Honeycomb management agent and FD.io +Vector Packet Processor (VPP). + +Here's a more detailed list of the individual software components involved: + +**Openstack Neutron ML2 OpenDaylight Plugin**: Handles Neutron data base +synchronization and interaction with the southbound controller using a REST +interface. + +**ODL GBP Neutron Mapper**: Maps neutron elements like networks, subnets, +security groups, etc. to GBP entities: Creates policy and configuration for +tenants (endpoints, resolved policies, forwarding rules). + +**ODL GBP Neutron VPP Mapper**: Maps Neutron ports to VPP endpoints in GBP. + +**ODL GBP Location Manager**: Provides real location for endpoints (i.e. Which +physical node an endpoint is connected to). + +**GBP Renderer Manager**: Creates configuration for Renderers (like e.g. +VPP-Renderer or OVS-Renderer). The GBP Renderer Manager is the central point +for dispatching of data to specific device renderers. It uses the information +derived from the GBP end-point and its topology entries to dispatch the task +of configuration to a specific device renderer by writing a renderer policy +configuration into the registered renderer's policy store. The renderer +manager also monitors, by being a data change listener on the VPP Renderer +Policy States, for any errors in the application of a rendered configuration. + +**GBP VPP Renderer Interface Manager**: Listens to VPP endpoints in the +Config DataStore and configures associated interfaces on VPP via HoneyComb. + +**GBP VPP Renderer Renderer Policy Manager**: Manages the creation of +bridge domains using VBD and assigns interfaces to bridge domains. + +**Virtual Bridge Domain Manager (VBD)**: Creates bridge domains (i.e. in case +of VXLAN creates full mesh of VXLAN tunnels, configures split horizon on +tunnel endpoints etc.). VDB configures VXLAN tunnels always into a full-mesh +with split-horizon group forwarding applied on any domain facing tunnel +interface (i.e. forwarding behavior will be that used for VPLS). + +**Virtual Packet Processor (VPP) and Honeycomb server**: The VPP is the +accelerated data plane forwarding engine relying on vhost user interfaces +towards Virtual Machines created by the Nova Agent. The Honeycomb NETCONF +configuration server is responsible for driving the configuration of the VPP, +and collecting the operational data. + +**Nova Agent**: The Nova Agent, a sub-component of the overall Openstack +architecture, is responsible for interacting with the compute node's host +Libvirt API to drive the life-cycle of Virtual Machines. It, along with the +compute node software, are assumed to be capable of supporting vhost user +interfaces. + +The picture below shows the key components. + +.. image:: FDS-basic-components.jpg + +To provide a better understanding how the above mentioned components interact +with each other, the following diagram shows how the example of creating a +vhost-user port on VPP through Openstack Neutron: + +To create or update a port, Neutron will send a request to ODL Neutron +Northbound which contains the UUID, along with the host-id as "vpp" and +vif-type as "vhost-user". The GBP Neutron mapper turns the "Neutron speak" of +"ports" into the generic connectivity model that GroupBasedPolicy uses. +Neutron "ports" become generic "GBP Endpoints" which can be consumed by the +GBP Renderer Manager. The GBP Renderer Manager resolves the policy for the +endpoint, i.e. it determines which communication relationships apply to the +specific endpoint, and hands the resolution to a device specific renderer, +which is the VPP renderer in the given case here. VPP renderer turns the +generic policy into VPP specific configuration. Note that in case the policy +would need to be applied to a different device, e.g. an OpenVSwitch (OVS), +then an "OVS Renderer" would be used. VPP Renderer and the topology manager +("Virtual Bridge Domain" manager - i.e. VBD) cooperate to create the actual +network configuration. VPP Renderer configures the interfaces to the virtual +machines (VM), i.e. the vhost-user interface in the given case here and +attaches them to a bridge domain on VPP. VBD handles the setup of connectivity +between bridge domains on individual VPPs, i.e. it maintains the VXLAN tunnels +in the given case here. Both VPP Renderer as well as VBD communicate with the +device through Netconf/YANG. All compute and control nodes run an instance of +VPP and the VPP-configuration agent "Honeycomb". Honeycomb serves as a +Netconf/YANG server, receives the configuration commands from VBD and VPP +Renderer and drives VPP configuration using VPP's local Java APIs. + +.. image:: FDS-simple-callflow.png + +Scenario Configuration +====================== + +To enable the "apex-os-odl-fdio-ha" scenario check the appropriate +settings in the APEX configuration files. Those are typically found in +/etc/opnfv-apex. + +File "deploy_settings.yaml": Choose Opendaylight as controller with version +"carbon" and enable vpp as forwarder. "odl_routing_node" chooses the node +which is used as the layer 3 gateway for the tenant networks. +"odl_routing_node" is an optional parameter. If omitted, the VPP on the first +compute node will serve as layer 3 gateway:: + + global_params: + ha_enabled: true + + deploy_options: + sdn_controller: opendaylight + sdn_l3: true + odl_version: carbon + odl_routing_node: overcloud-novacompute-0 + tacker: true + congress: true + sfc: false + vpn: false + vpp: true + dataplane: fdio + performance: + Controller: + kernel: + hugepages: 1024 + hugepagesz: 2M + intel_iommu: 'on' + iommu: pt + isolcpus: 1,2 + vpp: + main-core: 1 + corelist-workers: 2 + uio-driver: uio_pci_generic + Compute: + kernel: + hugepagesz: 2M + hugepages: 2048 + intel_iommu: 'on' + iommu: pt + isolcpus: 1,2 + vpp: + main-core: 1 + corelist-workers: 2 + uio-driver: uio_pci_generic + +Limitations, Issues and Workarounds +=================================== + +For specific information on limitations and issues, please refer to the APEX +installer release notes. + +References +========== + + + * FastDataStacks OPNFV project wiki: https://wiki.opnfv.org/display/fds + * Fast Data (FD.io): https://fd.io/ + * FD.io Vector Packet Processor (VPP): https://wiki.fd.io/view/VPP + * OpenDaylight Controller: https://www.opendaylight.org/ + * OPNFV Danube release - more information: http://www.opnfv.org/danube diff --git a/docs/scenarios/os-odl-fdio-noha/FDS-L3-noha-sample-setup.png b/docs/scenarios/os-odl-fdio-noha/FDS-L3-noha-sample-setup.png new file mode 100755 index 0000000..27c8335 Binary files /dev/null and b/docs/scenarios/os-odl-fdio-noha/FDS-L3-noha-sample-setup.png differ diff --git a/docs/scenarios/os-odl-fdio-noha/FDS-basic-components.jpg b/docs/scenarios/os-odl-fdio-noha/FDS-basic-components.jpg new file mode 100755 index 0000000..e92851f Binary files /dev/null and b/docs/scenarios/os-odl-fdio-noha/FDS-basic-components.jpg differ diff --git a/docs/scenarios/os-odl-fdio-noha/FDS-odl_l3-noha-overview.png b/docs/scenarios/os-odl-fdio-noha/FDS-odl_l3-noha-overview.png new file mode 100755 index 0000000..1193ea4 Binary files /dev/null and b/docs/scenarios/os-odl-fdio-noha/FDS-odl_l3-noha-overview.png differ diff --git a/docs/scenarios/os-odl-fdio-noha/FDS-simple-callflow.png b/docs/scenarios/os-odl-fdio-noha/FDS-simple-callflow.png new file mode 100755 index 0000000..04546aa Binary files /dev/null and b/docs/scenarios/os-odl-fdio-noha/FDS-simple-callflow.png differ diff --git a/docs/scenarios/os-odl-fdio-noha/index.rst b/docs/scenarios/os-odl-fdio-noha/index.rst new file mode 100644 index 0000000..283080e --- /dev/null +++ b/docs/scenarios/os-odl-fdio-noha/index.rst @@ -0,0 +1,20 @@ +.. _os-odl-fdio-noha: + +.. OPNFV - Open Platform for Network Function Virtualization +.. This work is licensed under a Creative Commons Attribution 4.0 International License. +.. http://creativecommons.org/licenses/by/4.0 + + +*********************************************************************** +Fast Data Stacks Scenario: os-odl-fdio-noha Overview and Description +*********************************************************************** + +Scenario: "OpenStack - Opendaylight - FD.io" (apex-os-odl-fdio-noha) +is a scenario developed as part of the FastDataStacks +OPNFV project. + +.. toctree:: + :numbered: + :maxdepth: 2 + + scenario.description.rst diff --git a/docs/scenarios/os-odl-fdio-noha/scenario.description.rst b/docs/scenarios/os-odl-fdio-noha/scenario.description.rst new file mode 100755 index 0000000..c9927ff --- /dev/null +++ b/docs/scenarios/os-odl-fdio-noha/scenario.description.rst @@ -0,0 +1,272 @@ +.. OPNFV - Open Platform for Network Function Virtualization +.. This work is licensed under a Creative Commons Attribution 4.0 International License. +.. http://creativecommons.org/licenses/by/4.0 + +Scenario: "OpenStack - OpenDaylight - FD.io" +====================================================== + +Scenario: apex-os-odl-fdio-noha + +"apex-os-odl-fdio-noha" is a scenario developed as part of the +FastDataStacks OPNFV project. The main components of the +"apex-os-odl-fdio-noha" scenario are: + + - APEX (TripleO) installer (please also see APEX installer documentation) + - Openstack (in non-HA configuration) + - OpenDaylight controller (non-clustered) + controlling layer 2 and layer 3 networking + - FD.io/VPP virtual forwarder for tenant networking + +Introduction +============ + +NFV and virtualized high performance applications, such as video processing, +require a "fast data stack" solution that provides both carrier grade +forwarding performance, scalability and open extensibility, along with +functionality for realizing application policies and controlling a complex +network topology. + +A solution stack is only as good as its foundation. Key foundational assets for +NFV infrastructure are + + * The virtual forwarder: The virtual forwarder needs to be a feature rich, + high performance, highly scale virtual switch-router. It needs to leverage + hardware accelerators when available and run in user space. + In addition, it should be modular and easily extensible. + * Forwarder diversity: A solution stack should support a variety of + forwarders, hardware forwarders (physical switches and routers) + as well as software forwarders. This way virtual and physical + forwarding domains can be seamlessly glued together. + * Policy driven connectivity: Connectivity should respect and + reflect different business + +In order to meet the desired qualities of an NFV infrastructure, the +following components were chosen for the "Openstack - OpenDaylight - FD.io" +scenario: + + * FD.io Vector Packet Processor (VPP) - a highly scalable, + high performance, extensible virtual forwarder + * OpenDaylight Controller - an extensible controller platform which + offers the ability to separate business logic from networking + constructs, supports a diverse set of network devices + (virtual and physical) via the "group based policy (GBP)" + component, and can be clustered to achieve a highly available + deployment. + + +The "Openstack - OpenDaylight - FD.io" scenario provides the capability to +realize a set of use-cases relevant to the deployment of NFV nodes instantiated +by means of an Openstack orchestration system on FD.io/VPP enabled compute +nodes. The role of the Opendaylight network controller in this integration is +twofold. It provides a network device configuration and topology abstraction +via the Openstack Neutron interface, while providing the capability to realize +more complex network policies by means of Group Based Policies. Furthermore it +also provides the capabilities to monitor as well as visualize the operation of +the virtual network devices and their topologies. In supporting the general +use-case of instantiatiting an NFV instance, two specific types of network +transport use cases are realized: + + * NFV instances with VPP data-plane forwarding using a VLAN provider network + * NFV instances with VPP data-plane forwarding using a VXLAN overlay + transport network + +A deployment of the "apex-os-odl-fdio-noha" scenario consists of 4 or more +servers: + + * 1 Jumphost hosting the APEX installer - running the Undercloud + * 1 Controlhost, which runs the Overcloud as well as + OpenDaylight as a network controller + * 2 or more Computehosts. One of the Computehosts also serves as + layer 3 gateway for the tenant networks. + +.. image:: FDS-odl_l3-noha-overview.png + +Tenant and public networking leverages FD.io/VPP. On one of the compute nodes, +VPP binds to both, the tenant networking interface as well as to the public +networking interface. This means that VPP is used for communication within +a tenant network, between tenant networks, as well as between a tenant network +and the Internet. Note that this setup slightly differs from the usual +centralized L3 setup with qrouter on the control node. This setup was chosen +to limit the configuration changes for the introduction of FD.io/VPP. The +OpenDaylight network controller is used to setup and manage layer 2 and +layer 3 networking for the scenario - with Group Based Policy (GBP) being the +key component. Tenant networking can either leverage VXLAN (in which case a +full mesh of VXLAN tunnels is created) or VLANs. + +The picture below shows an example setup with two compute and one control +node. Note that the external network is connected via compute node 0 through +VPP. VPP provides all layer 3 services which are provided in a "vanilla" +OpenStack deployment, including SNAT and DNAT, as well as north-south +and east-west traffic filtering for security purposes ("security groups"). + +.. image:: FDS-L3-noha-sample-setup.png + +Features of the scenario +------------------------ + +Main features of the "apex-os-odl-fdio-noha" scenario: + + * Automated installation using the APEX installer + * Fast and scalable tenant networking using FD.io/VPP as forwarder + * Layer 2 networking using VLANs or VXLAN, managed + and controlled through OpenDaylight + * Layer 3 connectivitiy for tenant networks supplied + through FD.io/VPP. Layer 3 features, including security groups as well as + floating IP addresses (i.e. NAT) are implemented by the FD.io/VPP forwarder + * Manual and automatic (via DHCP) addressing on tenant networks + +Scenario components and composition +=================================== + +The apex-os-odl-fdio-noha scenario combines components from three key open +source projects: OpenStack, OpenDaylight, and Fast Data (FD.io). The key +components that realize the apex-os-odl-fdio-noha scenario and which differ +from a regular, OVS-based scenario, are the OpenStack ML2 OpenDaylight plugin, +OpenDaylight Neutron Northbound, OpenDaylight Group Based Policy, OpenDaylight +Virtual Bridge Domain Manager, FD.io Honeycomb management agent and FD.io +Vector Packet Processor (VPP). + +Here's a more detailed list of the individual software components involved: + +**Openstack Neutron ML2 OpenDaylight Plugin**: Handles Neutron data base +synchronization and interaction with the southbound controller using a REST +interface. + +**ODL GBP Neutron Mapper**: Maps neutron elements like networks, subnets, +security groups, etc. to GBP entities: Creates policy and configuration for +tenants (endpoints, resolved policies, forwarding rules). + +**ODL GBP Neutron VPP Mapper**: Maps Neutron ports to VPP endpoints in GBP. + +**ODL GBP Location Manager**: Provides real location for endpoints (i.e. Which +physical node an endpoint is connected to). + +**GBP Renderer Manager**: Creates configuration for Renderers (like e.g. +VPP-Renderer or OVS-Renderer). The GBP Renderer Manager is the central point +for dispatching of data to specific device renderers. It uses the information +derived from the GBP end-point and its topology entries to dispatch the task +of configuration to a specific device renderer by writing a renderer policy +configuration into the registered renderer's policy store. The renderer +manager also monitors, by being a data change listener on the VPP Renderer +Policy States, for any errors in the application of a rendered configuration. + +**GBP VPP Renderer Interface Manager**: Listens to VPP endpoints in the +Config DataStore and configures associated interfaces on VPP via HoneyComb. + +**GBP VPP Renderer Renderer Policy Manager**: Manages the creation of +bridge domains using VBD and assigns interfaces to bridge domains. + +**Virtual Bridge Domain Manager (VBD)**: Creates bridge domains (i.e. in case +of VXLAN creates full mesh of VXLAN tunnels, configures split horizon on +tunnel endpoints etc.). VDB configures VXLAN tunnels always into a full-mesh +with split-horizon group forwarding applied on any domain facing tunnel +interface (i.e. forwarding behavior will be that used for VPLS). + +**Virtual Packet Processor (VPP) and Honeycomb server**: The VPP is the +accelerated data plane forwarding engine relying on vhost user interfaces +towards Virtual Machines created by the Nova Agent. The Honeycomb NETCONF +configuration server is responsible for driving the configuration of the VPP, +and collecting the operational data. + +**Nova Agent**: The Nova Agent, a sub-component of the overall Openstack +architecture, is responsible for interacting with the compute node's host +Libvirt API to drive the life-cycle of Virtual Machines. It, along with the +compute node software, are assumed to be capable of supporting vhost user +interfaces. + +The picture below shows the key components. + +.. image:: FDS-basic-components.jpg + +To provide a better understanding how the above mentioned components interact +with each other, the following diagram shows how the example of creating a +vhost-user port on VPP through Openstack Neutron: + +To create or update a port, Neutron will send a request to ODL Neutron +Northbound which contains the UUID, along with the host-id as "vpp" and +vif-type as "vhost-user". The GBP Neutron mapper turns the "Neutron speak" of +"ports" into the generic connectivity model that GroupBasedPolicy uses. +Neutron "ports" become generic "GBP Endpoints" which can be consumed by the +GBP Renderer Manager. The GBP Renderer Manager resolves the policy for the +endpoint, i.e. it determines which communication relationships apply to the +specific endpoint, and hands the resolution to a device specific renderer, +which is the VPP renderer in the given case here. VPP renderer turns the +generic policy into VPP specific configuration. Note that in case the policy +would need to be applied to a different device, e.g. an OpenVSwitch (OVS), +then an "OVS Renderer" would be used. VPP Renderer and the topology manager +("Virtual Bridge Domain" manager - i.e. VBD) cooperate to create the actual +network configuration. VPP Renderer configures the interfaces to the virtual +machines (VM), i.e. the vhost-user interface in the given case here and +attaches them to a bridge domain on VPP. VBD handles the setup of connectivity +between bridge domains on individual VPPs, i.e. it maintains the VXLAN tunnels +in the given case here. Both VPP Renderer as well as VBD communicate with the +device through Netconf/YANG. All compute and control nodes run an instance of +VPP and the VPP-configuration agent "Honeycomb". Honeycomb serves as a +Netconf/YANG server, receives the configuration commands from VBD and VPP +Renderer and drives VPP configuration using VPP's local Java APIs. + +.. image:: FDS-simple-callflow.png + +Scenario Configuration +====================== + +To enable the "apex-os-odl-fdio-noha" scenario check the appropriate +settings in the APEX configuration files. Those are typically found in +/etc/opnfv-apex. + +File "deploy_settings.yaml": Choose Opendaylight as controller with version +"carbon" and enable vpp as forwarder. "odl_routing_node" chooses the node +which is used as the layer 3 gateway for the tenant networks. +"odl_routing_node" is an optional parameter. If omitted, the VPP on the first +compute node will serve as layer 3 gateway:: + + deploy_options: + sdn_controller: opendaylight + sdn_l3: true + odl_version: carbon + odl_routing_node: overcloud-novacompute-0 + tacker: true + congress: true + sfc: false + vpn: false + vpp: true + dataplane: fdio + performance: + Controller: + kernel: + hugepages: 1024 + hugepagesz: 2M + intel_iommu: 'on' + iommu: pt + isolcpus: 1,2 + vpp: + main-core: 1 + corelist-workers: 2 + uio-driver: uio_pci_generic + Compute: + kernel: + hugepagesz: 2M + hugepages: 2048 + intel_iommu: 'on' + iommu: pt + isolcpus: 1,2 + vpp: + main-core: 1 + corelist-workers: 2 + uio-driver: uio_pci_generic + +Limitations, Issues and Workarounds +=================================== + +For specific information on limitations and issues, please refer to the APEX +installer release notes. + +References +========== + + + * FastDataStacks OPNFV project wiki: https://wiki.opnfv.org/display/fds + * Fast Data (FD.io): https://fd.io/ + * FD.io Vector Packet Processor (VPP): https://wiki.fd.io/view/VPP + * OpenDaylight Controller: https://www.opendaylight.org/ + * OPNFV Danube release - more information: http://www.opnfv.org/danube diff --git a/docs/scenarios/os-odl_l2-fdio-ha/FDS-L3-tenant-connectivity.png b/docs/scenarios/os-odl_l2-fdio-ha/FDS-L3-tenant-connectivity.png deleted file mode 100755 index 9de77e5..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-ha/FDS-L3-tenant-connectivity.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-ha/FDS-basic-callflow.jpg b/docs/scenarios/os-odl_l2-fdio-ha/FDS-basic-callflow.jpg deleted file mode 100755 index 96464f8..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-ha/FDS-basic-callflow.jpg and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-ha/FDS-basic-components.jpg b/docs/scenarios/os-odl_l2-fdio-ha/FDS-basic-components.jpg deleted file mode 100755 index e92851f..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-ha/FDS-basic-components.jpg and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-ha/FDS-odl_l2-ha-overview.png b/docs/scenarios/os-odl_l2-fdio-ha/FDS-odl_l2-ha-overview.png deleted file mode 100755 index 78526da..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-ha/FDS-odl_l2-ha-overview.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-ha/FDS-odl_l2-overview.png b/docs/scenarios/os-odl_l2-fdio-ha/FDS-odl_l2-overview.png deleted file mode 100755 index 2755099..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-ha/FDS-odl_l2-overview.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-ha/FDS-simple-callflow.png b/docs/scenarios/os-odl_l2-fdio-ha/FDS-simple-callflow.png deleted file mode 100755 index 04546aa..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-ha/FDS-simple-callflow.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-ha/index.rst b/docs/scenarios/os-odl_l2-fdio-ha/index.rst deleted file mode 100644 index b014451..0000000 --- a/docs/scenarios/os-odl_l2-fdio-ha/index.rst +++ /dev/null @@ -1,20 +0,0 @@ -.. _os-odl_l2-fdio-ha: - -.. OPNFV - Open Platform for Network Function Virtualization -.. This work is licensed under a Creative Commons Attribution 4.0 International License. -.. http://creativecommons.org/licenses/by/4.0 - - -********************************************************************* -Fast Data Stacks Scenario: os-odl_l2-fdio-ha Overview and Description -********************************************************************* - -Scenario: "OpenStack - Opendaylight (L2) - FD.io" (apex-os-odl_l2-fdio-ha) -is a scenario developed as part of the FastDataStacks -OPNFV project. - -.. toctree:: - :numbered: - :maxdepth: 2 - - scenario.description.rst diff --git a/docs/scenarios/os-odl_l2-fdio-ha/os-odl_l2-fdio-ha-colorado2_1.png b/docs/scenarios/os-odl_l2-fdio-ha/os-odl_l2-fdio-ha-colorado2_1.png deleted file mode 100755 index aa23495..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-ha/os-odl_l2-fdio-ha-colorado2_1.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-ha/scenario.description.rst b/docs/scenarios/os-odl_l2-fdio-ha/scenario.description.rst deleted file mode 100755 index a81e8ed..0000000 --- a/docs/scenarios/os-odl_l2-fdio-ha/scenario.description.rst +++ /dev/null @@ -1,299 +0,0 @@ -.. OPNFV - Open Platform for Network Function Virtualization -.. This work is licensed under a Creative Commons Attribution 4.0 International License. -.. http://creativecommons.org/licenses/by/4.0 - -Scenario: "OpenStack - OpenDaylight (Layer 2) - FD.io" -====================================================== - -Scenario: apex-os-odl_l2-fdio-ha - -"apex-os-odl_l2-fdio-ha" is a scenario developed as part of the -FastDataStacks OPNFV project. The main components of the -"apex-os-odl_l2-fdio-ha" scenario are: - - - APEX (TripleO) installer (please also see APEX installer documentation) - - Openstack (in HA configuration) - - OpenDaylight controller in clustered mode controlling layer 2 networking - - FD.io/VPP virtual forwarder for tenant networking - -Introduction -============ - -NFV and virtualized high performance applications, such as video processing, -require a "fast data stack" solution that provides both carrier grade -forwarding performance, scalability and open extensibility, along with -functionality for realizing application policies and controlling a complex -network topology. - -A solution stack is only as good as its foundation. Key foundational assets for -NFV infrastructure are - * The virtual forwarder: The virtual forwarder needs to be a feature rich, - high performance, highly scale virtual switch-router. It needs to leverage - hardware accelerators when available and run in user space. - In addition, it should be modular and easily extensible. - * Forwarder diversity: A solution stack should support a variety of - forwarders, hardware forwarders (physical switches and routers) - as well as software forwarders. This way virtual and physical - forwarding domains can be seamlessly glued together. - * Policy driven connectivity: Connectivity should respect and - reflect different business - -In order to meet the desired qualities of an NFV infrastructure, the -following components were chosen for the "Openstack - OpenDaylight - FD.io" -scenario: - * FD.io Vector Packet Processor (VPP) - a highly scalable, - high performance, extensible virtual forwarder - * OpenDaylight Controller - an extensible controller platform which - offers the ability to separate business logic from networking - constructs, supports a diverse set of network devices - (virtual and physical) via the "group based policy (GBP)" - component, and can be clustered to achieve a highly available - deployment - as done in this scenario. - -The "Openstack - OpenDaylight - FD.io" scenario provides the capability to -realize a set of use-cases relevant to the deployment of NFV nodes instantiated -by means of an Openstack orchestration system on FD.io/VPP enabled compute -nodes. The role of the Opendaylight network controller in this integration is -twofold. It provides a network device configuration and topology abstraction -via the Openstack Neutron interface, while providing the capability to realize -more complex network policies by means of Group Based Policies. Furthermore it -also provides the capabilities to monitor as well as visualize the operation of -the virtual network devices and their topologies. -In supporting the general use-case of instantiatiting an NFV instance, two -specific types of network transport use cases are realized: - - * NFV instances with VPP data-plane forwarding using a VLAN provider network - * NFV instances with VPP data-plane forwarding using a VXLAN overlay - transport network - -A deployment of the "apex-os-odl_l2-fdio-ha" scenario consists of 4 or more -servers: - - * 1 Jumphost hosting the APEX installer - running the Undercloud - * 3 Controlhosts, which run the Overcloud as well as OpenDaylight - as a network controller. OpenDaylight is deployed in clustered - mode and runs on all 3 control nodes. - * 2 or more Computehosts - -.. image:: FDS-odl_l2-ha-overview.png - -Tenant networking leverages FD.io/VPP. Open VSwitch (OVS) is used for all other -connectivity, in particular the connectivity to public networking / the -Internet (i.e. br-ext) is performed via OVS as in any standard OpenStack -deployment. The OpenDaylight network controller is used to setup and manage -layer 2 networking for the scenario. Tenant networking can either leverage -VXLAN (in which case a full mesh of VXLAN tunnels is created) or VLANs. Layer 3 -connectivity for a tenant network is provided centrally via qrouter on the -control node. As in a standard OpenStack deployment, the Layer3 agent -configures the qrouter and associated rulesets for security (security groups) -and NAT (floating IPs). Public IP network connectivity for a tenant network is -provided by interconnecting the VPP-based bridge domain representing the tenant -network to qrouter using a tap interface. The setup is depicted below: - -.. image:: FDS-L3-tenant-connectivity.png - -With high availability factored in the setup looks like the following. - -.. image:: os-odl_l2-fdio-ha-colorado2_1.png - -Note that the picture only shows two Controllernodes for reasons of -simplicity. A HA deployment will always include 3 Controllernodes. - - -Features of the scenario ------------------------- - -Main features of the "apex-os-odl_l2-fdio-ha" scenario: - - * Automated installation using the APEX installer - * Fast and scalable tenant networking using FD.io/VPP as forwarder - * Layer 2 networking using VLANs or VXLAN, managed and - controlled through OpenDaylight - * Layer 3 connectivitiy for tenant networks supplied centrally on - the Control node through standard OpenStack mechanisms. - All layer 3 features apply, including floating IPs (i.e. NAT) - and security groups. - * Manual and automatic (via DHCP) addressing on tenant networks - * OpenDaylight controller high availability (clustering) - * OpenStack high availability - -Scenario components and composition -=================================== - -The apex-os-odl_l2-fdio-ha scenario combines components from three key open -source projects: OpenStack, OpenDaylight, and Fast Data (FD.io). The key -components that realize the apex-os-odl_l2-fdio-ha scenario and which differ -from a regular, OVS-based scenario, are the OpenStack ML2 OpenDaylight plugin, -OpenDaylight Neutron Northbound, OpenDaylight Group Based Policy, OpenDaylight -Virtual Bridge Domain Manager, FD.io Honeycomb management agent and FD.io -Vector Packet Processor (VPP). - -Here's a more detailed list of the individual software components involved: - -**Openstack Neutron ML2 OpenDaylight Plugin**: Handles Neutron data base -synchronization and interaction with the southbound controller using a REST -interface. - -**ODL GBP Neutron Mapper**: Maps neutron elements like networks, subnets, -security groups, etc. to GBP entities: Creates policy and configuration for -tenants (endpoints, resolved policies, forwarding rules). - -**ODL GBP Neutron VPP Mapper**: Maps Neutron ports to VPP endpoints in GBP. - -**ODL GBP Location Manager**: Provides real location for endpoints (i.e. Which -physical node an endpoint is connected to). - -**GBP Renderer Manager**: Creates configuration for Renderers (like e.g. -VPP-Renderer or OVS-Renderer). The GBP Renderer Manager is the central point -for dispatching of data to specific device renderers. It uses the information -derived from the GBP end-point and its topology entries to dispatch the task -of configuration to a specific device renderer by writing a renderer policy -configuration into the registered renderer's policy store. The renderer -manager also monitors, by being a data change listener on the VPP Renderer -Policy States, for any errors in the application of a rendered configuration. - -**GBP VPP Renderer Interface Manager**: Listens to VPP endpoints in the -Config DataStore and configures associated interfaces on VPP via HoneyComb. - -**GBP VPP Renderer Renderer Policy Manager**: Manages the creation of -bridge domains using VBD and assigns interfaces to bridge domains. - -**Virtual Bridge Domain Manager (VBD)**: Creates bridge domains (i.e. in case -of VXLAN creates full mesh of VXLAN tunnels, configures split horizon on -tunnel endpoints etc.). VDB configures VXLAN tunnels always into a full-mesh -with split-horizon group forwarding applied on any domain facing tunnel -interface (i.e. forwarding behavior will be that used for VPLS). - -**Virtual Packet Processor (VPP) and Honeycomb server**: The VPP is the -accelerated data plane forwarding engine relying on vhost user interfaces -towards Virtual Machines created by the Nova Agent. The Honeycomb NETCONF -configuration server is responsible for driving the configuration of the VPP, -and collecting the operational data. - -**Nova Agent**: The Nova Agent, a sub-component of the overall Openstack -architecture, is responsible for interacting with the compute node's host -Libvirt API to drive the life-cycle of Virtual Machines. It, along with the -compute node software, are assumed to be capable of supporting vhost user -interfaces. - -The picture below shows the key components. - -.. image:: FDS-basic-components.jpg - -To provide a better understanding how the above mentioned components interact -with each other, the following diagram shows how the example of creating a -vhost-user port on VPP through Openstack Neutron: - -To create or update a port, Neutron will send a request to ODL Neutron -Northbound which contains the UUID, along with the host-id as "vpp" and -vif-type as "vhost-user". The GBP Neutron mapper turns the "Neutron speak" of -"ports" into the generic connectivity model that GroupBasedPolicy uses. -Neutron "ports" become generic "GBP Endpoints" which can be consumed by the -GBP Renderer Manager. The GBP Renderer Manager resolves the policy for the -endpoint, i.e. it determines which communication relationships apply to the -specific endpoint, and hands the resolution to a device specific renderer, -which is the VPP renderer in the given case here. VPP renderer turns the -generic policy into VPP specific configuration. Note that in case the policy -would need to be applied to a different device, e.g. an OpenVSwitch (OVS), -then an "OVS Renderer" would be used. VPP Renderer and the topology manager -("Virtual Bridge Domain" manager - i.e. VBD) cooperate to create the actual -network configuration. VPP Renderer configures the interfaces to the virtual -machines (VM), i.e. the vhost-user interface in the given case here and -attaches them to a bridge domain on VPP. VBD handles the setup of connectivity -between bridge domains on individual VPPs, i.e. it maintains the VXLAN tunnels -in the given case here. Both VPP Renderer as well as VBD communicate with the -device through Netconf/YANG. All compute and control nodes run an instance of -VPP and the VPP-configuration agent "Honeycomb". Honeycomb serves as a -Netconf/YANG server, receives the configuration commands from VBD and VPP -Renderer and drives VPP configuration using VPP's local Java APIs. - -.. image:: FDS-simple-callflow.png - - -Scenario Configuration -====================== - -To enable the "apex-os-odl_l2-fdio-ha" scenario check the appropriate -settings in the APEX configuration files. Those are typically found in -/etc/opnfv-apex. - -File "deploy_settings.yaml" choose opendaylight as controller with version -"carbon" and enable vpp as forwarder. Also make sure that you set -"ha_enabled" to "true" in the global_params section. "ha_enabled" is the -only real difference from a configuration file perspective between the -scenario with high availability when compared to the ODL-L2 scenario -without high-availability support. "hugepages" need to set to a -sufficiently large value for VPP to work. The default value for VPP is -1024, but this only allows for a few VMs to be started. If feasible, -choose a significantly larger number on the compute nodes:: - - global_params: - ha_enabled: true - - deploy_options: - sdn_controller: opendaylight - sdn_l3: false - odl_version: carbon - tacker: true - congress: true - sfc: false - vpn: false - vpp: true - dataplane: fdio - performance: - Controller: - kernel: - hugepages: 1024 - hugepagesz: 2M - intel_iommu: 'on' - iommu: pt - isolcpus: 1,2 - vpp: - main-core: 1 - corelist-workers: 2 - uio-driver: uio_pci_generic - Compute: - kernel: - hugepagesz: 2M - hugepages: 2048 - intel_iommu: 'on' - iommu: pt - isolcpus: 1,2 - vpp: - main-core: 1 - corelist-workers: 2 - uio-driver: uio_pci_generic - - -Validated deployment environments -================================= - -The "os-odl_l2-fdio-ha" scenario has been deployed and tested -on the following sets of hardware: - * Linux Foundation lab (Chassis: Cisco UCS-B-5108 blade server, - NICs: 8 external / 32 internal 10GE ports, - RAM: 32G (4 x 8GB DDR4-2133-MHz RDIMM/PC4-17000/single rank/x4/1.2v), - CPU: 3.50 GHz E5-2637 v3/135W 4C/15MB Cache/DDR4 2133MHz - Disk: 1.2 TB 6G SAS 10K rpm SFF HDD) see also: - https://wiki.opnfv.org/display/pharos/Lflab+Hosting - * OPNFV CENGN lab (https://wiki.opnfv.org/display/pharos/CENGN+Pharos+Lab) - * Cisco internal development labs (UCS-B and UCS-C) - -Limitations, Issues and Workarounds -=================================== - -For specific information on limitations and issues, please refer to the APEX -installer release notes. Note that this high availability scenario -deploys OpenStack in HA mode *and* OpenDaylight in cluster mode. - - -References -========== - - - * FastDataStacks OPNFV project wiki: https://wiki.opnfv.org/display/fds - * Fast Data (FD.io): https://fd.io/ - * FD.io Vector Packet Processor (VPP): https://wiki.fd.io/view/VPP - * OpenDaylight Controller: https://www.opendaylight.org/ - * OPNFV Danube release - more information: http://www.opnfv.org/danube - diff --git a/docs/scenarios/os-odl_l2-fdio-noha/FDS-L3-tenant-connectivity.png b/docs/scenarios/os-odl_l2-fdio-noha/FDS-L3-tenant-connectivity.png deleted file mode 100755 index 9de77e5..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-noha/FDS-L3-tenant-connectivity.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-noha/FDS-basic-callflow.jpg b/docs/scenarios/os-odl_l2-fdio-noha/FDS-basic-callflow.jpg deleted file mode 100755 index 96464f8..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-noha/FDS-basic-callflow.jpg and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-noha/FDS-basic-components.jpg b/docs/scenarios/os-odl_l2-fdio-noha/FDS-basic-components.jpg deleted file mode 100755 index e92851f..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-noha/FDS-basic-components.jpg and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-noha/FDS-odl_l2-overview.png b/docs/scenarios/os-odl_l2-fdio-noha/FDS-odl_l2-overview.png deleted file mode 100755 index 2755099..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-noha/FDS-odl_l2-overview.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-noha/FDS-simple-callflow.png b/docs/scenarios/os-odl_l2-fdio-noha/FDS-simple-callflow.png deleted file mode 100755 index 04546aa..0000000 Binary files a/docs/scenarios/os-odl_l2-fdio-noha/FDS-simple-callflow.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l2-fdio-noha/index.rst b/docs/scenarios/os-odl_l2-fdio-noha/index.rst deleted file mode 100644 index a324f11..0000000 --- a/docs/scenarios/os-odl_l2-fdio-noha/index.rst +++ /dev/null @@ -1,20 +0,0 @@ -.. _os-odl_l2-fdio-noha: - -.. OPNFV - Open Platform for Network Function Virtualization -.. This work is licensed under a Creative Commons Attribution 4.0 International License. -.. http://creativecommons.org/licenses/by/4.0 - - -*********************************************************************** -Fast Data Stacks Scenario: os-odl_l2-fdio-noha Overview and Description -*********************************************************************** - -Scenario: "OpenStack - Opendaylight (L2) - FD.io" (apex-os-odl_l2-fdio-noha) -is a scenario developed as part of the FastDataStacks -OPNFV project. - -.. toctree:: - :numbered: - :maxdepth: 2 - - scenario.description.rst diff --git a/docs/scenarios/os-odl_l2-fdio-noha/scenario.description.rst b/docs/scenarios/os-odl_l2-fdio-noha/scenario.description.rst deleted file mode 100755 index b392070..0000000 --- a/docs/scenarios/os-odl_l2-fdio-noha/scenario.description.rst +++ /dev/null @@ -1,281 +0,0 @@ -.. OPNFV - Open Platform for Network Function Virtualization -.. This work is licensed under a Creative Commons Attribution 4.0 International License. -.. http://creativecommons.org/licenses/by/4.0 - -Scenario: "OpenStack - OpenDaylight (Layer 2) - FD.io" -====================================================== - -Scenario: apex-os-odl_l2-fdio-noha - -"apex-os-odl_l2-fdio-noha" is a scenario developed as part of the -FastDataStacks OPNFV project. The main components of the -"apex-os-odl_l2-fdio-noha" scenario are: - - - APEX (TripleO) installer (please also see APEX installer documentation) - - Openstack (in non-HA configuration) - - OpenDaylight controller (non-clustered) controlling layer 2 networking - - FD.io/VPP virtual forwarder for tenant networking - -Introduction -============ - -NFV and virtualized high performance applications, such as video processing, -require a "fast data stack" solution that provides both carrier grade -forwarding performance, scalability and open extensibility, along with -functionality for realizing application policies and controlling a complex -network topology. - -A solution stack is only as good as its foundation. Key foundational assets for -NFV infrastructure are - * The virtual forwarder: The virtual forwarder needs to be a feature rich, - high performance, highly scale virtual switch-router. It needs to leverage - hardware accelerators when available and run in user space. - In addition, it should be modular and easily extensible. - * Forwarder diversity: A solution stack should support a variety of - forwarders, hardware forwarders (physical switches and routers) - as well as software forwarders. This way virtual and physical - forwarding domains can be seamlessly glued together. - * Policy driven connectivity: Connectivity should respect and - reflect different business - -In order to meet the desired qualities of an NFV infrastructure, the -following components were chosen for the "Openstack - OpenDaylight - FD.io" -scenario: - * FD.io Vector Packet Processor (VPP) - a highly scalable, - high performance, extensible virtual forwarder - * OpenDaylight Controller - an extensible controller platform which - offers the ability to separate business logic from networking - constructs, supports a diverse set of network devices - (virtual and physical) via the "group based policy (GBP)" - component, and can be clustered to achieve a highly available - deployment. - -The "Openstack - OpenDaylight - FD.io" scenario provides the capability to -realize a set of use-cases relevant to the deployment of NFV nodes instantiated -by means of an Openstack orchestration system on FD.io/VPP enabled compute -nodes. The role of the Opendaylight network controller in this integration is -twofold. It provides a network device configuration and topology abstraction -via the Openstack Neutron interface, while providing the capability to realize -more complex network policies by means of Group Based Policies. Furthermore it -also provides the capabilities to monitor as well as visualize the operation of -the virtual network devices and their topologies. -In supporting the general use-case of instantiatiting an NFV instance, two -specific types of network transport use cases are realized: - - * NFV instances with VPP data-plane forwarding using a VLAN provider network - * NFV instances with VPP data-plane forwarding using a VXLAN overlay - transport network - -A deployment of the "apex-os-odl_l2-fdio-noha" scenario consists of 4 or more -servers: - - * 1 Jumphost hosting the APEX installer - running the Undercloud - * 1 Controlhost, which runs the Overcloud as well as OpenDaylight - as a network controller - * 2 or more Computehosts - -.. image:: FDS-odl_l2-overview.png - -Tenant networking leverages FD.io/VPP. Open VSwitch (OVS) is used for all -other connectivity, in particular the connectivity to public networking / the -Internet (i.e. br-ext) is performed via OVS as in any standard OpenStack -deployment. The OpenDaylight network controller is used to setup and manage -layer 2 networking for the scenario. Tenant networking can either leverage -VXLAN (in which case a full mesh of VXLAN tunnels is created) or VLANs. Layer 3 -connectivity for a tenant network is provided centrally via qrouter on the -control node. As in a standard OpenStack deployment, the Layer3 agent -configures the qrouter and associated rulesets for security (security groups) -and NAT (floating IPs). Public IP network connectivity for a tenant network is -provided by interconnecting the VPP-based bridge domain representing the tenant -network to qrouter using a tap interface. The setup is depicted below: - -.. image:: FDS-L3-tenant-connectivity.png - -Features of the scenario ------------------------- - -Main features of the "apex-os-odl_l2-fdio-noha" scenario: - - * Automated installation using the APEX installer - * Fast and scalable tenant networking using FD.io/VPP as forwarder - * Layer 2 networking using VLANs or VXLAN, managed and - controlled through OpenDaylight - * Layer 3 connectivitiy for tenant networks supplied centrally on - the Control node through standard OpenStack mechanisms. - All layer 3 features apply, including floating IPs (i.e. NAT) - and security groups. - * Manual and automatic (via DHCP) addressing on tenant networks - -Scenario components and composition -=================================== - -The apex-os-odl_l2-fdio-noha scenario combines components from three key open -source projects: OpenStack, OpenDaylight, and Fast Data (FD.io). The key -components that realize the apex-os-odl_l2-fdio-noha scenario and which differ -from a regular, OVS-based scenario, are the OpenStack ML2 OpenDaylight plugin, -OpenDaylight Neutron Northbound, OpenDaylight (ODL) Group Based Policy (GBP), -OpenDaylight Virtual Bridge Domain Manager (VBD), FD.io Honeycomb management -agent and FD.io Vector Packet Processor (VPP). - -Here's a more detailed list of the individual software components involved: - -**Openstack Neutron ML2 OpenDaylight Plugin**: Handles Neutron data base -synchronization and interaction with the southbound controller using a REST -interface. - -**ODL GBP Neutron Mapper**: Maps neutron elements like networks, subnets, -security groups, etc. to GBP entities: Creates policy and configuration for -tenants (endpoints, resolved policies, forwarding rules). - -**ODL GBP Neutron VPP Mapper**: Maps Neutron ports to VPP endpoints in GBP. - -**ODL GBP Location Manager**: Provides real location for endpoints (i.e. Which -physical node an endpoint is connected to). - -**GBP Renderer Manager**: Creates configuration for Renderers (like e.g. -VPP-Renderer or OVS-Renderer). The GBP Renderer Manager is the central point -for dispatching of data to specific device renderers. It uses the information -derived from the GBP end-point and its topology entries to dispatch the task -of configuration to a specific device renderer by writing a renderer policy -configuration into the registered renderer's policy store. The renderer -manager also monitors, by being a data change listener on the VPP Renderer -Policy States, for any errors in the application of a rendered configuration. - -**GBP VPP Renderer Interface Manager**: Listens to VPP endpoints in the -Config DataStore and configures associated interfaces on VPP via HoneyComb. - -**GBP VPP Renderer Renderer Policy Manager**: Manages the creation of -bridge domains using VBD and assigns interfaces to bridge domains. - -**Virtual Bridge Domain Manager (VBD)**: Creates bridge domains (i.e. in case -of VXLAN creates full mesh of VXLAN tunnels, configures split horizon on -tunnel endpoints etc.). VDB configures VXLAN tunnels always into a full-mesh -with split-horizon group forwarding applied on any domain facing tunnel -interface (i.e. forwarding behavior will be that used for VPLS). - -**Virtual Packet Processor (VPP) and Honeycomb server**: The VPP is the -accelerated data plane forwarding engine relying on vhost user interfaces -towards Virtual Machines created by the Nova Agent. The Honeycomb NETCONF -configuration server is responsible for driving the configuration of the VPP, -and collecting the operational data. - -**Nova Agent**: The Nova Agent, a sub-component of the overall Openstack -architecture, is responsible for interacting with the compute node's host -Libvirt API to drive the life-cycle of Virtual Machines. It, along with the -compute node software, are assumed to be capable of supporting vhost user -interfaces. - -The picture below shows the key components. - -.. image:: FDS-basic-components.jpg - -To provide a better understanding how the above mentioned components interact -with each other, the following diagram shows how the example of creating a -vhost-user port on VPP through Openstack Neutron: - -To create or update a port, Neutron will send a request to ODL Neutron -Northbound which contains the UUID, along with the host-id as "vpp" and -vif-type as "vhost-user". The GBP Neutron mapper turns the "Neutron speak" of -"ports" into the generic connectivity model that GroupBasedPolicy uses. -Neutron "ports" become generic "GBP Endpoints" which can be consumed by the -GBP Renderer Manager. The GBP Renderer Manager resolves the policy for the -endpoint, i.e. it determines which communication relationships apply to the -specific endpoint, and hands the resolution to a device specific renderer, -which is the VPP renderer in the given case here. VPP renderer turns the -generic policy into VPP specific configuration. Note that in case the policy -would need to be applied to a different device, e.g. an OpenVSwitch (OVS), -then an "OVS Renderer" would be used. VPP Renderer and the topology manager -("Virtual Bridge Domain" manager - i.e. VBD) cooperate to create the actual -network configuration. VPP Renderer configures the interfaces to the virtual -machines (VM), i.e. the vhost-user interface in the given case here and -attaches them to a bridge domain on VPP. VBD handles the setup of connectivity -between bridge domains on individual VPPs, i.e. it maintains the VXLAN tunnels -in the given case here. Both VPP Renderer as well as VBD communicate with the -device through Netconf/YANG. All compute and control nodes run an instance of -VPP and the VPP-configuration agent "Honeycomb". Honeycomb serves as a -Netconf/YANG server, receives the configuration commands from VBD and VPP -Renderer and drives VPP configuration using VPP's local Java APIs. - -.. image:: FDS-simple-callflow.png - - -Scenario Configuration -====================== - -To enable the "apex-os-odl_l2-fdio-noha" scenario check the appropriate -settings in the APEX configuration files. Those are typically found in -/etc/opnfv-apex. - -File "deploy_settings.yaml" choose opendaylight as controller with version -"carbon" and enable vpp as forwarder. "hugepages" need to set to a -sufficiently large value for VPP to work. The default value for VPP is 1024, -but this only allows for a few VMs to be started. If feasible, choose a -significantly larger number on the compute nodes:: - - global_params: - ha_enabled: false - - deploy_options: - sdn_controller: opendaylight - sdn_l3: false - odl_version: carbon - tacker: true - congress: true - sfc: false - vpn: false - vpp: true - dataplane: fdio - performance: - Controller: - kernel: - hugepages: 1024 - hugepagesz: 2M - intel_iommu: 'on' - iommu: pt - isolcpus: 1,2 - vpp: - main-core: 1 - corelist-workers: 2 - uio-driver: uio_pci_generic - Compute: - kernel: - hugepagesz: 2M - hugepages: 2048 - intel_iommu: 'on' - iommu: pt - isolcpus: 1,2 - vpp: - main-core: 1 - corelist-workers: 2 - uio-driver: uio_pci_generic - - -Validated deployment environments -================================= - -The "os-odl_l2-fdio-noha" scenario has been deployed and tested -on the following sets of hardware: - * Linux Foundation lab (Chassis: Cisco UCS-B-5108 blade server, - NICs: 8 external / 32 internal 10GE ports, - RAM: 32G (4 x 8GB DDR4-2133-MHz RDIMM/PC4-17000/single rank/x4/1.2v), - CPU: 3.50 GHz E5-2637 v3/135W 4C/15MB Cache/DDR4 2133MHz - Disk: 1.2 TB 6G SAS 10K rpm SFF HDD) see also: - https://wiki.opnfv.org/display/pharos/Lflab+Hosting - * OPNFV CENGN lab (https://wiki.opnfv.org/display/pharos/CENGN+Pharos+Lab) - * Cisco internal development labs (UCS-B and UCS-C) - -Limitations, Issues and Workarounds -=================================== - -For specific information on limitations and issues, please refer to the APEX installer release notes. - -References -========== - - - * FastDataStacks OPNFV project wiki: https://wiki.opnfv.org/display/fds - * Fast Data (FD.io): https://fd.io/ - * FD.io Vector Packet Processor (VPP): https://wiki.fd.io/view/VPP - * OpenDaylight Controller: https://www.opendaylight.org/ - * OPNFV Carbon release - more information: http://www.opnfv.org/carbon - diff --git a/docs/scenarios/os-odl_l3-fdio-ha/FDS-L3-DVR-example.png b/docs/scenarios/os-odl_l3-fdio-ha/FDS-L3-DVR-example.png deleted file mode 100755 index 18932c3..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-ha/FDS-L3-DVR-example.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-ha/FDS-L3-noha-sample-setup.png b/docs/scenarios/os-odl_l3-fdio-ha/FDS-L3-noha-sample-setup.png deleted file mode 100755 index 27c8335..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-ha/FDS-L3-noha-sample-setup.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-ha/FDS-basic-callflow.jpg b/docs/scenarios/os-odl_l3-fdio-ha/FDS-basic-callflow.jpg deleted file mode 100755 index 96464f8..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-ha/FDS-basic-callflow.jpg and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-ha/FDS-basic-components.jpg b/docs/scenarios/os-odl_l3-fdio-ha/FDS-basic-components.jpg deleted file mode 100755 index e92851f..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-ha/FDS-basic-components.jpg and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-ha/FDS-odl_l3-noha-overview.png b/docs/scenarios/os-odl_l3-fdio-ha/FDS-odl_l3-noha-overview.png deleted file mode 100755 index 1193ea4..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-ha/FDS-odl_l3-noha-overview.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-ha/FDS-odl_l3-overview.png b/docs/scenarios/os-odl_l3-fdio-ha/FDS-odl_l3-overview.png deleted file mode 100755 index 5b90c93..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-ha/FDS-odl_l3-overview.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-ha/FDS-simple-callflow.png b/docs/scenarios/os-odl_l3-fdio-ha/FDS-simple-callflow.png deleted file mode 100755 index 04546aa..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-ha/FDS-simple-callflow.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-ha/index.rst b/docs/scenarios/os-odl_l3-fdio-ha/index.rst deleted file mode 100644 index 3ccdda7..0000000 --- a/docs/scenarios/os-odl_l3-fdio-ha/index.rst +++ /dev/null @@ -1,20 +0,0 @@ -.. _os-odl_l3-fdio-ha: - -.. OPNFV - Open Platform for Network Function Virtualization -.. This work is licensed under a Creative Commons Attribution 4.0 International License. -.. http://creativecommons.org/licenses/by/4.0 - - -********************************************************************* -Fast Data Stacks Scenario: os-odl_l3-fdio-ha Overview and Description -********************************************************************* - -Scenario: "OpenStack - Opendaylight (L3) - FD.io" (apex-os-odl_l3-fdio-ha) -is a scenario developed as part of the FastDataStacks -OPNFV project. - -.. toctree:: - :numbered: - :maxdepth: 2 - - scenario.description.rst diff --git a/docs/scenarios/os-odl_l3-fdio-ha/scenario.description.rst b/docs/scenarios/os-odl_l3-fdio-ha/scenario.description.rst deleted file mode 100755 index 4a69a32..0000000 --- a/docs/scenarios/os-odl_l3-fdio-ha/scenario.description.rst +++ /dev/null @@ -1,280 +0,0 @@ -.. OPNFV - Open Platform for Network Function Virtualization -.. This work is licensed under a Creative Commons Attribution 4.0 International License. -.. http://creativecommons.org/licenses/by/4.0 - -Scenario: "OpenStack - OpenDaylight (Layer 3) - FD.io" -====================================================== - -Scenario: apex-os-odl_l3-fdio-ha - -"apex-os-odl_l3-fdio-ha" is a scenario developed as part of the -FastDataStacks OPNFV project. The main components of the -"apex-os-odl_l3-fdio-ha" scenario are: - - - APEX (TripleO) installer (please also see APEX installer documentation) - - Openstack (in HA configuration) - - OpenDaylight controller (clustered) - controlling layer 2 and layer 3 networking - - FD.io/VPP virtual forwarder for tenant networking - -Introduction -============ - -NFV and virtualized high performance applications, such as video processing, -require a "fast data stack" solution that provides both carrier grade -forwarding performance, scalability and open extensibility, along with -functionality for realizing application policies and controlling a complex -network topology. - -A solution stack is only as good as its foundation. Key foundational assets for -NFV infrastructure are - * The virtual forwarder: The virtual forwarder needs to be a feature rich, - high performance, highly scale virtual switch-router. It needs to leverage - hardware accelerators when available and run in user space. - In addition, it should be modular and easily extensible. - * Forwarder diversity: A solution stack should support a variety of - forwarders, hardware forwarders (physical switches and routers) - as well as software forwarders. This way virtual and physical - forwarding domains can be seamlessly glued together. - * Policy driven connectivity: Connectivity should respect and - reflect different business - -In order to meet the desired qualities of an NFV infrastructure, the -following components were chosen for the "Openstack - OpenDaylight - FD.io" -scenario: - * FD.io Vector Packet Processor (VPP) - a highly scalable, - high performance, extensible virtual forwarder - * OpenDaylight Controller - an extensible controller platform which - offers the ability to separate business logic from networking - constructs, supports a diverse set of network devices - (virtual and physical) via the "group based policy (GBP)" - component, and can be clustered to achieve a highly available - deployment. - - -The "Openstack - OpenDaylight - FD.io" scenario provides the capability to -realize a set of use-cases relevant to the deployment of NFV nodes instantiated -by means of an Openstack orchestration system on FD.io/VPP enabled compute -nodes. The role of the Opendaylight network controller in this integration is -twofold. It provides a network device configuration and topology abstraction -via the Openstack Neutron interface, while providing the capability to realize -more complex network policies by means of Group Based Policies. Furthermore it -also provides the capabilities to monitor as well as visualize the operation of -the virtual network devices and their topologies. In supporting the general -use-case of instantiatiting an NFV instance, two specific types of network -transport use cases are realized: - - * NFV instances with VPP data-plane forwarding using a VLAN provider network - * NFV instances with VPP data-plane forwarding using a VXLAN overlay - transport network - -A deployment of the "apex-os-odl_l3-fdio-ha" scenario consists of 6 or more -servers: - - * 1 Jumphost hosting the APEX installer - running the Undercloud - * 3 Controlhosts, which run the Overcloud as well as - 3 instances of OpenDaylight as a network controller - * 2 or more Computehosts. One of the Computehosts also serves as - layer 3 gateway for the tenant networks. - -The following picture depicts a non-HA setup of the L3 scenario. - -.. image:: FDS-odl_l3-noha-overview.png - -Tenant and public networking leverages FD.io/VPP. On one of the compute nodes, -VPP binds to both, the tenant networking interface as well as to the public -networking interface. This means that VPP is used for communication within -a tenant network, between tenant networks, as well as between a tenant network -and the Internet. Note that this setup slightly differs from the usual -centralized L3 setup with qrouter on the control node. This setup was chosen -to limit the configuration changes for the introduction of FD.io/VPP. The -OpenDaylight network controller is used to setup and manage layer 2 and -layer 3 networking for the scenario - with Group Based Policy (GBP) being the -key component. Tenant networking can either leverage VXLAN (in which case a -full mesh of VXLAN tunnels is created) or VLANs. - -The picture below shows an example setup with two compute and one control -node. Note that the external network is connected via compute node 0 through -VPP. VPP provides all layer 3 services which are provided in a "vanilla" -OpenStack deployment, including SNAT and DNAT, as well as north-south -and east-west traffic filtering for security purposes ("security groups"). - -.. image:: FDS-L3-noha-sample-setup.png - -Features of the scenario ------------------------- - -Main features of the "apex-os-odl_l3-fdio-ha" scenario: - - * Automated installation using the APEX installer - * Fast and scalable tenant networking using FD.io/VPP as forwarder - * Layer 2 networking using VLANs or VXLAN, managed - and controlled through OpenDaylight - * Layer 3 connectivitiy for tenant networks supplied - through FD.io/VPP. Layer 3 features, including security groups as well as - floating IP addresses (i.e. NAT) are implemented by the FD.io/VPP forwarder - * Manual and automatic (via DHCP) addressing on tenant networks - * OpenStack high-availability and OpenDaylight clustered mode - -Scenario components and composition -=================================== - -The apex-os-odl_l3-fdio-ha scenario combines components from three key open -source projects: OpenStack, OpenDaylight, and Fast Data (FD.io). The key -components that realize the apex-os-odl_l3-fdio-ha scenario and which differ -from a regular, OVS-based scenario, are the OpenStack ML2 OpenDaylight plugin, -OpenDaylight Neutron Northbound, OpenDaylight Group Based Policy, OpenDaylight -Virtual Bridge Domain Manager, FD.io Honeycomb management agent and FD.io -Vector Packet Processor (VPP). - -Note that the key components of the OpenDaylight based scenarios of -FastDataStacks are the same. The Layer 2 scenario "apex-os-odl_l2-fdio-ha" -and the Layer 3 scenario "apex-os-odl_l3-fdio-ha" share the same components. - -Here's a more detailed list of the individual software components involved: - -**Openstack Neutron ML2 OpenDaylight Plugin**: Handles Neutron data base -synchronization and interaction with the southbound controller using a REST -interface. - -**ODL GBP Neutron Mapper**: Maps neutron elements like networks, subnets, -security groups, etc. to GBP entities: Creates policy and configuration for -tenants (endpoints, resolved policies, forwarding rules). - -**ODL GBP Neutron VPP Mapper**: Maps Neutron ports to VPP endpoints in GBP. - -**ODL GBP Location Manager**: Provides real location for endpoints (i.e. Which -physical node an endpoint is connected to). - -**GBP Renderer Manager**: Creates configuration for Renderers (like e.g. -VPP-Renderer or OVS-Renderer). The GBP Renderer Manager is the central point -for dispatching of data to specific device renderers. It uses the information -derived from the GBP end-point and its topology entries to dispatch the task -of configuration to a specific device renderer by writing a renderer policy -configuration into the registered renderer's policy store. The renderer -manager also monitors, by being a data change listener on the VPP Renderer -Policy States, for any errors in the application of a rendered configuration. - -**GBP VPP Renderer Interface Manager**: Listens to VPP endpoints in the -Config DataStore and configures associated interfaces on VPP via HoneyComb. - -**GBP VPP Renderer Renderer Policy Manager**: Manages the creation of -bridge domains using VBD and assigns interfaces to bridge domains. - -**Virtual Bridge Domain Manager (VBD)**: Creates bridge domains (i.e. in case -of VXLAN creates full mesh of VXLAN tunnels, configures split horizon on -tunnel endpoints etc.). VDB configures VXLAN tunnels always into a full-mesh -with split-horizon group forwarding applied on any domain facing tunnel -interface (i.e. forwarding behavior will be that used for VPLS). - -**Virtual Packet Processor (VPP) and Honeycomb server**: The VPP is the -accelerated data plane forwarding engine relying on vhost user interfaces -towards Virtual Machines created by the Nova Agent. The Honeycomb NETCONF -configuration server is responsible for driving the configuration of the VPP, -and collecting the operational data. - -**Nova Agent**: The Nova Agent, a sub-component of the overall Openstack -architecture, is responsible for interacting with the compute node's host -Libvirt API to drive the life-cycle of Virtual Machines. It, along with the -compute node software, are assumed to be capable of supporting vhost user -interfaces. - -The picture below shows the key components. - -.. image:: FDS-basic-components.jpg - -To provide a better understanding how the above mentioned components interact -with each other, the following diagram shows how the example of creating a -vhost-user port on VPP through Openstack Neutron: - -To create or update a port, Neutron will send a request to ODL Neutron -Northbound which contains the UUID, along with the host-id as "vpp" and -vif-type as "vhost-user". The GBP Neutron mapper turns the "Neutron speak" of -"ports" into the generic connectivity model that GroupBasedPolicy uses. -Neutron "ports" become generic "GBP Endpoints" which can be consumed by the -GBP Renderer Manager. The GBP Renderer Manager resolves the policy for the -endpoint, i.e. it determines which communication relationships apply to the -specific endpoint, and hands the resolution to a device specific renderer, -which is the VPP renderer in the given case here. VPP renderer turns the -generic policy into VPP specific configuration. Note that in case the policy -would need to be applied to a different device, e.g. an OpenVSwitch (OVS), -then an "OVS Renderer" would be used. VPP Renderer and the topology manager -("Virtual Bridge Domain" manager - i.e. VBD) cooperate to create the actual -network configuration. VPP Renderer configures the interfaces to the virtual -machines (VM), i.e. the vhost-user interface in the given case here and -attaches them to a bridge domain on VPP. VBD handles the setup of connectivity -between bridge domains on individual VPPs, i.e. it maintains the VXLAN tunnels -in the given case here. Both VPP Renderer as well as VBD communicate with the -device through Netconf/YANG. All compute and control nodes run an instance of -VPP and the VPP-configuration agent "Honeycomb". Honeycomb serves as a -Netconf/YANG server, receives the configuration commands from VBD and VPP -Renderer and drives VPP configuration using VPP's local Java APIs. - -.. image:: FDS-simple-callflow.png - -Scenario Configuration -====================== - -To enable the "apex-os-odl_l3-fdio-ha" scenario check the appropriate -settings in the APEX configuration files. Those are typically found in -/etc/opnfv-apex. - -File "deploy_settings.yaml": Choose Opendaylight as controller with version -"carbon" and enable vpp as forwarder. "odl_routing_node" chooses the node -which is used as the layer 3 gateway for the tenant networks. -"odl_routing_node" is an optional parameter. If omitted, the VPP on the first -compute node will serve as layer 3 gateway:: - - global_params: - ha_enabled: true - - deploy_options: - sdn_controller: opendaylight - sdn_l3: true - odl_version: carbon - odl_routing_node: overcloud-novacompute-0 - tacker: true - congress: true - sfc: false - vpn: false - vpp: true - dataplane: fdio - performance: - Controller: - kernel: - hugepages: 1024 - hugepagesz: 2M - intel_iommu: 'on' - iommu: pt - isolcpus: 1,2 - vpp: - main-core: 1 - corelist-workers: 2 - uio-driver: uio_pci_generic - Compute: - kernel: - hugepagesz: 2M - hugepages: 2048 - intel_iommu: 'on' - iommu: pt - isolcpus: 1,2 - vpp: - main-core: 1 - corelist-workers: 2 - uio-driver: uio_pci_generic - -Limitations, Issues and Workarounds -=================================== - -For specific information on limitations and issues, please refer to the APEX -installer release notes. - -References -========== - - - * FastDataStacks OPNFV project wiki: https://wiki.opnfv.org/display/fds - * Fast Data (FD.io): https://fd.io/ - * FD.io Vector Packet Processor (VPP): https://wiki.fd.io/view/VPP - * OpenDaylight Controller: https://www.opendaylight.org/ - * OPNFV Danube release - more information: http://www.opnfv.org/danube diff --git a/docs/scenarios/os-odl_l3-fdio-noha/FDS-L3-DVR-example.png b/docs/scenarios/os-odl_l3-fdio-noha/FDS-L3-DVR-example.png deleted file mode 100755 index 18932c3..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-noha/FDS-L3-DVR-example.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-noha/FDS-L3-noha-sample-setup.png b/docs/scenarios/os-odl_l3-fdio-noha/FDS-L3-noha-sample-setup.png deleted file mode 100755 index 27c8335..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-noha/FDS-L3-noha-sample-setup.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-noha/FDS-basic-callflow.jpg b/docs/scenarios/os-odl_l3-fdio-noha/FDS-basic-callflow.jpg deleted file mode 100755 index 96464f8..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-noha/FDS-basic-callflow.jpg and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-noha/FDS-basic-components.jpg b/docs/scenarios/os-odl_l3-fdio-noha/FDS-basic-components.jpg deleted file mode 100755 index e92851f..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-noha/FDS-basic-components.jpg and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-noha/FDS-odl_l3-noha-overview.png b/docs/scenarios/os-odl_l3-fdio-noha/FDS-odl_l3-noha-overview.png deleted file mode 100755 index 1193ea4..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-noha/FDS-odl_l3-noha-overview.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-noha/FDS-odl_l3-overview.png b/docs/scenarios/os-odl_l3-fdio-noha/FDS-odl_l3-overview.png deleted file mode 100755 index 5b90c93..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-noha/FDS-odl_l3-overview.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-noha/FDS-simple-callflow.png b/docs/scenarios/os-odl_l3-fdio-noha/FDS-simple-callflow.png deleted file mode 100755 index 04546aa..0000000 Binary files a/docs/scenarios/os-odl_l3-fdio-noha/FDS-simple-callflow.png and /dev/null differ diff --git a/docs/scenarios/os-odl_l3-fdio-noha/index.rst b/docs/scenarios/os-odl_l3-fdio-noha/index.rst deleted file mode 100644 index e053d57..0000000 --- a/docs/scenarios/os-odl_l3-fdio-noha/index.rst +++ /dev/null @@ -1,20 +0,0 @@ -.. _os-odl_l3-fdio-noha: - -.. OPNFV - Open Platform for Network Function Virtualization -.. This work is licensed under a Creative Commons Attribution 4.0 International License. -.. http://creativecommons.org/licenses/by/4.0 - - -*********************************************************************** -Fast Data Stacks Scenario: os-odl_l3-fdio-noha Overview and Description -*********************************************************************** - -Scenario: "OpenStack - Opendaylight (L3) - FD.io" (apex-os-odl_l3-fdio-noha) -is a scenario developed as part of the FastDataStacks -OPNFV project. - -.. toctree:: - :numbered: - :maxdepth: 2 - - scenario.description.rst diff --git a/docs/scenarios/os-odl_l3-fdio-noha/scenario.description.rst b/docs/scenarios/os-odl_l3-fdio-noha/scenario.description.rst deleted file mode 100755 index 6bbf12b..0000000 --- a/docs/scenarios/os-odl_l3-fdio-noha/scenario.description.rst +++ /dev/null @@ -1,274 +0,0 @@ -.. OPNFV - Open Platform for Network Function Virtualization -.. This work is licensed under a Creative Commons Attribution 4.0 International License. -.. http://creativecommons.org/licenses/by/4.0 - -Scenario: "OpenStack - OpenDaylight (Layer 3) - FD.io" -====================================================== - -Scenario: apex-os-odl_l3-fdio-noha - -"apex-os-odl_l3-fdio-noha" is a scenario developed as part of the -FastDataStacks OPNFV project. The main components of the -"apex-os-odl_l3-fdio-noha" scenario are: - - - APEX (TripleO) installer (please also see APEX installer documentation) - - Openstack (in non-HA configuration) - - OpenDaylight controller (non-clustered) - controlling layer 2 and layer 3 networking - - FD.io/VPP virtual forwarder for tenant networking - -Introduction -============ - -NFV and virtualized high performance applications, such as video processing, -require a "fast data stack" solution that provides both carrier grade -forwarding performance, scalability and open extensibility, along with -functionality for realizing application policies and controlling a complex -network topology. - -A solution stack is only as good as its foundation. Key foundational assets for -NFV infrastructure are - * The virtual forwarder: The virtual forwarder needs to be a feature rich, - high performance, highly scale virtual switch-router. It needs to leverage - hardware accelerators when available and run in user space. - In addition, it should be modular and easily extensible. - * Forwarder diversity: A solution stack should support a variety of - forwarders, hardware forwarders (physical switches and routers) - as well as software forwarders. This way virtual and physical - forwarding domains can be seamlessly glued together. - * Policy driven connectivity: Connectivity should respect and - reflect different business - -In order to meet the desired qualities of an NFV infrastructure, the -following components were chosen for the "Openstack - OpenDaylight - FD.io" -scenario: - * FD.io Vector Packet Processor (VPP) - a highly scalable, - high performance, extensible virtual forwarder - * OpenDaylight Controller - an extensible controller platform which - offers the ability to separate business logic from networking - constructs, supports a diverse set of network devices - (virtual and physical) via the "group based policy (GBP)" - component, and can be clustered to achieve a highly available - deployment. - - -The "Openstack - OpenDaylight - FD.io" scenario provides the capability to -realize a set of use-cases relevant to the deployment of NFV nodes instantiated -by means of an Openstack orchestration system on FD.io/VPP enabled compute -nodes. The role of the Opendaylight network controller in this integration is -twofold. It provides a network device configuration and topology abstraction -via the Openstack Neutron interface, while providing the capability to realize -more complex network policies by means of Group Based Policies. Furthermore it -also provides the capabilities to monitor as well as visualize the operation of -the virtual network devices and their topologies. In supporting the general -use-case of instantiatiting an NFV instance, two specific types of network -transport use cases are realized: - - * NFV instances with VPP data-plane forwarding using a VLAN provider network - * NFV instances with VPP data-plane forwarding using a VXLAN overlay - transport network - -A deployment of the "apex-os-odl_l3-fdio-noha" scenario consists of 4 or more -servers: - - * 1 Jumphost hosting the APEX installer - running the Undercloud - * 1 Controlhost, which runs the Overcloud as well as - OpenDaylight as a network controller - * 2 or more Computehosts. One of the Computehosts also serves as - layer 3 gateway for the tenant networks. - -.. image:: FDS-odl_l3-noha-overview.png - -Tenant and public networking leverages FD.io/VPP. On one of the compute nodes, -VPP binds to both, the tenant networking interface as well as to the public -networking interface. This means that VPP is used for communication within -a tenant network, between tenant networks, as well as between a tenant network -and the Internet. Note that this setup slightly differs from the usual -centralized L3 setup with qrouter on the control node. This setup was chosen -to limit the configuration changes for the introduction of FD.io/VPP. The -OpenDaylight network controller is used to setup and manage layer 2 and -layer 3 networking for the scenario - with Group Based Policy (GBP) being the -key component. Tenant networking can either leverage VXLAN (in which case a -full mesh of VXLAN tunnels is created) or VLANs. - -The picture below shows an example setup with two compute and one control -node. Note that the external network is connected via compute node 0 through -VPP. VPP provides all layer 3 services which are provided in a "vanilla" -OpenStack deployment, including SNAT and DNAT, as well as north-south -and east-west traffic filtering for security purposes ("security groups"). - -.. image:: FDS-L3-noha-sample-setup.png - -Features of the scenario ------------------------- - -Main features of the "apex-os-odl_l3-fdio-noha" scenario: - - * Automated installation using the APEX installer - * Fast and scalable tenant networking using FD.io/VPP as forwarder - * Layer 2 networking using VLANs or VXLAN, managed - and controlled through OpenDaylight - * Layer 3 connectivitiy for tenant networks supplied - through FD.io/VPP. Layer 3 features, including security groups as well as - floating IP addresses (i.e. NAT) are implemented by the FD.io/VPP forwarder - * Manual and automatic (via DHCP) addressing on tenant networks - -Scenario components and composition -=================================== - -The apex-os-odl_l3-fdio-noha scenario combines components from three key open -source projects: OpenStack, OpenDaylight, and Fast Data (FD.io). The key -components that realize the apex-os-odl_l3-fdio-noha scenario and which differ -from a regular, OVS-based scenario, are the OpenStack ML2 OpenDaylight plugin, -OpenDaylight Neutron Northbound, OpenDaylight Group Based Policy, OpenDaylight -Virtual Bridge Domain Manager, FD.io Honeycomb management agent and FD.io -Vector Packet Processor (VPP). - -Note that the key components of the OpenDaylight based scenarios of -FastDataStacks are the same. The Layer 2 scenario "apex-os-odl_l2-fdio-noha" -and the Layer 3 scenario "apex-os-odl_l3-fdio-noha" share the same components. - -Here's a more detailed list of the individual software components involved: - -**Openstack Neutron ML2 OpenDaylight Plugin**: Handles Neutron data base -synchronization and interaction with the southbound controller using a REST -interface. - -**ODL GBP Neutron Mapper**: Maps neutron elements like networks, subnets, -security groups, etc. to GBP entities: Creates policy and configuration for -tenants (endpoints, resolved policies, forwarding rules). - -**ODL GBP Neutron VPP Mapper**: Maps Neutron ports to VPP endpoints in GBP. - -**ODL GBP Location Manager**: Provides real location for endpoints (i.e. Which -physical node an endpoint is connected to). - -**GBP Renderer Manager**: Creates configuration for Renderers (like e.g. -VPP-Renderer or OVS-Renderer). The GBP Renderer Manager is the central point -for dispatching of data to specific device renderers. It uses the information -derived from the GBP end-point and its topology entries to dispatch the task -of configuration to a specific device renderer by writing a renderer policy -configuration into the registered renderer's policy store. The renderer -manager also monitors, by being a data change listener on the VPP Renderer -Policy States, for any errors in the application of a rendered configuration. - -**GBP VPP Renderer Interface Manager**: Listens to VPP endpoints in the -Config DataStore and configures associated interfaces on VPP via HoneyComb. - -**GBP VPP Renderer Renderer Policy Manager**: Manages the creation of -bridge domains using VBD and assigns interfaces to bridge domains. - -**Virtual Bridge Domain Manager (VBD)**: Creates bridge domains (i.e. in case -of VXLAN creates full mesh of VXLAN tunnels, configures split horizon on -tunnel endpoints etc.). VDB configures VXLAN tunnels always into a full-mesh -with split-horizon group forwarding applied on any domain facing tunnel -interface (i.e. forwarding behavior will be that used for VPLS). - -**Virtual Packet Processor (VPP) and Honeycomb server**: The VPP is the -accelerated data plane forwarding engine relying on vhost user interfaces -towards Virtual Machines created by the Nova Agent. The Honeycomb NETCONF -configuration server is responsible for driving the configuration of the VPP, -and collecting the operational data. - -**Nova Agent**: The Nova Agent, a sub-component of the overall Openstack -architecture, is responsible for interacting with the compute node's host -Libvirt API to drive the life-cycle of Virtual Machines. It, along with the -compute node software, are assumed to be capable of supporting vhost user -interfaces. - -The picture below shows the key components. - -.. image:: FDS-basic-components.jpg - -To provide a better understanding how the above mentioned components interact -with each other, the following diagram shows how the example of creating a -vhost-user port on VPP through Openstack Neutron: - -To create or update a port, Neutron will send a request to ODL Neutron -Northbound which contains the UUID, along with the host-id as "vpp" and -vif-type as "vhost-user". The GBP Neutron mapper turns the "Neutron speak" of -"ports" into the generic connectivity model that GroupBasedPolicy uses. -Neutron "ports" become generic "GBP Endpoints" which can be consumed by the -GBP Renderer Manager. The GBP Renderer Manager resolves the policy for the -endpoint, i.e. it determines which communication relationships apply to the -specific endpoint, and hands the resolution to a device specific renderer, -which is the VPP renderer in the given case here. VPP renderer turns the -generic policy into VPP specific configuration. Note that in case the policy -would need to be applied to a different device, e.g. an OpenVSwitch (OVS), -then an "OVS Renderer" would be used. VPP Renderer and the topology manager -("Virtual Bridge Domain" manager - i.e. VBD) cooperate to create the actual -network configuration. VPP Renderer configures the interfaces to the virtual -machines (VM), i.e. the vhost-user interface in the given case here and -attaches them to a bridge domain on VPP. VBD handles the setup of connectivity -between bridge domains on individual VPPs, i.e. it maintains the VXLAN tunnels -in the given case here. Both VPP Renderer as well as VBD communicate with the -device through Netconf/YANG. All compute and control nodes run an instance of -VPP and the VPP-configuration agent "Honeycomb". Honeycomb serves as a -Netconf/YANG server, receives the configuration commands from VBD and VPP -Renderer and drives VPP configuration using VPP's local Java APIs. - -.. image:: FDS-simple-callflow.png - -Scenario Configuration -====================== - -To enable the "apex-os-odl_l3-fdio-noha" scenario check the appropriate -settings in the APEX configuration files. Those are typically found in -/etc/opnfv-apex. - -File "deploy_settings.yaml": Choose Opendaylight as controller with version -"carbon" and enable vpp as forwarder. "odl_routing_node" chooses the node -which is used as the layer 3 gateway for the tenant networks. -"odl_routing_node" is an optional parameter. If omitted, the VPP on the first -compute node will serve as layer 3 gateway:: - - deploy_options: - sdn_controller: opendaylight - sdn_l3: true - odl_version: carbon - odl_routing_node: overcloud-novacompute-0 - tacker: true - congress: true - sfc: false - vpn: false - vpp: true - dataplane: fdio - performance: - Controller: - kernel: - hugepages: 1024 - hugepagesz: 2M - intel_iommu: 'on' - iommu: pt - isolcpus: 1,2 - vpp: - main-core: 1 - corelist-workers: 2 - uio-driver: uio_pci_generic - Compute: - kernel: - hugepagesz: 2M - hugepages: 2048 - intel_iommu: 'on' - iommu: pt - isolcpus: 1,2 - vpp: - main-core: 1 - corelist-workers: 2 - uio-driver: uio_pci_generic - -Limitations, Issues and Workarounds -=================================== - -For specific information on limitations and issues, please refer to the APEX -installer release notes. - -References -========== - - - * FastDataStacks OPNFV project wiki: https://wiki.opnfv.org/display/fds - * Fast Data (FD.io): https://fd.io/ - * FD.io Vector Packet Processor (VPP): https://wiki.fd.io/view/VPP - * OpenDaylight Controller: https://www.opendaylight.org/ - * OPNFV Danube release - more information: http://www.opnfv.org/danube -- cgit 1.2.3-korg