diff options
-rw-r--r-- | docs/release/release-notes/Auto-release-notes.rst | 173 | ||||
-rw-r--r-- | docs/release/release-notes/auto-proj-parameters.png | bin | 0 -> 32716 bytes | |||
-rw-r--r-- | docs/release/release-notes/auto-project-activities.png | bin | 58789 -> 25995 bytes |
3 files changed, 121 insertions, 52 deletions
diff --git a/docs/release/release-notes/Auto-release-notes.rst b/docs/release/release-notes/Auto-release-notes.rst index 26445b6..c30bd4c 100644 --- a/docs/release/release-notes/Auto-release-notes.rst +++ b/docs/release/release-notes/Auto-release-notes.rst @@ -7,13 +7,13 @@ Auto Release Notes ================== -This document provides the release notes for the Fraser release of Auto. +This document provides the release notes for the Gambia 7.0 release of Auto. Important notes for this release ================================ -The initial release for Auto was in Fraser 6.0 (project inception: July 2017). This is the second point release, in Fraser 6.2. +The initial release for Auto was in Fraser 6.0 (project inception: July 2017). Summary @@ -23,14 +23,14 @@ Overview ^^^^^^^^ OPNFV is an SDNFV system integration project for open-source components, which so far have been mostly limited to -the NFVI+VIM as generally described by ETSI. +the NFVI+VIM as generally described by `ETSI <https://www.etsi.org/technologies-clusters/technologies/nfv>`_. In particular, OPNFV has yet to integrate higher-level automation features for VNFs and end-to-end Services. -As an OPNFV project, Auto ("ONAP-Automated OPNFV") will focus on ONAP component integration and verification with +As an OPNFV project, Auto (*ONAP-Automated OPNFV*) will focus on ONAP component integration and verification with OPNFV reference platforms/scenarios, through primarily a post-install process, in order to avoid impact to OPNFV -installer projects. As much as possible, this will use a generic installation/integration process (not specific to -any OPNFV installer's technology). +installer projects (Fuel/MCP, Compass4NFV, Apex/TripleO, Daisy4NFV). As much as possible, this will use a generic +installation/integration process (not specific to any OPNFV installer's technology). * `ONAP <https://www.onap.org/>`_ (a Linux Foundation Project) is an open source software platform that delivers robust capabilities for the design, creation, orchestration, monitoring, and life cycle management of @@ -38,7 +38,15 @@ any OPNFV installer's technology). Auto aims at validating the business value of ONAP in general, but especially within an OPNFV infrastructure (integration of ONAP and OPNFV). Business value is measured in terms of improved service quality (performance, -reliability, ...) and OPEX reduction (VNF management simplification, power consumption reduction, ...). +reliability, ...) and OPEX reduction (VNF management simplification, power consumption reduction, ...), as +demonstrated by use cases. + +Auto also validates multi-architecture software (binary images and containers) availability of ONAP and OPNFV: +CPUs (x86, ARM) and Clouds (MultiVIM) + +In other words, Auto is a turnkey approach to automatically deploy an integrated open-source virtual network +based on OPNFV (as infrastructure) and ONAP (as end-to-end service manager), that demonstrates business value +to end-users (IT/Telco service providers, enterprises). While all of ONAP is in scope, as it proceeds, the Auto project will focus on specific aspects of this integration @@ -48,7 +56,7 @@ and verification in each release. Some example topics and work items include: * How ONAP SDN-C uses OPNFV existing features, e.g. NetReady, in a two-layer controller architecture in which the upper layer (global controller) is replaceable, and the lower layer can use different vendor’s local controller to interact with SDN-C. For interaction with multiple cloud infrastructures, the MultiVIM ONAP component will be used. -* How ONAP leverages OPNFV installers (Fuel/MCP, Compass4NFV, Apex/TripleO, Daisy4NFV, JOID) to provide a cloud +* How ONAP leverages OPNFV installers (Fuel/MCP, Compass4NFV, Apex/TripleO, Daisy4NFV) to provide a cloud instance (starting with OpenStack) on which to install the tool ONAP * What data collection interface VNF and controllers provide to ONAP DCAE, and (through DCAE), to closed-loop control functions such as Policy Tests which verify interoperability of ONAP automation/lifecycle features with specific NFVI @@ -72,7 +80,7 @@ It is understood that: .. image:: auto-proj-rn01.png -The current ONAP architecture overview can be found `here <http://onap.readthedocs.io/en/latest/guides/onap-developer/architecture/onap-architecture.html>`_. +The current ONAP architecture overview can be found `here <https://onap.readthedocs.io/en/latest/guides/onap-developer/architecture/onap-architecture.html>`_. For reference, the ONAP-Beijing architecture diagram is replicated here: @@ -89,17 +97,18 @@ Within OPNFV, Auto leverages tools and collaborates with other projects: * FuncTest for software verification (CI/CD, Pass/Fail) * Yardstick for metric management (quantitative measurements) * VES (VNF Event Stream) and Barometer for VNF monitoring (feed to ONAP/DCAE) + * Edge Cloud as use case * leverage OPNFV tools and infrastructure: * Pharos as LaaS: transient pods (3-week bookings) and permanent Arm pod (6 servers) - * possibly other labs from the community + * `WorksOnArm <http://worksonarm.com/cluster>`_ (`GitHub link <http://github.com/worksonarm/cluster>`_) + * possibly other labs from the community (Huawei pod-12, 6 servers, x86) * JJB/Jenkins for CI/CD (and follow OPNFV scenario convention) * Gerrit/Git for code and documents reviewing and archiving (similar to ONAP: Linux Foundation umbrella) * follow OPNFV releases (Releng group) - Testability ^^^^^^^^^^^ @@ -123,21 +132,26 @@ value (i.e., find/determine policies and controls which yield optimized ONAP bus More precisely, the following list shows parameters that could be applied to an Auto full run of test cases: * Auto test cases for given use cases -* OPNFV installer {Fuel/MCP, Compass4NFV, Apex/TripleO, Daisy4NFV, JOID} +* OPNFV installer {Fuel/MCP, Compass4NFV, Apex/TripleO, Daisy4NFV} * OPNFV availability scenario {HA, noHA} -* cloud where ONAP runs {OpenStack, AWS, GCP, Azure, ...} -* ONAP installation type {bare metal or virtual server, VM or container, ...} and options {MultiVIM single|distributed, ...} -* VNFs {vFW, vCPE, vAAA, vDHCP, vDNS, vHSS, ...} and VNF-based services {vIMS, vEPC, ...} +* environment where ONAP runs {bare metal servers, VMs from clouds (OpenStack, AWS, GCP, Azure, ...), containers} +* ONAP installation type {bare metal, VM, or container, ...} and options {MultiVIM single|distributed, ...} +* VNF types {vFW, vCPE, vAAA, vDHCP, vDNS, vHSS, ...} and VNF-based services {vIMS, vEPC, ...} * cloud where VNFs run {OpenStack, AWS, GCP, Azure, ...} -* VNF type {VM-based, container} -* CPU architectures {x86/AMD64, ARM/aarch64} for ONAP software and for VNFs +* VNF host type {VM, container} +* CPU architectures {x86/AMD64, ARM/aarch64} for ONAP software and for VNF software; not really important for Auto software; * pod size and technology (RAM, storage, CPU cores/threads, NICs) -* traffic types and amounts/volumes +* traffic types and amounts/volumes; traffic generators (although that should not really matter); * ONAP configuration {especially policies and closed-loop controls; monitoring types for DCAE: VES, ...} * versions of every component {Linux OS (Ubuntu, CentOS), OPNFV release, clouds, ONAP, VNFs, ...} +The diagram below shows Auto parameters: + +.. image:: auto-proj-parameters.png -Illustration of Auto analysis loop based on test case executions: + +The next figure is an illustration of the Auto analysis loop (design, configuration, execution, result analysis) +based on test cases covering as many parameters as possible : .. image:: auto-proj-tests.png @@ -150,14 +164,14 @@ Auto currently defines three use cases: Edge Cloud (UC1), Resiliency Improvement including end-to-end composite services of which a Cloud Manager may not be aware (VMs or containers could be recovered by a Cloud Manager, but not necessarily an end-to-end service built on top of VMs or containers). * enterprise-grade performance of vCPEs (certification during onboarding, then real-time performance assurance with - SLAs and HA as well as scaling). + SLAs and HA, as well as scaling). The use cases define test cases, which initially will be independent, but which might eventually be integrated to `FuncTest <https://wiki.opnfv.org/display/functest/Opnfv+Functional+Testing>`_. -Additional use cases can be added in the future, such as vIMS (example: project Clearwater) or residential vHGW (virtual -Home Gateways). The interest for vHGW is to reduce overall power consumption: even in idle mode, physical HGWs in -residential premises consume a lot of energy. Virtualizing that service to the Service Provider edge data center would -allow to minimize that consumption. +Additional use cases can be added in the future, such as vIMS (example: project `Clearwater <http://www.projectclearwater.org/>`_) +or residential vHGW (virtual Home Gateways). The interest for vHGW is to reduce overall power consumption: even in idle mode, +physical HGWs in residential premises consume a lot of energy. Virtualizing that service to the Service Provider edge data center +would allow to minimize that consumption. Lab environment @@ -172,39 +186,53 @@ x86 pod at UNH IOL. A transition is in progress, to leverage OPNFV LaaS (Lab-as-a-Service) pods (`Pharos <https://labs.opnfv.org/>`_). These pods can be booked for 3 weeks only (with an extension for a maximum of 2 weeks), so they are not a permanent resource. -A repeatable automated installation procedure is being developed. +For ONAP-Beijing, a repeatable automated installation procedure is being developed, using 3 Pharos servers (x86 for now). +Also, a more permanent ONAP installation is in progress at a Huawei lab (pod-12, consisting of 6 x86 servers, +1 as jump server, the other 5 with this example allocation: 3 for ONAP components, and 2 for an OPNFV infratructure: +Openstack installed by Compass4NFV). ONAP-based onboarding and deployment of VNFs is in progress (ONAP-Amsterdam pre-loading of VNFs must still done outside of ONAP: for VM-based VNFs, users need to prepare OpenStack stacks (using Heat templates), then make an instance snapshot which serves as the binary image of the VNF). -An initial version of a script to prepare an OpenStack instance for ONAP (creation of a public and a private network, -with a router) has been developed. It leverages OpenStack SDK. +A script to prepare an OpenStack instance for ONAP (creation of a public and a private network, with a router, +pre-loading of images and flavors, creation of a security group and an ONAP user) has been developed. It leverages +OpenStack SDK. It has a delete option, so it can be invoked to delete these objects for example in a tear-down procedure. Integration with Arm servers has started (exploring binary compatibility): -* OpenStack is currently installed on a 6-server pod of Arm servers +* The Auto project has a specific 6-server pod of Arm servers, which is currently loaned to ONAP integration team, + to build ONAP images * A set of 14 additional Arm servers was deployed at UNH, for increased capacity -* Arm-compatible Docker images are in the process of being developed +* ONAP Docker registry: ONAP-specific images for ARM are being built, with the purpose of populating ONAP nexus2 + (Maven2 artifacts) and nexus3 (Docker containers) repositories at Linux Foundation. Docker images are + multi-architecture, and the manifest of an image may contain 1 or more layers (for example 2 layers: x86/AMD64 + and ARM/aarch64). One of ONAP-Casablanca architectural requirements is to be CPU-architecture independent. + There are almost 150 Docker containers in a complete ONAP instance. Currently, more disk space is being added + to the ARM nodes (configuration of Nova, and/or additional actual physical storage space). + -Test case implementation for the three use cases has started. +Test case design and implementation for the three use cases has started. OPNFV CI/CD integration with JJD (Jenkins Job Description) has started: see the Auto plan description -`here <https://wiki.opnfv.org/display/AUTO/CI+Plan+for+Auto>`_. The permanent resource for that is the 6-server Arm +`here <https://wiki.opnfv.org/display/AUTO/CI+for+Auto>`_. The permanent resource for that is the 6-server Arm pod, hosted at UNH. The CI directory from the Auto repository is `here <https://git.opnfv.org/auto/tree/ci>`_ + Finally, the following figure illustrates Auto in terms of project activities: .. image:: auto-project-activities.png -Note: a demo was delivered at the OpenStack Summit in Vancouver on May 21st 2018, to illustrate the deployment of a WordPress application -(WordPress is a platform for websites and blogs) deployed on a multi-architecture cloud (mix of x86 and Arm servers). +Note: a demo was delivered at the OpenStack Summit in Vancouver on May 21st 2018, to illustrate the deployment of +a WordPress application (WordPress is a platform for websites and blogs) deployed on a multi-architecture cloud (mix +of x86 and Arm servers). This shows how service providers and enterprises can diversify their data centers with servers of different architectures, -and select architectures best suited to each use case (mapping application components to architectures: DBs, interactive servers, -number-crunching modules, ...). +and select architectures best suited to each use case (mapping application components to architectures: DBs, +interactive servers, number-crunching modules, ...). This prefigures how other examples such as ONAP, VIMs, and VNFs could also be deployed on heterogeneous multi-architecture -environments (open infrastructure), orchestrated by Kubernetes. The Auto installation scripts could expand on that approach. +environments (open infrastructure), orchestrated by Kubernetes. The Auto installation scripts covering all the parameters +described above could expand on that approach. .. image:: auto-proj-openstacksummit1805.png @@ -218,13 +246,13 @@ Release Data | **Project** | Auto | | | | +--------------------------------------+--------------------------------------+ -| **Repo/commit-ID** | auto/opnfv-6.2.0 | +| **Repo/commit-ID** | auto/opnfv-7.0.0 | | | | +--------------------------------------+--------------------------------------+ -| **Release designation** | Fraser 6.2 | +| **Release designation** | Gambia 7.0 | | | | +--------------------------------------+--------------------------------------+ -| **Release date** | 2018-06-29 | +| **Release date** | 2018-11-02 | | | | +--------------------------------------+--------------------------------------+ | **Purpose of the delivery** | Official OPNFV release | @@ -273,22 +301,66 @@ Point release 6.2: * initial scripts for OPNFV CI/CD, registration of Jenkins slave on `Arm pod <https://build.opnfv.org/ci/view/auto/>`_ * updated script for configuring OpenStack instance for ONAP, using OpenStack SDK 0.14 -Notable activities since release 6.1, which may result in new features for Gambia 7.0: +Point release 7.0: + +* progress on Docker registry of ONAP's Arm images +* progress on ONAP installation script for 3-server cluster of UNH servers +* CI scripts for OPNFV installers: Fuel/MCP (x86), Compass, Apex/TripleO (must run twice) +* initial CI script for Daisy4NFV (work in progress) +* JOID script, but supported only until R6.2, not Gambia 7.0 +* completed script for configuring OpenStack instance for ONAP, using OpenStack SDK 0.17 +* use of an additional lab resource for Auto development: 6-server x86 pod (huawei-pod12) + + -* researching how to configure multiple Pharos servers in a cluster for Kubernetes -* started to evaluate Compass4nfv as another OpenStack installer; issues with Python version (2 or 3) -* common meeting with Functest -* Plugfest: initiated collaboration with ONAP/MultiVIM (including support for ONAP installation) **JIRA TICKETS for this release:** + +`JIRA Auto Gambia 7.0.0 Done <https://jira.opnfv.org/issues/?filter=12403>`_ + +Manual selection of significant JIRA tickets for this version's highlights: + +--------------------------------------+--------------------------------------+ | **JIRA REFERENCE** | **SLOGAN** | | | | +--------------------------------------+--------------------------------------+ -| AUTO-38, auto-resiliency-vif-001: | UC2: validate VM suspension command | -| 2/3 Test Logic | and measurement of Recovery Time | +| AUTO-37 | Get DCAE running onto Pharos | +| | deployment | ++--------------------------------------+--------------------------------------+ +| AUTO-42 | Use Compass4NFV to create an | +| | OpenStack instance on a UNH pod | ++--------------------------------------+--------------------------------------+ +| AUTO-43 | String together scripts for Fuel, | +| | Tool installation, ONAP preparation | ++--------------------------------------+--------------------------------------+ +| AUTO-44 | Build ONAP components for arm64 | +| | platform | ++--------------------------------------+--------------------------------------+ +| AUTO-45 | CI: Jenkins definition of verify and | +| | merge jobs | ++--------------------------------------+--------------------------------------+ +| AUTO-46 | Use Apex to create an OpenStack | +| | instance on a UNH pod | ++--------------------------------------+--------------------------------------+ +| AUTO-47 | Install ONAP with Kubernetes on LaaS | +| | | ++--------------------------------------+--------------------------------------+ +| AUTO-48 | Create documentation for ONAP | +| | deployment with Kubernetes on LaaS | ++--------------------------------------+--------------------------------------+ +| AUTO-49 | Automate ONAP deployment with | +| | Kubernetes on LaaS | ++--------------------------------------+--------------------------------------+ +| AUTO-51 | huawei-pod12: Prepare IDF and PDF | +| | files | ++--------------------------------------+--------------------------------------+ +| AUTO-52 | Deploy a running ONAP instance on | +| | huawei-pod12 | ++--------------------------------------+--------------------------------------+ +| AUTO-54 | Use Daisy4nfv to create an OpenStack | +| | instance on a UNH pod | +--------------------------------------+--------------------------------------+ | | | | | | @@ -319,7 +391,7 @@ Deliverables Software deliverables ^^^^^^^^^^^^^^^^^^^^^ -6.2 release: in-progress install scripts, CI scripts, and test case implementations. +7.0 release: in-progress Docker ARM images, install scripts, CI scripts, and test case implementations. Documentation deliverables @@ -341,9 +413,6 @@ Known Limitations, Issues and Workarounds System Limitations ^^^^^^^^^^^^^^^^^^ -* ONAP still to be validated for Arm servers (many Docker images are ready) -* ONAP installation still to be automated in a repeatable way, and need to configure cluster of Pharos servers - Known issues @@ -393,8 +462,8 @@ None at this point. References ========== -For more information on the OPNFV Fraser release, please see: -http://opnfv.org/fraser +For more information on the OPNFV Gambia release, please see: +http://opnfv.org/gambia Auto Wiki pages: diff --git a/docs/release/release-notes/auto-proj-parameters.png b/docs/release/release-notes/auto-proj-parameters.png Binary files differnew file mode 100644 index 0000000..a0cbe2e --- /dev/null +++ b/docs/release/release-notes/auto-proj-parameters.png diff --git a/docs/release/release-notes/auto-project-activities.png b/docs/release/release-notes/auto-project-activities.png Binary files differindex c50bd72..d25ac2a 100644 --- a/docs/release/release-notes/auto-project-activities.png +++ b/docs/release/release-notes/auto-project-activities.png |