summaryrefslogtreecommitdiffstats
path: root/docs
diff options
context:
space:
mode:
Diffstat (limited to 'docs')
-rw-r--r--docs/conf.py15
-rw-r--r--docs/conf.yaml12
-rw-r--r--docs/images/dovetail_generic_framework.pngbin0 -> 92053 bytes
-rw-r--r--docs/index.rst28
-rw-r--r--docs/release/release-notes/dovetail-release.rst338
-rw-r--r--docs/release/release-notes/index.rst283
-rw-r--r--docs/requirements.txt5
-rw-r--r--docs/testing/developer/genericframework/index.rst413
-rw-r--r--docs/testing/developer/testcaserequirements/index.rst37
-rw-r--r--docs/testing/developer/testscope/index.rst616
-rw-r--r--docs/testing/user/certificationworkflow/ApplicationForm.rst6
-rw-r--r--docs/testing/user/certificationworkflow/index.rst122
-rw-r--r--docs/testing/user/instructionsperinstaller/index.rst14
-rw-r--r--docs/testing/user/instructionsperinstaller/testing_instructions_apex.rst158
-rw-r--r--docs/testing/user/instructionsperinstaller/testing_instructions_xci.rst220
-rw-r--r--docs/testing/user/ovpaddendum/exemption-strict-API-validation.rst185
-rw-r--r--docs/testing/user/ovpaddendum/index.rst474
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_log_files.pngbin6504 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_log_setup.pngbin37346 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_log_test_count.pngbin6930 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_missing_ha.pngbin80192 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_pass_fraction.pngbin117383 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_pass_percentage.pngbin75765 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_result_overview.pngbin105270 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_result_review.pngbin44980 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_run_results.pngbin13149 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_test_count.pngbin4012 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/ovp_top_nav.pngbin23642 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/sut_endpoints.pngbin119897 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/danube/images/sut_info.pngbin29227 -> 0 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/ovp_log_files_functest_image.pngbin0 -> 39227 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/ovp_log_setup.pngbin0 -> 12385 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/ovp_pass_fraction.pngbin0 -> 36240 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/ovp_pass_percentage.pngbin0 -> 8757 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/ovp_result_review.pngbin0 -> 13652 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/ovp_top_nav.pngbin0 -> 20489 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/ovp_vping_ssh.png (renamed from docs/testing/user/reviewerguide/danube/images/ovp_vping_ssh.png)bin5246 -> 5246 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/ovp_vping_user.png (renamed from docs/testing/user/reviewerguide/danube/images/ovp_vping_user.png)bin5667 -> 5667 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/review_status.pngbin0 -> 9887 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/sut_endpoints.pngbin0 -> 74700 bytes
-rw-r--r--docs/testing/user/reviewerguide/images/sut_info.pngbin0 -> 12822 bytes
-rw-r--r--docs/testing/user/reviewerguide/index.rst271
-rw-r--r--docs/testing/user/systempreparation/index.rst8
-rw-r--r--docs/testing/user/testspecification/faultmanagement/index.rst0
-rw-r--r--docs/testing/user/testspecification/forwardingpackets/index.rst147
-rw-r--r--docs/testing/user/testspecification/highavailability/index.rst396
-rw-r--r--docs/testing/user/testspecification/index.rst45
-rw-r--r--docs/testing/user/testspecification/security_patrole/index.rst327
-rw-r--r--docs/testing/user/testspecification/security_patrole_vxlan_dependent/index.rst51
-rw-r--r--docs/testing/user/testspecification/stress/index.rst159
-rw-r--r--docs/testing/user/testspecification/tempest_compute/index.rst132
-rw-r--r--docs/testing/user/testspecification/tempest_identity_v3/index.rst83
-rw-r--r--docs/testing/user/testspecification/tempest_image/index.rst42
-rw-r--r--docs/testing/user/testspecification/tempest_ipv6/index.rst140
-rw-r--r--docs/testing/user/testspecification/tempest_ipv6/ipv6_api.rst (renamed from docs/testing/user/testspecification/ipv6/index.rst)828
-rw-r--r--docs/testing/user/testspecification/tempest_ipv6/ipv6_scenario.rst624
-rw-r--r--docs/testing/user/testspecification/tempest_multi_node_scheduling/index.rst (renamed from docs/testing/user/testspecification/multiplenodes/index.rst)31
-rw-r--r--docs/testing/user/testspecification/tempest_network_api/index.rst130
-rw-r--r--docs/testing/user/testspecification/tempest_network_scenario/index.rst (renamed from docs/testing/user/testspecification/dynamicnetwork/index.rst)35
-rw-r--r--docs/testing/user/testspecification/tempest_network_security/index.rst (renamed from docs/testing/user/testspecification/securitygroup/index.rst)27
-rw-r--r--docs/testing/user/testspecification/tempest_osinterop/index.rst38
-rw-r--r--docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_compute.rst (renamed from docs/testing/user/testspecification/vimoperationscompute/index.rst)322
-rw-r--r--docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_identity.rst (renamed from docs/testing/user/testspecification/vimoperationsidentity/index.rst)90
-rw-r--r--docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_image.rst (renamed from docs/testing/user/testspecification/vimoperationsimage/index.rst)63
-rw-r--r--docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_network.rst (renamed from docs/testing/user/testspecification/vimoperationsnetwork/index.rst)300
-rw-r--r--docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_volume.rst (renamed from docs/testing/user/testspecification/vimoperationsvolume/index.rst)364
-rw-r--r--docs/testing/user/testspecification/tempest_trunk_ports/index.rst130
-rw-r--r--docs/testing/user/testspecification/tempest_vm_lifecycle/index.rst (renamed from docs/testing/user/testspecification/machinelifecycle/index.rst)47
-rw-r--r--docs/testing/user/testspecification/tempest_volume/index.rst128
-rw-r--r--docs/testing/user/testspecification/vnf/index.rst222
-rw-r--r--docs/testing/user/testspecification/vping/index.rst139
-rw-r--r--docs/testing/user/testspecification/vpn/index.rst524
-rw-r--r--docs/testing/user/userguide/api_testing_guide.rst373
-rw-r--r--docs/testing/user/userguide/cli_reference.rst183
-rw-r--r--docs/testing/user/userguide/images/tocsa_vnf_test_environment.pngbin0 -> 101795 bytes
-rw-r--r--docs/testing/user/userguide/images/tosca_vnf_test_flow.pngbin0 -> 40614 bytes
-rw-r--r--docs/testing/user/userguide/index.rst8
-rw-r--r--docs/testing/user/userguide/testing_guide.rst600
-rw-r--r--docs/testing/user/userguide/vnf_test_guide.rst714
79 files changed, 6980 insertions, 3667 deletions
diff --git a/docs/conf.py b/docs/conf.py
new file mode 100644
index 00000000..a3d9ebb7
--- /dev/null
+++ b/docs/conf.py
@@ -0,0 +1,15 @@
+##############################################################################
+# Copyright (c) 2019 opnfv.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+
+from docs_conf.conf import *
+
+linkcheck_ignore = [
+ 'https://github.com/opnfv/functest/blob/master/docs/testing/user/userguide/test_details.rst#',
+ 'https://review.openstack.org/#'
+]
diff --git a/docs/conf.yaml b/docs/conf.yaml
new file mode 100644
index 00000000..fa2e75d7
--- /dev/null
+++ b/docs/conf.yaml
@@ -0,0 +1,12 @@
+##############################################################################
+# Copyright (c) 2019 opnfv.
+#
+# All rights reserved. This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+# http://www.apache.org/licenses/LICENSE-2.0
+##############################################################################
+
+---
+project_cfg: opnfv
+project: dovetail
diff --git a/docs/images/dovetail_generic_framework.png b/docs/images/dovetail_generic_framework.png
new file mode 100644
index 00000000..b73a0678
--- /dev/null
+++ b/docs/images/dovetail_generic_framework.png
Binary files differ
diff --git a/docs/index.rst b/docs/index.rst
new file mode 100644
index 00000000..5b83a60c
--- /dev/null
+++ b/docs/index.rst
@@ -0,0 +1,28 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. SPDX-License-Identifier: CC-BY-4.0
+.. (c) Open Platform for NFV Project, Inc. and its contributors
+
+.. _<project-name>: dovetail
+
+========
+Dovetail
+========
+
+.. toctree::
+ :numbered:
+ :maxdepth: 1
+
+ testing/user/certificationworkflow/ApplicationForm
+ testing/user/certificationworkflow/index
+ testing/user/instructionsperinstaller/index
+ testing/user/ovpaddendum/exemption-strict-API-validation
+ testing/user/ovpaddendum/index
+ testing/user/reviewerguide/index
+ testing/user/systempreparation/index
+ testing/user/testspecification/index
+ testing/user/userguide/index
+
+ testing/developer/testcaserequirements/index
+ testing/developer/genericframework/index
+
+ release/release-notes/index
diff --git a/docs/release/release-notes/dovetail-release.rst b/docs/release/release-notes/dovetail-release.rst
deleted file mode 100644
index e70e2e61..00000000
--- a/docs/release/release-notes/dovetail-release.rst
+++ /dev/null
@@ -1,338 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. SPDX-License-Identifier: CC-BY-4.0
-
-=======
-License
-=======
-
-OPNFV Danube release note for Dovetail Docs
-are licensed under a Creative Commons Attribution 4.0 International License.
-You should have received a copy of the license along with this.
-If not, see <http://creativecommons.org/licenses/by/4.0/>.
-
-==================================================================
-OPNFV Verified Program (OVP) 2018.01 / Dovetail 1.0.0 Release Note
-==================================================================
-
-Abstract
-========
-
-This document describes the release note of the OPNFV Verified Program and the Dovetail project.
-
-
-Version History
-===============
-
-+------------+----------+--------------------------+
-| **Date** | **Ver.** | **Comment** |
-| | | |
-+------------+----------+--------------------------+
-| 2018-01-21 | 1.0.0 | Dovetail for OVP 2018.01 |
-| | | Danube release |
-+------------+----------+--------------------------+
-
-OPNFV Danube Release
-====================
-
-The OPNFV Verified Program (OVP) allows vendors and operators to obtain 'OPNFV Verified'
-status based on an agreed upon set of compliance verification test cases that align to OPNFV
-releases. The reference System under Test (SUT) are the NFV components deployed by the OPNFV
-installers for a given release, where OVP 2018.01 is based on the Danube release. Participants
-of the program can verify commercial or open source offerings against an OVP release. This implies
-that the SUT used for verification has interfaces, components, functions and behaviors that align
-to OPNFV installer integrations.
-
-Dovetail is the overall framework used to execute tests and collect results for OVP. Dovetail does
-not deliver test content directly. Rather, test content is developed in other OPNFV test frameworks
-such as Functest and upstream test communities such as OpenStack's RefStack/Tempest projects.
-Dovetail leverages this upstream test content and provides a common set of test platform services
-for the OVP.
-
-Dovetail works in conjunction with a web portal interface dubbed the 'OVP web portal' to allow
-users to upload test results to a centralized community repository. This facilitates user
-collaboration, result sharing, self-testing and community reviews. It also serves as a hub for
-new participants to learn about the program and access key resources. The link for this portal
-is at:
-
- * https://verified.opnfv.org
-
-Use of the OVP web portal is open to all and only requires a valid Linux Foundation or OpenStack
-ID to login. Users are welcome to use the portal to upload, inspect and share results in a private
-manner. In order to submit results for official review, the first step is apply for acceptance
-into the program with the participation form provided in the link:
-
- * https://na3.docusign.net/Member/PowerFormSigning.aspx?PowerFormId=e03e78d7-c32e-47d3-8292-350b747a0105
-
-Test Suites & Test Areas
-------------------------
-
-OVP/Dovetail groups test cases into test suites and test areas. Test suites are currently a basic
-categorization around releases for the most part. Executing the test suite 'ovp.1.0.0' without
-further specification will run all the test cases in the OVP 2018.01 release. Test suites are
-divided into test areas that can be executed separately.
-
-Test areas include a division into 'mandatory' and 'optional' in an overarching categorization.
-The mandatory test area is further subdivided into the following three test areas, which can
-be executed independently:
-
- * osinterop
- * vping
- * ha
-
-The optional test area is further subdivided into the following three test areas:
-
- * ipv6
- * tempest
- * sdnvpn
-
-All the mandatory test areas are required to be executed with passing results for all inclusive
-test cases for results to be reviewed and approved by the community made up of peer reviewers.
-The optional test areas are not required to be executed for the official compliance verification
-review in the OVP 2018.01 release. However, execution of these areas is encouraged, as some
-optional test areas may become mandatory in future releases.
-
-Test Cases and Sub Test Cases
-^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
-
-Each test area consists of multiple test cases where each test case can be a single test or
-broken down into sub test cases. A listing of test cases with the number of sub test cases noted
-in parenthesis is shown below for the OVP 2018.01 release.
-
-**Mandatory**
- * dovetail.osinterop.tc001 (205)
- * dovetail.vping.tc001 (1)
- * dovetail.vping.tc002 (1)
- * dovetail.ha.tc001 (1)
- * dovetail.ha.tc002 (1)
- * dovetail.ha.tc003 (1)
- * dovetail.ha.tc004 (1)
- * dovetail.ha.tc005 (1)
- * dovetail.ha.tc006 (1)
- * dovetail.ha.tc007 (1)
- * dovetail.ha.tc008 (1)
-
-There are a total of 215 mandatory test cases (osinterop: 205, vping: 2, ha: 8).
-
-**Optional**
- * dovetail.ipv6.tc001 (3)
- * dovetail.ipv6.tc002 (1)
- * dovetail.ipv6.tc003 (1)
- * dovetail.ipv6.tc004 (2)
- * dovetail.ipv6.tc005 (2)
- * dovetail.ipv6.tc006 (1)
- * dovetail.ipv6.tc007 (1)
- * dovetail.ipv6.tc008 (1)
- * dovetail.ipv6.tc009 (1)
- * dovetail.ipv6.tc010 (1)
- * dovetail.ipv6.tc011 (1)
- * dovetail.ipv6.tc012 (1)
- * dovetail.ipv6.tc013 (1)
- * dovetail.ipv6.tc014 (1)
- * dovetail.ipv6.tc015 (1)
- * dovetail.ipv6.tc016 (1)
- * dovetail.ipv6.tc017 (1)
- * dovetail.ipv6.tc018 (1)
- * dovetail.ipv6.tc019 (1)
- * dovetail.ipv6.tc020 (1)
- * dovetail.ipv6.tc021 (1)
- * dovetail.ipv6.tc022 (1)
- * dovetail.ipv6.tc023 (1)
- * dovetail.ipv6.tc024 (1)
- * dovetail.ipv6.tc025 (1)
- * dovetail.tempest.tc001 (1)
- * dovetail.tempest.tc002 (6)
- * dovetail.tempest.tc003 (5)
- * dovetail.tempest.tc004 (12)
- * dovetail.tempest.tc005 (6)
- * dovetail.sdnvpn.tc001 (1)
- * dovetail.sdnvpn.tc002 (1)
- * dovetail.sdnvpn.tc004 (1)
- * dovetail.sdnvpn.tc008 (1)
-
-There are a total of 63 optional test cases (ipv6: 29, tempest: 30, sdnvpn: 4).
-
-Further details on test area breakdown with expanded test and sub test case names is available at:
-
- * http://docs.opnfv.org/en/stable-danube/submodules/dovetail/docs/testing/developer/testscope/index.html
-
-OPNFV Test Projects and Components
-----------------------------------
-
-The OPNFV test frameworks integrated into the Dovetail framework that deliver test content are:
-
- * Functest (leverages OpenStack RefStack/Tempest projects in addition to supplying native test cases)
- * Yardstick
-
-Other upstream components integrated into the Dovetail framework are:
-
- * TestAPI
- * MongoDB
-
-The above components are part of the OPNFV test collection framework. Further information on how
-this framework is used to collect test results can be found at:
-
- * http://docs.opnfv.org/en/stable-danube/submodules/functest/docs/testing/developer/devguide/#test-collection-framework
-
-The test frameworks and components above are packaged as Docker containers for Dovetail to employ.
-Dovetail creates OVP-specific containers for Functest and TestAPI, while it uses the default
-Yardstick Danube container. Additionally, a generic container version of MongoDB is used.
-Installation instructions for Dovetail and its dependent containers can be found in the user
-guide at:
-
- * http://docs.opnfv.org/en/stable-danube/submodules/dovetail/docs/testing/user/userguide/testing_guide.html
-
-Acceptence and Marketing
-------------------------
-
-Upon successful community review of results for OVP 2018.01, the OPNFV C&C Committee on behalf of
-the Board of Directors can award a product 'OPNFV Verified' status. Use of 'OPNFV Verified'
-Program Marks shall be awarded to the platform used for compliance verification. The category label
-of 'Infrastructure' is used within the Program Marks logo and limits the scope of this OVP release
-to a SUT consisting of NFVI and VIM components using ETSI terminology. It does not provide
-compliance verification for specific VNFs in any fashion. The date '2018.01' corresponds to a
-reference SUT that aligns to the OPNFV Danube release and currently aligns to the Dovetail
-framework version 1.0.0.
-
-Organizations shall not use the Program Marks in any way that would associate it with any
-individual or company logo or brand, beyond the association to the specific platform to which it
-was awarded. While OpenStack RefStack interoperability and Tempest integration test cases are
-executed as part of the OVP 2018.01 compliance verification test suites, the OVP does not grant or
-award OpenStack Marks in any fashion. 'OPNFV Verified' status does not assert readiness for
-commercial deployment.
-
-Please refer to the program governance guidelines and term & conditions documents for additional
-details using the respective links:
-
- * https://www.opnfv.org/wp-content/uploads/sites/12/2018/01/OVP-Governance-Guidelines-1.0.1-012218.pdf
- * https://www.opnfv.org/wp-content/uploads/sites/12/2018/01/OVP-Terms-and-Conditions-011918.pdf
-
-Release Data
-============
-
-+--------------------------------------+---------------------------------------+
-| **Project** | Dovetail |
-| | |
-+--------------------------------------+---------------------------------------+
-| **Repo tag** | ovp.1.0.0 |
-| | |
-+--------------------------------------+---------------------------------------+
-| **Release designation** | OPNFV Verified Program (OVP) |
-| | 2018.01 (Danube) |
-+--------------------------------------+---------------------------------------+
-| **Release date** | January 21st 2018 |
-| | |
-+--------------------------------------+---------------------------------------+
-| **Purpose of the delivery** | Support OVP 2018.01 release with |
-| | OPNFV Danube release as reference SUT |
-+--------------------------------------+---------------------------------------+
-
-Deliverables
-============
-
-Software
---------
-
-+-----------------+----------------------+-------------+
-| Docker | Docker Image | Tag |
-| Container | | |
-+=================+======================+=============+
-| dovetail | opnfv/dovetail | ovp.1.0.0 |
-+-----------------+----------------------+-------------+
-| functest | opnfv/functest | ovp.1.0.0 |
-+-----------------+----------------------+-------------+
-| yardstick | opnfv/yardstick | danube.3.2 |
-+-----------------+----------------------+-------------+
-| testapi | opnfv/testapi | ovp.1.0.0 |
-+-----------------+----------------------+-------------+
-| mongo | mongo | 3.2.1 |
-+-----------------+----------------------+-------------+
-
-
- - Dovetail Docker images: https://hub.docker.com/r/opnfv/dovetail
-
- - Functest Docker images: https://hub.docker.com/r/opnfv/functest
-
- - Yardstick Docker images: https://hub.docker.com/r/opnfv/yardstick
-
- - TestAPI Docker images: https://hub.docker.com/r/opnfv/testapi
-
- - MongoDB Docker images: https://hub.docker.com/r/mongo
-
-
-Documents
----------
-
- - System Preparation Guide: http://docs.opnfv.org/en/stable-danube/submodules/dovetail/docs/testing/user/systempreparation/index.html
-
- - User Guide: http://docs.opnfv.org/en/stable-danube/submodules/dovetail/docs/testing/user/userguide/testing_guide.html
-
- - Test Specifications: http://docs.opnfv.org/en/stable-danube/submodules/dovetail/docs/testing/user/testspecification/index.html
-
- - Dovetail CLI Reference: http://docs.opnfv.org/en/stable-danube/submodules/dovetail/docs/testing/user/userguide/cli_reference.html
-
- - Process Workflow: http://docs.opnfv.org/en/stable-danube/submodules/dovetail/docs/testing/user/certificationworkflow/index.html
-
- - Reviewer Guide: http://docs.opnfv.org/en/stable-danube/submodules/dovetail/docs/testing/user/reviewerguide/index.html
-
-
-Version Change
-==============
-
-This is the first major release of OVP/Dovetail. Please refer to the link below for minor
-version changes during pre-release and beta phases.
-
- * https://wiki.opnfv.org/display/dovetail/Running+history+for+the+dovetail+tool
-
-Testing with OPNFV Danube Installers
-====================================
-
-OVP 2018.01 and Dovetail 1.0.0 are known to be have been tested with the following OPNFV
-Danube installer versions.
-
-+-----------------+----------------------+
-| Installer | Version |
-+=================+======================+
-| Apex | danube.3.1 |
-+-----------------+----------------------+
-| Compass | danube.3.1 |
-+-----------------+----------------------+
-| Fuel | danube.3.0 |
-+-----------------+----------------------+
-
-
-Danube Known Restrictions/Issues
-==================================
-
-Please refer to the following link for known issues with the Dovetail Danube release:
-
- * https://wiki.opnfv.org/display/dovetail/Running+history+for+the+dovetail+tool#Runninghistoryforthedovetailtool-4.KnownIssuesList
-
-Open JIRA Tickets
-=================
-
-+------------------+-----------------------------------------------+
-| JIRA | Description |
-+==================+===============================================+
-| | |
-| | |
-+------------------+-----------------------------------------------+
-
-All blocking tickets have been fixed.
-
-
-Useful Links
-============
-
- - OVP Web Portal: https://verified.opnfv.org
-
- - Wiki Project Page: https://wiki.opnfv.org/display/dovetail
-
- - Dovetail Repo: https://git.opnfv.org/dovetail/
-
- - Dovetail CI dashboard: https://build.opnfv.org/ci/view/dovetail/
-
- - JIRA dashboard: https://jira.opnfv.org/secure/RapidBoard.jspa?rapidView=149
-
- - Dovetail IRC Channel: #opnfv-dovetail
-
- - Dovetail Test Configuration: https://git.opnfv.org/dovetail/tree/dovetail/compliance/ovp.1.0.0.yml
diff --git a/docs/release/release-notes/index.rst b/docs/release/release-notes/index.rst
index ad4cade0..2de92c55 100644
--- a/docs/release/release-notes/index.rst
+++ b/docs/release/release-notes/index.rst
@@ -3,18 +3,281 @@
.. _dovetail-releasenotes:
-**********************
-Dovetail Release Notes
-**********************
+======================================================================
+OPNFV Verification Program (OVP) 2019.12 / Dovetail 3.0.0 Release Note
+======================================================================
-.. toctree::
- :numbered:
- :maxdepth: 4
- dovetail-release.rst
+OPNFV 2019.12 Release
+=====================
-Revision: _sha1_
+The OPNFV Verification Program (OVP) allows vendors and operators to obtain 'OPNFV Verified'
+status based on an agreed upon set of compliance verification test cases that align to OPNFV
+releases. The reference System under Test (SUT) is either the NFV components deployed by the OPNFV
+installers for a given release, where OVP 2019.12 is based on the OPNFV Hunter release, or a VNF
+being on-boarded and orchestrated by the ONAP El Alto release. Participants of the program can
+verify commercial or open source offerings against an OVP release. This implies that the SUT
+used for verification has interfaces, components, functions and behaviors that align to OPNFV
+installer integrations and ONAP deployments.
-:Author: Eddie Arrage (eddie.arrage@huawei.com)
+Dovetail is the overall framework used to execute tests and collect results for the OVP
+Infrastructure badge. Dovetail does not deliver test content directly. Rather, test content
+is developed in other OPNFV test frameworks such as Functest and upstream test communities such
+as OpenStack's RefStack/Tempest projects. Dovetail leverages this upstream test content and
+provides a common set of test platform services for the OVP.
-Build date: |today|
+Approved test tools (OPNFV Dovetail, ONAP VTP, and ONAP VVP) work in conjunction with a web portal
+interface dubbed the 'OVP web portal' to allow users to upload test results to a centralized community
+repository. This facilitates user collaboration, result sharing, self-testing and community reviews.
+It also serves as a hub for new participants to learn about the program and access key resources. The
+link for this portal is at: `OPNFV Verification Program <https://nfvi-verified.lfnetworking.org>`_.
+
+Use of the OVP web portal is open to all and only requires a valid Linux Foundation
+ID to login. Users are welcome to use the portal to upload, inspect and share results in a private
+manner. In order to submit results for official review, the first step is apply for acceptance
+into the program with the participation form provided in the link: `OPNFV Verification Program
+Participation Form <https://na3.docusign.net/Member/PowerFormSigning.aspx?PowerFormId=dc24bf38-ea41-40d4-9e58-9babc6eec778>`_
+
+NFVI Test Suites and Test Areas
+-------------------------------
+
+OVP/Dovetail groups test cases into test suites and test areas. Test suites are currently a basic
+categorization around releases for the most part. Executing the test suite 'ovp.2019.12' without
+further specification will run all the test cases in the OVP 2019.12 release. Test suites are
+divided into test areas that can be executed separately.
+
+Test cases include a division into **'mandatory'** and **'optional'** in an overarching
+categorization.
+
+All the mandatory test cases are required to be executed with passing results for all inclusive
+test cases for results to be reviewed and approved by the community made up of peer reviewers.
+The optional test cases are not required to be executed for the official compliance verification
+review in the OVP 2019.12 release. However, execution of these cases is encouraged, as some
+optional test cases may become mandatory in future releases.
+
+NFVI Test Cases and Sub Test Cases
+----------------------------------
+
+Each test area consists of multiple test cases where each test case can be a single test or
+broken down into sub test cases. A listing of test cases with the number of sub test cases noted
+in parenthesis is shown below for the OVP 2019.12 release.
+
+**Mandatory**
+
+- bottlenecks.stress.ping (1)
+- functest.vping.userdata (1)
+- functest.vping.ssh (1)
+- functest.tempest.osinterop (219)
+- functest.tempest.compute (12)
+- functest.tempest.identity_v3 (11)
+- functest.tempest.image (2)
+- functest.tempest.network_api (14)
+- functest.tempest.volume (2)
+- functest.tempest.neutron_trunk_ports (38)
+- functest.tempest.ipv6_api (21)
+- functest.security.patrole (124)
+- yardstick.ha.nova_api (1)
+- yardstick.ha.neutron_server (1)
+- yardstick.ha.keystone (1)
+- yardstick.ha.glance_api (1)
+- yardstick.ha.cinder_api (1)
+- yardstick.ha.cpu_load (1)
+- yardstick.ha.disk_load (1)
+- yardstick.ha.haproxy (1)
+- yardstick.ha.rabbitmq (1)
+- yardstick.ha.database (1)
+
+
+There are a total of 456 mandatory test cases.
+
+**Optional**
+
+- functest.vnf.vims (1)
+- functest.vnf.vepc (1)
+- functest.tempest.ipv6_scenario (8)
+- functest.tempest.multi_node_scheduling (6)
+- functest.tempest.network_security (6)
+- functest.tempest.vm_lifecycle (12)
+- functest.tempest.network_scenario (5)
+- functest.tempest.bgpvpn (21)
+- functest.security.patrole_vxlan_dependent (2)
+- yardstick.ha.neutron_l3_agent (1)
+- yardstick.ha.controller_restart (1)
+
+There are a total of 64 optional test cases.
+
+OPNFV Test Projects and Components
+----------------------------------
+
+The OPNFV test frameworks integrated into the Dovetail framework that deliver test content are:
+
+ - Functest (leverages OpenStack RefStack/Tempest projects in addition to supplying native test cases)
+ - Yardstick
+ - Bottlenecks
+
+ONAP Test Projects and Components
+---------------------------------
+
+The ONAP test projects and components used with this OVP release to provide the test requirements
+and test scripting are:
+
+- VNFRQTS
+- VNFSDK
+- VVP
+
+Acceptence and Marketing
+------------------------
+
+Upon successful community review of results for OVP 2019.12, the Linux Foundation Compliance
+Verification Committee (LFN CVC) on behalf of the Board of Directors can award a product 'OPNFV
+Verified' status. Use of 'OPNFV Verified' Program Marks shall be awarded to the platform used
+for compliance verification. The category label of 'Infrastructure' is used within the Program
+Marks logo and limits the scope of this OVP release to a SUT consisting of NFVI and VIM components
+using ETSI terminology. It does not provide compliance verification for specific VNFs in any fashion.
+The date '2019.12' corresponds to a reference SUT that aligns to the OPNFV Hunter release and
+currently aligns to the Dovetail framework version 3.0.0.
+
+Organizations shall not use the Program Marks in any way that would associate it with any
+individual or company logo or brand, beyond the association to the specific platform to which it
+was awarded. While OpenStack RefStack interoperability and Tempest integration test cases are
+executed as part of the OVP 2019.12 compliance verification test suites, the OVP does not grant or
+award OpenStack Marks in any fashion. 'OPNFV Verified' status does not assert readiness for
+commercial deployment.
+
+Please refer to the program governance guidelines and term & conditions documents for additional
+details using the respective links:
+
+* `OVP Governance Guidelines <https://www.opnfv.org/wp-content/uploads/sites/12/2018/01/OVP-Governance-Guidelines-1.0.1-012218.pdf>`_
+* `OVP Terms and Conditions <https://www.opnfv.org/wp-content/uploads/sites/12/2018/01/OVP-Terms-and-Conditions-011918.pdf>`_
+
+Release Data
+============
+
++--------------------------------------+---------------------------------------+
+| **Project** | Dovetail |
+| | |
++--------------------------------------+---------------------------------------+
+| **Repo tag** | ovp-3.0.1 |
+| | |
++--------------------------------------+---------------------------------------+
+| **Release designation** | OPNFV Verification Program (OVP) |
+| | 2019.12 (Hunter) |
++--------------------------------------+---------------------------------------+
+| **Release date** | December 2019 |
+| | |
++--------------------------------------+---------------------------------------+
+| **Purpose of the delivery** | Support OVP 2019.12 release with |
+| | OPNFV Hunter release as reference SUT |
++--------------------------------------+---------------------------------------+
+| **Notes** | Point release ovp-3.0.1 updates, |
+| | changes, and corrects the |
+| | documentation only. |
++--------------------------------------+---------------------------------------+
+
+
+Deliverables
+============
+
+Software
+--------
+
+OPNFV Software
+""""""""""""""
+
++-------------------------+-----------------------------------+----------------+
+| **Docker Container** | **Docker Image** | **Tag** |
++-------------------------+-----------------------------------+----------------+
+| dovetail | opnfv/dovetail | ovp-3.0.0 |
++-------------------------+-----------------------------------+----------------+
+| functest | opnfv/functest-smoke | hunter |
++-------------------------+-----------------------------------+----------------+
+| functest | opnfv/functest-healthcheck | hunter |
++-------------------------+-----------------------------------+----------------+
+| functest | opnfv/functest-vnf | hunter |
++-------------------------+-----------------------------------+----------------+
+| yardstick | opnfv/yardstick | opnfv-8.0.0 |
++-------------------------+-----------------------------------+----------------+
+| bottlenecks | opnfv/bottlenecks | 8.0.1-latest |
++-------------------------+-----------------------------------+----------------+
+
+**Docker images:**
+
+- `Dovetail Docker images <https://hub.docker.com/r/opnfv/dovetail>`_
+- `Functest-smoke Docker images <https://hub.docker.com/r/opnfv/functest-smoke/>`_
+- `Functest-healthcheck Docker images <https://hub.docker.com/r/opnfv/functest-healthcheck/>`_
+- `Functest-vnf Docker images <https://hub.docker.com/r/opnfv/functest-vnf/>`_
+- `Yardstick Docker images <https://hub.docker.com/r/opnfv/yardstick/>`_
+- `Bottlenecks Docker images <https://hub.docker.com/r/opnfv/bottlenecks/>`_
+
+ONAP Software
+"""""""""""""
+
++-------------------------+--------------------------------------------------------------+
+| **Item** | **Repo Link** |
++-------------------------+--------------------------------------------------------------+
+| VTP/VNFSDK Test Scripts | `<https://gerrit.onap.org/r/admin/repos/vnfsdk/refrepo>`_ |
++-------------------------+--------------------------------------------------------------+
+| VVP Robot Test Scripts | `<https://gerrit.onap.org/r/admin/repos/oom>`_ |
++-------------------------+--------------------------------------------------------------+
+
+Documents
+---------
+
+- `System Preparation Guide <https://opnfv-dovetail.readthedocs.io/en/stable-hunter/testing/user/systempreparation/index.html>`_
+
+- `NFVI User Guide <https://opnfv-dovetail.readthedocs.io/en/stable-hunter/testing/user/userguide/testing_guide.html>`_
+
+- `VNF User Guide: <https://opnfv-dovetail.readthedocs.io/en/stable-hunter/testing/user/userguide/testing_guide.html>`_
+
+- `OVP NFVI Test Specifications <https://opnfv-dovetail.readthedocs.io/en/stable-hunter/testing/user/testspecification/index.html>`_
+
+- `ONAP VNF Test Specifications <https://docs.onap.org/en/elalto/submodules/vnfrqts/testcases.git/docs/index.html>`_
+
+- `Dovetail CLI Reference <https://opnfv-dovetail.readthedocs.io/en/stable-hunter/testing/user/userguide/cli_reference.html>`_
+
+- `Dovetail RESTful API <https://opnfv-dovetail.readthedocs.io/en/stable-hunter/testing/user/userguide/api_testing_guide.html>`_
+
+- `OVP Workflow <https://opnfv-dovetail.readthedocs.io/en/stable-hunter/testing/user/certificationworkflow/index.html>`_
+
+- `OVP Reviewer Guide <https://opnfv-dovetail.readthedocs.io/en/stable-hunter/testing/user/reviewerguide/index.html>`_
+
+
+Testing with OPNFV Hunter Installers
+====================================
+
+OVP 2019.12 and Dovetail 3.0.0 are known to be have been tested with the following OPNFV
+Hunter installer versions.
+
++-----------------+----------------------+
+| Installer | Version |
++=================+======================+
+| Fuel | stable/hunter |
++-----------------+----------------------+
+
+
+Hunter Known Restrictions/Issues
+================================
+
+Please refer to the OPNFV and ONAP JIRA for known issues with each applicable project:
+
+- `<https://jira.opnfv.org/projects/DOVETAIL>`_
+- `<https://jira.onap.org/projects/VVP>`_
+- `<https://jira.onap.org/projects/VNFSDK>`_
+
+
+Useful Links
+============
+
+ - `OVP Web Portal <https://nfvi-verified.lfnetworking.org>`_
+
+ - `Wiki Project Page <https://wiki.opnfv.org/display/dovetail>`_
+
+ - `Dovetail Repo <https://git.opnfv.org/dovetail/>`_
+
+ - `Dovetail CI dashboard <https://build.opnfv.org/ci/view/dovetail/>`_
+
+ - `JIRA dashboard <https://jira.opnfv.org/secure/RapidBoard.jspa?rapidView=149>`_
+
+ - Dovetail IRC Channel: #opnfv-dovetail
+
+ - `Dovetail Test Configuration <https://git.opnfv.org/dovetail/tree/etc/compliance/ovp.2019.12.yaml>`_
diff --git a/docs/requirements.txt b/docs/requirements.txt
new file mode 100644
index 00000000..44084358
--- /dev/null
+++ b/docs/requirements.txt
@@ -0,0 +1,5 @@
+lfdocs-conf
+sphinx_opnfv_theme
+# Uncomment the following line if your project uses Sphinx to document
+# HTTP APIs
+# sphinxcontrib-httpdomain
diff --git a/docs/testing/developer/genericframework/index.rst b/docs/testing/developer/genericframework/index.rst
new file mode 100644
index 00000000..75721469
--- /dev/null
+++ b/docs/testing/developer/genericframework/index.rst
@@ -0,0 +1,413 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) Huawei Technologies Co.,Ltd, and others
+
+====================================
+Dovetail as a Generic Test Framework
+====================================
+
+.. toctree::
+ :maxdepth: 2
+
+
+Overview
+========
+
+Dovetail is responsible for the technical realization of the OPNFV Verification
+Program (OVP) and other compliance verification projects within the scope of
+the Linux Foundation Networking (LFN) umbrella projects.
+Dovetail provides a generic framework for executing a specific set of test cases
+which define the scope of a given compliance verification program, such as OVP.
+
+This document aims at introducing what Dovetail generic framework looks like and
+how to develop within this framework.
+
+
+Introduction of Dovetail Framework
+==================================
+
+The following diagram illustrates Dovetail generic framework.
+
+.. image:: ../../../images/dovetail_generic_framework.png
+ :align: center
+ :scale: 50%
+
+In this diagram, there are 5 main parts, `TestcaseFactory`, `TestRunnerFactory`,
+`CrawlerFactory`, `CheckerFactory` and `test case groups`.
+
+- **TestcaseFactory**: For each project, there needs to create its own
+ testcase class such as `FunctestTestcase` and `OnapVtpTestcase`. All these
+ classes are based on class `Testcase`. There are already many functions in this
+ base class which are mainly used to parse test case configuration files. If no
+ other special requirements exist, it only needs to initialize these classes with
+ different types. Otherwise, it needs to overload or add some functions.
+
+- **TestRunnerFactory**: Similar to `TestcaseFactory`, each project has its own
+ test runner class. Dovetail supports 2 kinds of test runners, `DockerRunner`
+ and `ShellRunner`. For projects based on Docker, it needs to create
+ their own test runner classes such as `FunctestRunner` which inherit from class
+ `DockerRunner`. For other projects that are based on Shell, it can use `ShellRunner`
+ directly. Test case runners provide many functions to support test cases runs
+ such as preparing test tool of each project, run all the commands defined by
+ each test case and clean the environment.
+
+- **Test case groups**: Each group is composed of one project configuration file
+ and a set of test cases belonging to this project. These groups are used as the
+ input of test runners to provide information of projects and test cases. For
+ `ShellRunner`, it only needs test case configuratons as the input.
+
+- **CrawlerFactory**: This is used to parse the results of test cases and record
+ them with unified format. The original result data report by each project is
+ different. So it needs to create different crawler classes for different projects
+ to parse their results.
+
+- **CheckerFactory**: This is used to check the result data generated by crawler.
+ Each project should have its own checker class due to the different requirements
+ of different projects.
+
+
+Development with Dovetail Framework
+===================================
+
+Everyone who is interested in developing Dovetail framework to integrate new upstream
+test cases will face one of the two following scenarios:
+
+- **Adding test cases that belong to integrated projects**: There are already some
+ projects integrated in Dovetail. These projects are coming from OPNFV (Open Platform
+ for NFV) and ONAP (Open Network Automation Platform) communities. It will be
+ much easier to add new test cases that belong to these projects.
+
+- **Adding test cases that not belong to integrated projects**: The test cases
+ may belong to other projects that haven't been integrated into Dovetail yet.
+ These projects could be in OPNFV, ONAP or other communities. This scenario is a
+ little more complicated.
+
+
+Test cases belonging to integrated projects
+-------------------------------------------
+
+Dovetail framework already includes a large amount of test cases. All these test
+cases are implemented by upstream projects in OPNFV and ONAP. The upstream
+projects already integrated in Dovetail are Functest, Yardstick and Bottlenecks
+from OPNFV and VNF SDK and VVP from ONAP.
+
+In order to add a test case belonging to one of these projects, there
+only need to add one test case configuration file which is in yaml format.
+Following is the introduction about how to use the file to add one new test case.
+Please refer to `Dovetail test case github
+<https://github.com/opnfv/dovetail/tree/master/etc/testcase>`_
+for all configuration files of all test cases.
+
+.. code-block:: bash
+
+ ---
+ Test case name in Dovetail:
+ name: Test case name in Dovetail
+ objective: Test case description
+ validate:
+ type: 'shell' or name of the project already integrated in Dovetail
+ testcase: The original test case name called in the project that it is developed
+ image_name: Name of the Docker image used to run this test
+ pre_condition:
+ - 'Commands needed to be executed before running this test'
+ - 'e.g. cp src_file dest_file'
+ cmds:
+ - 'Commands used to run this test case'
+ post_condition:
+ - 'Commands needed to be executed after running this test case'
+ report:
+ source_archive_files:
+ - test.log
+ dest_archive_files:
+ - path/to/archive/test.log
+ check_results_files:
+ - results.json
+ portal_key_file: path/to/key/logs/xxx.log
+ sub_testcase_list:
+ - sub_test_1
+ - sub_test_2
+ - sub_test_3
+
+This is the complete format of test case configuration file. Here are some
+detailed description for each of the configuration options.
+
+- **Test case name in Dovetail**: All test cases should be named as 'xxx.yyy.zzz'.
+ This is the alias in Dovetail and has no relationship with its name in its own
+ project. The first part is used to identify the project where this test case
+ come from (e.g. functest, onap-vtp). The second part is used to classify this
+ test case according to test area (e.g. healthcheck, ha). Dovetail supports to
+ run whole test cases in one test suite with the same test area. Also the area
+ is used to group all test cases and generate the summary report at the end of
+ the test. The last part is special for this test case itself (e.g. image,
+ haproxy, csar). It's better to keep the file name the same as the test case
+ name to make it easier to find the config file according to this test case
+ alias in Dovetail.
+
+- **validate**: This is the main section to define how to run this test case.
+
+ - **type**: This is the type of this test case. It can be `shell` which means
+ running this test case with Linux bash commands within Dovetail container. Also it
+ can be one of the projects already integrated in Dovetail (functest, yardstick,
+ bottlenecks, onap-vtp and onap-vvp). Then this type is used to map to its project
+ configuration yaml file. For example, in order to add a test case
+ in OPNFV project Functest to Dovetail framework, the type here should be
+ `functest`, and will map to `functest_config.yml` for more configurations
+ in project level. Please refer to `Dovetail project config github
+ <https://github.com/opnfv/dovetail/tree/master/etc/conf>`_ for more details.
+
+ - **testcase**: This is the name defined in its own project. One test case can
+ be uniquely identified by `type` and `testcase`. Take the test case
+ `functest.vping.ssh` as an example. Its `type` is 'functest' and `testcase`
+ is 'vping_ssh'. With these 2 properties, it can be uniquely identified. End users only
+ need to know that there is a test case named `functest.vping.ssh` in OVP
+ compliance test scope. Dovetail Framework will run `vping_ssh` within Functest
+ Docker container.
+
+ - **image_name**: [optional] If the type is `shell`, there is no need to give
+ this. For other types, there are default docker images defined in their project
+ configuration files. If this test case uses a different docker image, it needs
+ to overwrite it by adding `image_name` here. The `image_name` here should only
+ be the docker image name without tag. The tag is defined in project's configuration
+ file for all test cases belonging to this project.
+
+ - **pre_condition**: [optional] A list of all preparations needed by this
+ test case. If the list is the same as the default one in its project configuration
+ file, then there is no need to repeat it here. Otherwise, it's necessary to
+ overwrite it. If its type is `shell`, then all commands in `pre_condition`,
+ `cmds` and `post_condition` should be executable within Dovetail Ubuntu 14.04
+ Docker container. If its type is one of the Docker runner projects, then all
+ commands should be executable within their own containers. For Functest, it's
+ alpine 3.8. For Yardstick and Bottlenecks it's Ubuntu 16.04. For VNF SDK it's
+ Ubuntu 14.04. Also all these commands should not require network connection
+ because some commercial platforms may be offline environments in private labs.
+
+ - **cmds**: [optional] A list of all commands used to run this test case.
+
+ - **post_condition**: [optional] A list of all commands needed after executing
+ this test case such as some clean up operations.
+
+- **report**: This is the section for this test case to archive some log files and
+ provide the result file for reporting PASS or FAIL.
+
+ - **source_archive_files**: [optional] If there is no need to archive any files,
+ this section can be removed. Otherwise, this is a list of all source files
+ needed to be archived. All files generated by all integrated projects will be
+ put under `$DOVETAIL_HOME/results`. In order to classify and avoid overwriting
+ them, it needs to rename some important files or move them to new directories.
+ Navigating directory `$DOVETAIL_HOME/results` to find out all files
+ needed to be archived. The paths here should be relative ones according to
+ `$DOVETAIL_HOME/results`.
+
+ - **dest_archive_files**: [optional] This should be a list corresponding to the
+ list of `source_archive_files`. Also all paths here should be relative ones
+ according to `$DOVETAIL_HOME/results`.
+
+ - **check_results_files**: This should be a list of relative paths of
+ the result files generated by this test case. Dovetail will parse these files
+ to get the result (PASS or FAIL).
+
+ - **portal_key_file**: This should be the key log file of this test case which will
+ be used by the OVP portal for review.
+
+ - **sub_testcase_list**: [optional] This section is almost only for Tempest tests
+ in Functest. Take `functest.tempest.osinterop` as an example. The `sub_testcase_list`
+ here is an check list for this kind of tempest tests. Only when all sub test
+ cases list here are passed, this test case can be taken as PASS. The other kind
+ of tempest tests is `tempest_custom` such as `functest.tempest.image`. Besides
+ taking the `sub_testcase_list` as the check list, it's also used to generate an
+ input file of Functest to define the list of sub test cases to be tested.
+
+
+Test cases not belonging to integrated projects
+-----------------------------------------------
+
+If test cases waiting to be added into Dovetail do not belong to any project
+that is already integrated into Dovetail framework, then besides adding the test
+case configuration files introduced before, there are some other files needed to
+be added or modified.
+
+
+Step 1: Add a project configuration file
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+For a new test case that belongs to a new project, it needs to create a project
+configuration file to define this new project in Dovetail first. Now Dovetail
+only supports integration of projects by using their Docker images. If this test
+case should be run with shell runner, then can only add test case configuration
+files with `type` 'shell' as describing before and skip the following steps. Following is
+the introduction of how to use project configuration file to add one new project
+into Dovetail. Please refer to `Dovetail projects configuration github
+<https://github.com/opnfv/dovetail/tree/master/etc/conf>`_ for all configuration
+files of all integrated projects.
+
+.. code-block:: bash
+
+ ---
+
+ {% set validate_testcase = validate_testcase or '' %}
+ {% set testcase = testcase or '' %}
+ {% set dovetail_home = dovetail_home or '' %}
+ {% set debug = debug or 'false' %}
+ {% set build_tag = build_tag or '' %}
+ {% set userconfig_dir = '/tmp/userconfig' %}
+ {% set patches_dir = '/tmp/patches' %}
+ {% set result_dir = '/tmp/results' %}
+ {% set openrc_file = '/home/conf/env_file' %}
+
+ project name:
+ image_name: name of the docker image
+ docker_tag: tag of the docker image
+ opts:
+ detach: true
+ stdin_open: true
+ privileged: true
+ shell: '/bin/bash'
+ envs:
+ - 'CI_DEBUG={{debug}}'
+ - 'DEPLOY_SCENARIO={{deploy_scenario}}'
+ - 'ENV_NAME=env_value'
+ volumes:
+ - '{{dovetail_home}}/userconfig:{{userconfig_dir}}'
+ - '{{dovetail_home}}/results:{{result_dir}}'
+ - '/path/on/host:/path/in/container'
+ - '/path/of/host/file:/file/path/in/container'
+ mounts:
+ - 'source={{dovetail_home}}/pre_config/env_config.sh,target={{openrc_file}}
+ - 'source={{dovetail_home}}/pre_config,target=/home/opnfv/pre_config'
+ - 'source=/file/or/derectory/on/host,target=/file/or/derectory/in/container'
+ patches_dir: {{patches_dir}}
+ pre_condition:
+ - 'Commands needed to be executed before running this test'
+ cmds:
+ - 'Commands used to run this test case'
+ post_condition:
+ - 'Commands needed to be executed after running this test case'
+ openrc: absolute path of openstack credential files
+ extra_container:
+ - container1_name
+ - container2_name
+
+This is the complete format of project configuration file. Here are some
+detailed description for each of the configuration options.
+
+- **Jinja Template**: At the begining of this yaml file, it uses Jinja template
+ to define some parameters that will be used somewhere in this file (e.g. result_dir
+ and openrc_file). Besides those, there are some other parameters providing by Dovetail
+ framework as input of this file, and other parameters can be defined by using these
+ ones (e.g. testcase and dovetail_home). The whole input parameters which can be used
+ are list below.
+
+ - **attack_host**: This is the attack host name of the test case which calls this
+ project configuration file. It's only for HA test cases and can be given in HA
+ configuration file `pod.yaml`.
+
+ - **attack_process**: This is the attack process name of the test case which calls
+ this project configuration file. It's only for HA test cases and can be given in HA
+ configuration file `pod.yaml`.
+
+ - **build_tag**: This is a string includes the UUID generated by Dovetail.
+
+ - **cacert**: This is also only for OpenStack test cases. It is the absolute
+ path of the OpenStack certificate provided in `env_config.sh` file.
+
+ - **deploy_scenario**: This is the input when running Dovetail with option
+ `--deploy-scenario`.
+
+ - **debug**: This is `True` or `False` according to the command running test
+ cases with or without option `--debug`.
+
+ - **dovetail_home**: This is the `DOVETAIL_HOME` getting from the ENV.
+
+ - **os_insecure**: This is only for test cases aiming at OpenStack. This is
+ `True` or `False` according to `env_config.sh` file.
+
+ - **testcase**: This is the name of the test case which calls this project
+ configuration file. Different from `validate_testcase`, this is the alias
+ defined in Dovetail not in its own project.
+
+ - **validate_testcase**: This is the name of the test case instance which calls this
+ project configuration file. The name is provided by the configuration file
+ of this test case (validate -> testcase).
+
+- **project name**: This is the project name defined in Dovetail. For example
+ OPNFV Functest project is named as 'functest' here in Dovetail. This project
+ name will be used by test case configuration files as well as somewhere in
+ Dovetail source code.
+
+- **image_name**: This is the name of the default Docker image for most test cases
+ within this project. Each test case can overwrite it with its own configuration.
+
+- **docker_tag**: This is the tag of all Docker images for all test cases within
+ this project. For each release, it should use one Docker image with a stable
+ and official release version.
+
+- **opts**: Here are all options used to run Docker containers except 'image',
+ 'command', 'environment', 'volumes', 'mounts' and 'extra_hosts'. For example,
+ the options include 'detach', 'privileged' and 'tty'. The full list of all
+ options can be found in `Docker python SDK docs <https://docker-py.readthedocs.io/en/stable/containers.html>`_.
+
+- **shell**: This is the command used to run in the container.
+
+- **envs**: This is a list of all envs used to run Docker containers.
+
+- **volumes**: A volume mapping list used to run Docker containers. The source volumes
+ list here are allowed to be nonexistent and Docker will create new directories for them
+ on the host. Every project should at least map the `$DOVETAIL_HOME/results`
+ in the test host to containers to collect all result files.
+
+- **mounts**: A mount mapping list used to run Docker containers. More powerful alternative
+ to **volumes**. The source volumes list here are not allowed to be nonexistent.
+ Every project should at least mount the `$DOVETAIL_HOME/pre_config` in the test host to
+ containers to get config files.
+
+- **patches_dir**: [optional] This is an absolute path of the patches applied to
+ the containers.
+
+- **pre_condition**: A list of all default preparations needed by this project.
+ It can be overwritten by configurations of test cases.
+
+- **cmds**: A list of all default commands used to run all test cases within
+ this project. Also it can be overwritten by configurations of test cases.
+
+- **post_condition**: A list of all default cleaning commands needed by this
+ project.
+
+- **openrc**: [optional] If the system under test is OpenStack, then it needs to
+ provide the absolute path here to copy the credential file in the Test Host to
+ containers.
+
+- **extra_container**: [optional] The extra containers needed to be removed at the
+ end of the test. These containers are created by the test cases themselves at
+ runtime rather than created by Dovetail.
+
+
+Step 2: Add related classes
+^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+After adding the project and test case configuration files, there also need to
+add some related classes into the source code.
+
+- **Test Case class**: Each project should have its own test case class in
+ `testcase.py` for `TestcaseFactory`.
+
+- **Test Runner class**: Each project should have its own test runner class in
+ `test_runner.py` for `TestRunnerFactory`.
+
+- **Crawler class**: Each project should have its own test results crawler class
+ in `report.py` for `CrawlerFactory`.
+
+- **Checker class**: Each project should have its own test results checker class
+ in `report.py` for `CheckerFactory`.
+
+
+Step 3: Create related logs
+^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+If the classes added in step2 have function `create_log`, then need to call
+these functions in `run.py` to initial the log instances at the very begining.
+
+
+Step 4: Update unit tests
+^^^^^^^^^^^^^^^^^^^^^^^^^
+
+A patch is not going to be verified without 100% coverage when applying acceptance check.
diff --git a/docs/testing/developer/testcaserequirements/index.rst b/docs/testing/developer/testcaserequirements/index.rst
index 9084462e..b03a033c 100644
--- a/docs/testing/developer/testcaserequirements/index.rst
+++ b/docs/testing/developer/testcaserequirements/index.rst
@@ -4,9 +4,9 @@
.. _dovetail-test_case_requirements:
-==========================================================
-OPNFV Verified Program test case requirements
-==========================================================
+==========================
+OVP Test Case Requirements
+==========================
.. toctree::
:maxdepth: 2
@@ -16,20 +16,20 @@ OVP Test Suite Purpose and Goals
================================
The OVP test suite is intended to provide a method for validating the
-interfaces and behaviors of an NFVi platform according to the expected
-capabilities exposed in OPNFV. The behavioral foundation evaluated in these
+interfaces and behaviors of an NFVI platform according to the expected
+capabilities exposed in OPNFV. The behavioral foundation evaluated in these
tests should serve to provide a functional baseline for VNF deployment and
-portability across NFVi instances. All OVP tests are available in open source
+portability across NFVI instances. All OVP tests are available in open source
and are executed in open source test frameworks.
-Test case requirements
+Test Case Requirements
======================
The following requirements are mandatory for a test to be submitted for
consideration in the OVP test suite:
-- All test cases must be fully documented, in a common format. Please consider
+- All test cases must be fully documented, in a common format. Please consider
the existing :ref:`dovetail-test_case_specification` as examples.
- Clearly identifying the test procedure and expected results / metrics to
@@ -56,7 +56,7 @@ consideration in the OVP test suite:
- Test cases must pass on applicable OPNFV reference deployments and release
versions.
- - Tests must not require a specific NFVi platform composition or installation
+ - Tests must not require a specific NFVI platform composition or installation
tool.
- Tests and test tools must run independently of the method of platform
@@ -101,7 +101,7 @@ consideration in the OVP test suite:
- Use case specification
- Test preconditions
- Basic test flow execution description and test assertions
- - Pass fail criteria
+ - Pass/Fail criteria
- The following things may be documented for the test case:
@@ -120,21 +120,18 @@ Dovetail Test Suite Naming Convention
Test case naming and structuring must comply with the following conventions.
The fully qualified name of a test case must comprise three sections:
-`<community>.<test_area>.<test_case_name>`
+`<test_project>.<test_area>.<test_case_name>`
-- **community**: The fully qualified test case name must identify the community
- or upstream project which developed and maintains the test case. For test
- cases originating in OPNFV projects, the community identifier is 'opnfv'.
- Test cases consumed from the OpenStack tempest test suite, are named
- 'tempest', for example.
+- **test_project**: The fully qualified test case name must identify the test
+ project which developed and maintains the test case.
- **test_area**: The fully qualified test case name must identify the test case
- area. For test cases originating in OPNFV projects, the test case area must
- identify the project name.
+ area. The test case area is a single word identifier describing the broader
+ functional scope of a test case, such as ha (high-availability), tempest, vnf,
+ etc.
- **test_case_name**: The fully qualified test case name must include a concise
description of the purpose of the test case.
-An example of a fully qualified test case name is
-`opnfv.sdnvpn.router_association_floating_ip`.
+An example of a fully qualified test case name is `functest.tempest.compute`.
diff --git a/docs/testing/developer/testscope/index.rst b/docs/testing/developer/testscope/index.rst
deleted file mode 100644
index ed5d2e3d..00000000
--- a/docs/testing/developer/testscope/index.rst
+++ /dev/null
@@ -1,616 +0,0 @@
-.. This work is lit_snapshots_list_details_with_paramsensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. (c) OPNFV
-
-.. _dovetail-test_scope:
-
-=======================================================
-Compliance and Verification program accepted test cases
-=======================================================
-
- .. toctree::
- :maxdepth: 2
-
-
-Mandatory OVP Test Areas
-========================
-
-----------------------------------
-Test Area VIM Operations - Compute
-----------------------------------
-
-Image operations within the Compute API
----------------------------------------
-
-| tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON.test_create_delete_image
-| tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON.test_create_image_specify_multibyte_character_image_name
-
-
-Basic support Compute API for server actions such as reboot, rebuild, resize
-----------------------------------------------------------------------------
-
-| tempest.api.compute.servers.test_instance_actions.InstanceActionsTestJSON.test_get_instance_action
-| tempest.api.compute.servers.test_instance_actions.InstanceActionsTestJSON.test_list_instance_actions
-
-
-Generate, import, and delete SSH keys within Compute services
--------------------------------------------------------------
-
-| tempest.api.compute.servers.test_servers.ServersTestJSON.test_create_specify_keypair
-
-
-List supported versions of the Compute API
-------------------------------------------
-
-| tempest.api.compute.test_versions.TestVersions.test_list_api_versions
-
-
-Quotas management in Compute API
---------------------------------
-
-| tempest.api.compute.test_quotas.QuotasTestJSON.test_get_default_quotas
-| tempest.api.compute.test_quotas.QuotasTestJSON.test_get_quotas
-
-
-Basic server operations in the Compute API
-------------------------------------------
-
-| tempest.api.compute.servers.test_servers.ServersTestJSON.test_create_server_with_admin_password
-| tempest.api.compute.servers.test_servers.ServersTestJSON.test_create_with_existing_server_name
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_create_numeric_server_name
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_create_server_metadata_exceeds_length_limit
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_create_server_name_length_exceeds_256
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_create_with_invalid_flavor
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_create_with_invalid_image
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_create_with_invalid_network_uuid
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_delete_server_pass_id_exceeding_length_limit
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_delete_server_pass_negative_id
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_get_non_existent_server
-| tempest.api.compute.servers.test_create_server.ServersTestJSON.test_host_name_is_same_as_server_name
-| tempest.api.compute.servers.test_create_server.ServersTestManualDisk.test_host_name_is_same_as_server_name
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_invalid_ip_v6_address
-| tempest.api.compute.servers.test_create_server.ServersTestJSON.test_list_servers
-| tempest.api.compute.servers.test_create_server.ServersTestJSON.test_list_servers_with_detail
-| tempest.api.compute.servers.test_create_server.ServersTestManualDisk.test_list_servers
-| tempest.api.compute.servers.test_create_server.ServersTestManualDisk.test_list_servers_with_detail
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_detailed_filter_by_flavor
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_detailed_filter_by_image
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_detailed_filter_by_server_name
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_detailed_filter_by_server_status
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_detailed_limit_results
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_filter_by_flavor
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_filter_by_image
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_filter_by_limit
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_filter_by_server_name
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_filter_by_active_status
-| tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_filtered_by_name_wildcard
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_changes_since_future_date
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_changes_since_invalid_date
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_limits
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_limits_greater_than_actual_count
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_limits_pass_negative_value
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_limits_pass_string
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_non_existing_flavor
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_non_existing_image
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_non_existing_server_name
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_detail_server_is_deleted
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_status_non_existing
-| tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_with_a_deleted_server
-| tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_lock_unlock_server
-| tempest.api.compute.servers.test_server_metadata.ServerMetadataTestJSON.test_delete_server_metadata_item
-| tempest.api.compute.servers.test_server_metadata.ServerMetadataTestJSON.test_get_server_metadata_item
-| tempest.api.compute.servers.test_server_metadata.ServerMetadataTestJSON.test_list_server_metadata
-| tempest.api.compute.servers.test_server_metadata.ServerMetadataTestJSON.test_set_server_metadata
-| tempest.api.compute.servers.test_server_metadata.ServerMetadataTestJSON.test_set_server_metadata_item
-| tempest.api.compute.servers.test_server_metadata.ServerMetadataTestJSON.test_update_server_metadata
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_server_name_blank
-| tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_reboot_server_hard
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_reboot_non_existent_server
-| tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_rebuild_server
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_rebuild_deleted_server
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_rebuild_non_existent_server
-| tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_stop_start_server
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_stop_non_existent_server
-| tempest.api.compute.servers.test_servers.ServersTestJSON.test_update_access_server_address
-| tempest.api.compute.servers.test_servers.ServersTestJSON.test_update_server_name
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_update_name_of_non_existent_server
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_update_server_name_length_exceeds_256
-| tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON.test_update_server_set_empty_name
-| tempest.api.compute.servers.test_create_server.ServersTestJSON.test_verify_created_server_vcpus
-| tempest.api.compute.servers.test_create_server.ServersTestJSON.test_verify_server_details
-| tempest.api.compute.servers.test_create_server.ServersTestManualDisk.test_verify_created_server_vcpus
-| tempest.api.compute.servers.test_create_server.ServersTestManualDisk.test_verify_server_details
-
-
-Retrieve volume information through the Compute API
----------------------------------------------------
-
-| tempest.api.compute.volumes.test_attach_volume.AttachVolumeTestJSON.test_attach_detach_volume
-| tempest.api.compute.volumes.test_attach_volume.AttachVolumeTestJSON.test_list_get_volume_attachments
-
-
-
------------------------------------
-Test Area VIM Operations - Identity
------------------------------------
-
-API discovery operations within the Identity v3 API
----------------------------------------------------
-
-| tempest.api.identity.v3.test_api_discovery.TestApiDiscovery.test_api_media_types
-| tempest.api.identity.v3.test_api_discovery.TestApiDiscovery.test_api_version_resources
-| tempest.api.identity.v3.test_api_discovery.TestApiDiscovery.test_api_version_statuses
-
-
-Auth operations within the Identity API
----------------------------------------
-
-| tempest.api.identity.v3.test_tokens.TokensV3Test.test_create_token
-
-
---------------------------------
-Test Area VIM Operations - Image
---------------------------------
-
-Image deletion tests using the Glance v2 API
---------------------------------------------
-
-| tempest.api.image.v2.test_images.BasicOperationsImagesTest.test_delete_image
-| tempest.api.image.v2.test_images_negative.ImagesNegativeTest.test_delete_image_null_id
-| tempest.api.image.v2.test_images_negative.ImagesNegativeTest.test_delete_non_existing_image
-| tempest.api.image.v2.test_images_tags_negative.ImagesTagsNegativeTest.test_delete_non_existing_tag
-
-
-Image get tests using the Glance v2 API
----------------------------------------
-
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_get_image_schema
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_get_images_schema
-| tempest.api.image.v2.test_images_negative.ImagesNegativeTest.test_get_delete_deleted_image
-| tempest.api.image.v2.test_images_negative.ImagesNegativeTest.test_get_image_null_id
-| tempest.api.image.v2.test_images_negative.ImagesNegativeTest.test_get_non_existent_image
-
-
-CRUD image operations in Images API v2
---------------------------------------
-
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_list_no_params
-
-
-Image list tests using the Glance v2 API
-----------------------------------------
-
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_list_images_param_container_format
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_list_images_param_disk_format
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_list_images_param_limit
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_list_images_param_min_max_size
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_list_images_param_size
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_list_images_param_status
-| tempest.api.image.v2.test_images.ListUserImagesTest.test_list_images_param_visibility
-
-
-Image update tests using the Glance v2 API
-------------------------------------------
-
-| tempest.api.image.v2.test_images.BasicOperationsImagesTest.test_update_image
-| tempest.api.image.v2.test_images_tags.ImagesTagsTest.test_update_delete_tags_for_image
-| tempest.api.image.v2.test_images_tags_negative.ImagesTagsNegativeTest.test_update_tags_for_non_existing_image
-
-
-----------------------------------
-Test Area VIM Operations - Network
-----------------------------------
-
-Basic CRUD operations on L2 networks and L2 network ports
----------------------------------------------------------
-
-| tempest.api.network.test_networks.NetworksTest.test_create_delete_subnet_with_allocation_pools
-| tempest.api.network.test_networks.NetworksTest.test_create_delete_subnet_with_dhcp_enabled
-| tempest.api.network.test_networks.NetworksTest.test_create_delete_subnet_with_gw
-| tempest.api.network.test_networks.NetworksTest.test_create_delete_subnet_with_gw_and_allocation_pools
-| tempest.api.network.test_networks.NetworksTest.test_create_delete_subnet_with_host_routes_and_dns_nameservers
-| tempest.api.network.test_networks.NetworksTest.test_create_delete_subnet_without_gateway
-| tempest.api.network.test_networks.NetworksTest.test_create_delete_subnet_all_attributes
-| tempest.api.network.test_networks.NetworksTest.test_create_update_delete_network_subnet
-| tempest.api.network.test_networks.NetworksTest.test_delete_network_with_subnet
-| tempest.api.network.test_networks.NetworksTest.test_list_networks
-| tempest.api.network.test_networks.NetworksTest.test_list_networks_fields
-| tempest.api.network.test_networks.NetworksTest.test_list_subnets
-| tempest.api.network.test_networks.NetworksTest.test_list_subnets_fields
-| tempest.api.network.test_networks.NetworksTest.test_show_network
-| tempest.api.network.test_networks.NetworksTest.test_show_network_fields
-| tempest.api.network.test_networks.NetworksTest.test_show_subnet
-| tempest.api.network.test_networks.NetworksTest.test_show_subnet_fields
-| tempest.api.network.test_networks.NetworksTest.test_update_subnet_gw_dns_host_routes_dhcp
-| tempest.api.network.test_ports.PortsTestJSON.test_create_bulk_port
-| tempest.api.network.test_ports.PortsTestJSON.test_create_port_in_allowed_allocation_pools
-| tempest.api.network.test_ports.PortsTestJSON.test_create_update_delete_port
-| tempest.api.network.test_ports.PortsTestJSON.test_list_ports
-| tempest.api.network.test_ports.PortsTestJSON.test_list_ports_fields
-| tempest.api.network.test_ports.PortsTestJSON.test_show_port
-| tempest.api.network.test_ports.PortsTestJSON.test_show_port_fields
-| tempest.api.network.test_ports.PortsTestJSON.test_update_port_with_security_group_and_extra_attributes
-| tempest.api.network.test_ports.PortsTestJSON.test_update_port_with_two_security_groups_and_extra_attributes
-
-
-Basic CRUD operations on security groups
-----------------------------------------
-
-| tempest.api.network.test_security_groups.SecGroupTest.test_create_list_update_show_delete_security_group
-| tempest.api.network.test_security_groups.SecGroupTest.test_create_security_group_rule_with_additional_args
-| tempest.api.network.test_security_groups.SecGroupTest.test_create_security_group_rule_with_icmp_type_code
-| tempest.api.network.test_security_groups.SecGroupTest.test_create_security_group_rule_with_protocol_integer_value
-| tempest.api.network.test_security_groups.SecGroupTest.test_create_security_group_rule_with_remote_group_id
-| tempest.api.network.test_security_groups.SecGroupTest.test_create_security_group_rule_with_remote_ip_prefix
-| tempest.api.network.test_security_groups.SecGroupTest.test_create_show_delete_security_group_rule
-| tempest.api.network.test_security_groups.SecGroupTest.test_list_security_groups
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_create_additional_default_security_group_fails
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_create_duplicate_security_group_rule_fails
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_create_security_group_rule_with_bad_ethertype
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_create_security_group_rule_with_bad_protocol
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_create_security_group_rule_with_bad_remote_ip_prefix
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_create_security_group_rule_with_invalid_ports
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_create_security_group_rule_with_non_existent_remote_groupid
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_create_security_group_rule_with_non_existent_security_group
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_delete_non_existent_security_group
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_show_non_existent_security_group
-| tempest.api.network.test_security_groups_negative.NegativeSecGroupTest.test_show_non_existent_security_group_rule
-
-
----------------------------------
-Test Area VIM Operations - Volume
----------------------------------
-
-Volume attach and detach operations with the Cinder v2 API
-----------------------------------------------------------
-
-| tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_attach_detach_volume_to_instance
-| tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_get_volume_attachment
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_attach_volumes_with_nonexistent_volume_id
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_detach_volumes_with_invalid_volume_id
-
-
-Volume service availability zone operations with the Cinder v2 API
-------------------------------------------------------------------
-
-| tempest.api.volume.test_availability_zone.AvailabilityZoneV2TestJSON.test_get_availability_zone_list
-
-
-Volume cloning operations with the Cinder v2 API
-------------------------------------------------
-
-| tempest.api.volume.test_volumes_get.VolumesV2GetTest.test_volume_create_get_update_delete_as_clone
-
-
-Image copy-to-volume operations with the Cinder v2 API
-------------------------------------------------------
-
-| tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_volume_bootable
-| tempest.api.volume.test_volumes_get.VolumesV2GetTest.test_volume_create_get_update_delete_from_image
-
-
-Volume creation and deletion operations with the Cinder v2 API
---------------------------------------------------------------
-
-| tempest.api.volume.test_volumes_get.VolumesV2GetTest.test_volume_create_get_update_delete
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_invalid_size
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_nonexistent_source_volid
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_nonexistent_volume_type
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_without_passing_size
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_size_negative
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_size_zero
-
-
-Volume service extension listing operations with the Cinder v2 API
-------------------------------------------------------------------
-
-| tempest.api.volume.test_extensions.ExtensionsV2TestJSON.test_list_extensions
-
-
-Volume GET operations with the Cinder v2 API
---------------------------------------------
-
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_get_invalid_volume_id
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_get_volume_without_passing_volume_id
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_volume_get_nonexistent_volume_id
-
-
-Volume listing operations with the Cinder v2 API
-------------------------------------------------
-
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_by_name
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_details_by_name
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_param_display_name_and_status
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_with_detail_param_display_name_and_status
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_with_detail_param_metadata
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_with_details
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_with_param_metadata
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volumes_list_by_availability_zone
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volumes_list_by_status
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volumes_list_details_by_availability_zone
-| tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volumes_list_details_by_status
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_list_volumes_detail_with_invalid_status
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_list_volumes_detail_with_nonexistent_name
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_list_volumes_with_invalid_status
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_list_volumes_with_nonexistent_name
-| tempest.api.volume.v2.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_details_pagination
-| tempest.api.volume.v2.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_details_with_multiple_params
-| tempest.api.volume.v2.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_pagination
-
-
-Volume metadata operations with the Cinder v2 API
--------------------------------------------------
-
-| tempest.api.volume.test_volume_metadata.VolumesV2MetadataTest.test_crud_volume_metadata
-| tempest.api.volume.test_volume_metadata.VolumesV2MetadataTest.test_update_volume_metadata_item
-
-
-Verification of read-only status on volumes with the Cinder v2 API
-------------------------------------------------------------------
-
-| tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_volume_readonly_update
-
-
-Volume reservation operations with the Cinder v2 API
-----------------------------------------------------
-
-| tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_reserve_unreserve_volume
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_reserve_volume_with_negative_volume_status
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_reserve_volume_with_nonexistent_volume_id
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_unreserve_volume_with_nonexistent_volume_id
-
-
-Volume snapshot creation/deletion operations with the Cinder v2 API
--------------------------------------------------------------------
-
-| tempest.api.volume.test_snapshot_metadata.SnapshotV2MetadataTestJSON.test_crud_snapshot_metadata
-| tempest.api.volume.test_snapshot_metadata.SnapshotV2MetadataTestJSON.test_update_snapshot_metadata_item
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_nonexistent_snapshot_id
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_delete_invalid_volume_id
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_delete_volume_without_passing_volume_id
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_volume_delete_nonexistent_volume_id
-| tempest.api.volume.test_volumes_snapshots.VolumesV2SnapshotTestJSON.test_snapshot_create_get_list_update_delete
-| tempest.api.volume.test_volumes_snapshots.VolumesV2SnapshotTestJSON.test_volume_from_snapshot
-| tempest.api.volume.test_volumes_snapshots_list.VolumesV2SnapshotListTestJSON.test_snapshots_list_details_with_params
-| tempest.api.volume.test_volumes_snapshots_list.VolumesV2SnapshotListTestJSON.test_snapshots_list_with_params
-| tempest.api.volume.test_volumes_snapshots_negative.VolumesV2SnapshotNegativeTestJSON.test_create_snapshot_with_nonexistent_volume_id
-| tempest.api.volume.test_volumes_snapshots_negative.VolumesV2SnapshotNegativeTestJSON.test_create_snapshot_without_passing_volume_id
-
-
-Volume update operations with the Cinder v2 API
------------------------------------------------
-
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_update_volume_with_empty_volume_id
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_update_volume_with_invalid_volume_id
-| tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_update_volume_with_nonexistent_volume_id
-
-
----------------------------
-Test Area High Availability
----------------------------
-
-Verify high availability of OpenStack controller services
----------------------------------------------------------
-
-| dovetail.ha.tc001.nova-api_service_down
-| dovetail.ha.tc002.neutron-server_service_down
-| dovetail.ha.tc003.keystone_service_down
-| dovetail.ha.tc004.glance-api_service_down
-| dovetail.ha.tc005.cinder-api_service_down
-| dovetail.ha.tc006.cpu_overload
-| dovetail.ha.tc007.disk_I/O_overload
-| dovetail.ha.tc008.load_balance_service_down
-
-----------------------------------------
-Test Area vPing - Basic VNF Connectivity
-----------------------------------------
-
-| dovetail.vping.tc001.userdata
-| dovetail.vping.tc002.ssh
-
-
-
-Optional OVP Test Areas
-========================
-
-
------------------
-Test Area BGP VPN
------------------
-
-Verify association and dissasocitation of node using route targets
-------------------------------------------------------------------
-
-| dovetail.sdnvpn.tc001.subnet_connectivity
-| dovetail.sdnvpn.tc002.tenant_separation
-| dovetail.sdnvpn.tc004.router_association
-| dovetail.sdnvpn.tc008.router_association_floating_ip
-
---------------------------------------------------
-IPv6 Compliance Testing Methodology and Test Cases
---------------------------------------------------
-
-Test Case 1: Create and Delete an IPv6 Network, Port and Subnet
----------------------------------------------------------------
-
-| tempest.api.network.test_networks.BulkNetworkOpsIpV6Test.test_bulk_create_delete_network
-| tempest.api.network.test_networks.BulkNetworkOpsIpV6Test.test_bulk_create_delete_port
-| tempest.api.network.test_networks.BulkNetworkOpsIpV6Test.test_bulk_create_delete_subnet
-
-Test Case 2: Create, Update and Delete an IPv6 Network and Subnet
------------------------------------------------------------------
-
-| tempest.api.network.test_networks.NetworksIpV6Test.test_create_update_delete_network_subnet
-
-Test Case 3: Check External Network Visibility
-----------------------------------------------
-
-| tempest.api.network.test_networks.NetworksIpV6Test.test_external_network_visibility
-
-Test Case 4: List IPv6 Networks and Subnets of a Tenant
--------------------------------------------------------
-
-| tempest.api.network.test_networks.NetworksIpV6Test.test_list_networks
-| tempest.api.network.test_networks.NetworksIpV6Test.test_list_subnets
-
-Test Case 5: Show Information of an IPv6 Network and Subnet
------------------------------------------------------------
-
-| tempest.api.network.test_networks.NetworksIpV6Test.test_show_network
-| tempest.api.network.test_networks.NetworksIpV6Test.test_show_subnet
-
-Test Case 6: Create an IPv6 Port in Allowed Allocation Pools
-------------------------------------------------------------
-
-| tempest.api.network.test_ports.PortsIpV6TestJSON.test_create_port_in_allowed_allocation_pools
-
-Test Case 7: Create an IPv6 Port without Security Groups
---------------------------------------------------------
-
-| tempest.api.network.test_ports.PortsIpV6TestJSON.test_create_port_with_no_securitygroups
-
-Test Case 8: Create, Update and Delete an IPv6 Port
----------------------------------------------------
-
-| tempest.api.network.test_ports.PortsIpV6TestJSON.test_create_update_delete_port
-
-Test Case 9: List IPv6 Ports of a Tenant
-----------------------------------------
-
-| tempest.api.network.test_ports.PortsIpV6TestJSON.test_list_ports
-
-Test Case 10: Show Information of an IPv6 Port
-----------------------------------------------
-
-| tempest.api.network.test_ports.PortsIpV6TestJSON.test_show_port
-
-Test Case 11: Add Multiple Interfaces for an IPv6 Router
---------------------------------------------------------
-
-| tempest.api.network.test_routers.RoutersIpV6Test.test_add_multiple_router_interfaces
-
-Test Case 12: Add and Remove an IPv6 Router Interface with port_id
-------------------------------------------------------------------
-
-| tempest.api.network.test_routers.RoutersIpV6Test.test_add_remove_router_interface_with_port_id
-
-Test Case 13: Add and Remove an IPv6 Router Interface with subnet_id
---------------------------------------------------------------------
-
-| tempest.api.network.test_routers.RoutersIpV6Test.test_add_remove_router_interface_with_subnet_id
-
-Test Case 14: Create, Update, Delete, List and Show an IPv6 Router
-------------------------------------------------------------------
-
-| tempest.api.network.test_routers.RoutersIpV6Test.test_create_show_list_update_delete_router
-
-Test Case 15: Create, Update, Delete, List and Show an IPv6 Security Group
---------------------------------------------------------------------------
-
-| tempest.api.network.test_security_groups.SecGroupIPv6Test.test_create_list_update_show_delete_security_group
-
-Test Case 16: Create, Delete and Show Security Group Rules
-----------------------------------------------------------
-
-| tempest.api.network.test_security_groups.SecGroupIPv6Test.test_create_show_delete_security_group_rule
-
-Test Case 17: List All Security Groups
---------------------------------------
-
-| tempest.api.network.test_security_groups.SecGroupIPv6Test.test_list_security_groups
-
-Test Case 18: IPv6 Address Assignment - Dual Stack, SLAAC, DHCPv6 Stateless
----------------------------------------------------------------------------
-
-| tempest.scenario.test_network_v6.TestGettingAddress.test_dhcp6_stateless_from_os
-
-Test Case 19: IPv6 Address Assignment - Dual Net, Dual Stack, SLAAC, DHCPv6 Stateless
--------------------------------------------------------------------------------------
-
-| tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_dhcp6_stateless_from_os
-
-Test Case 20: IPv6 Address Assignment - Multiple Prefixes, Dual Stack, SLAAC, DHCPv6 Stateless
-----------------------------------------------------------------------------------------------
-
-| tempest.scenario.test_network_v6.TestGettingAddress.test_multi_prefix_dhcpv6_stateless
-
-Test Case 21: IPv6 Address Assignment - Dual Net, Multiple Prefixes, Dual Stack, SLAAC, DHCPv6 Stateless
---------------------------------------------------------------------------------------------------------
-
-| tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_multi_prefix_dhcpv6_stateless
-
-Test Case 22: IPv6 Address Assignment - Dual Stack, SLAAC
----------------------------------------------------------
-
-| tempest.scenario.test_network_v6.TestGettingAddress.test_slaac_from_os
-
-Test Case 23: IPv6 Address Assignment - Dual Net, Dual Stack, SLAAC
--------------------------------------------------------------------
-
-| tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_slaac_from_os
-
-Test Case 24: IPv6 Address Assignment - Multiple Prefixes, Dual Stack, SLAAC
-----------------------------------------------------------------------------
-
-| tempest.scenario.test_network_v6.TestGettingAddress.test_multi_prefix_slaac
-
-Test Case 25: IPv6 Address Assignment - Dual Net, Dual Stack, Multiple Prefixes, SLAAC
---------------------------------------------------------------------------------------
-
-| tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_multi_prefix_slaac
-
-------------------------------------------------------------------------
-Filtering Packets Based on Security Rules and Port Security in Data Path
-------------------------------------------------------------------------
-
-| tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_port_security_macspoofing_port
-| tempest.scenario.test_security_groups_basic_ops.TestSecurityGroupsBasicOps.test_cross_tenant_traffic
-| tempest.scenario.test_security_groups_basic_ops.TestSecurityGroupsBasicOps.test_in_tenant_traffic
-| tempest.scenario.test_security_groups_basic_ops.TestSecurityGroupsBasicOps.test_multiple_security_groups
-| tempest.scenario.test_security_groups_basic_ops.TestSecurityGroupsBasicOps.test_port_security_disable_security_group
-| tempest.scenario.test_security_groups_basic_ops.TestSecurityGroupsBasicOps.test_port_update_new_security_group
-
-------------------------------------------------------------
-Dynamic Network Runtime Operations Through the Life of a VNF
-------------------------------------------------------------
-
-| tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_network_basic_ops
-| tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_hotplug_nic
-| tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_subnet_details
-| tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_update_instance_port_admin_state
-| tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_update_router_admin_state
-
-----------------------------------------------------------------
-Correct Behavior after Common Virtual Machine Life Cycles Events
-----------------------------------------------------------------
-
-| tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario
-| tempest.scenario.test_network_advanced_server_ops.TestNetworkAdvancedServerOps.test_server_connectivity_cold_migration
-| tempest.scenario.test_network_advanced_server_ops.TestNetworkAdvancedServerOps.test_server_connectivity_pause_unpause
-| tempest.scenario.test_network_advanced_server_ops.TestNetworkAdvancedServerOps.test_server_connectivity_reboot
-| tempest.scenario.test_network_advanced_server_ops.TestNetworkAdvancedServerOps.test_server_connectivity_rebuild
-| tempest.scenario.test_network_advanced_server_ops.TestNetworkAdvancedServerOps.test_server_connectivity_resize
-| tempest.scenario.test_network_advanced_server_ops.TestNetworkAdvancedServerOps.test_server_connectivity_stop_start
-| tempest.scenario.test_network_advanced_server_ops.TestNetworkAdvancedServerOps.test_server_connectivity_suspend_resume
-| tempest.scenario.test_server_advanced_ops.TestServerAdvancedOps.test_server_sequence_suspend_resume
-| tempest.scenario.test_server_advanced_ops.TestServerAdvancedOps.test_resize_volume_backed_server_confirm
-| tempest.scenario.test_shelve_instance.TestShelveInstance.test_shelve_instance
-| tempest.scenario.test_shelve_instance.TestShelveInstance.test_shelve_volume_backed_instance
-
-------------------------------------------------------------
-Simple Virtual Machine Resource Scheduling on Multiple Nodes
-------------------------------------------------------------
-
-| tempest.scenario.test_server_multinode.TestServerMultinode.test_schedule_to_all_nodes
-| tempest.api.compute.servers.test_server_group.ServerGroupTestJSON.test_create_delete_multiple_server_groups_with_same_name_policy
-| tempest.api.compute.servers.test_server_group.ServerGroupTestJSON.test_create_delete_server_group_with_affinity_policy
-| tempest.api.compute.servers.test_server_group.ServerGroupTestJSON.test_create_delete_server_group_with_anti_affinity_policy
-| tempest.api.compute.servers.test_server_group.ServerGroupTestJSON.test_list_server_groups
-| tempest.api.compute.servers.test_server_group.ServerGroupTestJSON.test_show_server_group
-
---------------------------------------------------------
-Forwarding Packets Through Virtual Networks in Data Path
---------------------------------------------------------
-
-| tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_mtu_sized_frames
diff --git a/docs/testing/user/certificationworkflow/ApplicationForm.rst b/docs/testing/user/certificationworkflow/ApplicationForm.rst
index aac9a46e..1aa937b0 100644
--- a/docs/testing/user/certificationworkflow/ApplicationForm.rst
+++ b/docs/testing/user/certificationworkflow/ApplicationForm.rst
@@ -2,9 +2,9 @@
.. http://creativecommons.org/licenses/by/4.0
.. (c) OPNFV, Intel Corporation and others.
-=======================================
-OPNFV Verified Program Application Form
-=======================================
+===========================================
+OPNFV Verification Program Application Form
+===========================================
+----------------------------------+--------------------------------------------------------------------------------------------+
diff --git a/docs/testing/user/certificationworkflow/index.rst b/docs/testing/user/certificationworkflow/index.rst
index e0273601..bc43e775 100644
--- a/docs/testing/user/certificationworkflow/index.rst
+++ b/docs/testing/user/certificationworkflow/index.rst
@@ -4,61 +4,66 @@
.. _dovetail-certification_workflow:
-===============================
-OPNFV Verified Program Workflow
-===============================
+============
+OVP Workflow
+============
Introduction
============
-This document provides guidance for prospective participants on how to obtain 'OPNFV Verified'
-status. The OPNFV Verified Program (OVP) is administered by the OPNFV Compliance and Certification
-(C&C) committee.
+This document provides guidance for prospective participants on how to
+obtain "OPNFV Verified" for products and services. The OPNFV
+Verification Program (OVP) is administered by the LF Networking
+Compliance and Verification (C&V) Committee.
-For further information about the workflow and general inquiries about the
-program, please check out the `OVP web portal`_, or contact
-the C&C committee by email address verified@opnfv.org. This email address should be used
-for all communication with the OVP.
+For further information about the workflow and general inquiries about
+the program, please check out the `OVP landing page`_, or contact the
+C&V committee by email address ovp-support@lfnetworking.org. This email address
+should be used for all communication with the OVP.
Step 1: Participation Form Submission
=====================================
-A participant should start the process by submitting an online participation form. The participation
-form can found on the `OVP web portal`_ or directly at `OVP participation form`_ and the
-following information must be provided:
+A participant should start the process by submitting an online
+participation form. The participation form can be found on the `OVP
+landing page`_. Submitters are asked for the following information:
- Organization name
-- Organization website (if public)
-- Product name and/or identifier
-- Product specifications
-- Product public documentation
-- Product categories, choose one: (i) software and hardware (ii) software
- and third party hardware (please specify)
-- Primary contact name, business email, postal address and phone number
- Only the primary contact email address should be used for
- official communication with OPNFV OVP.
-- User ID for OVP web portal
- The OVP web portal supports the Linux Foundation user ID in the current release.
- If a new user ID is needed, visit https://identity.linuxfoundation.org.
-- Location where the verification testing is to be conducted. Choose one:
- (internal vendor lab, third-party lab)
-- If the test is to be conducted by a third-party lab, please specify
- name and contact information of the third-party lab, including email, address and
- phone number.
-- OVP software version for compliance verification
-- Testing date
-
-Once the participation form information is received and in order, an email response will be
-sent to the primary contact with confirmation and information to proceed. The primary contact
-specified in the participation form will be entered into OVP web portal back-end by the program
-administrator and will be permitted to submit results for review on behalf of their organization.
-
-There is no fee at this time for participation in the OVP.
+- Contact name
+- Contact address
+- Contact telephone
+- Contact email
+- Membership status in LF Networking
+- Qualifying Offerings Information
+- Testing Category
+- Product Category
+- User ID
+- Version of Testing Tools
+- Testing Date
+- Testing Location
+- Third-party lab information (if any)
+- Signature
+
+Once the participation form information is received and in order, an
+email response will be sent to the primary contact with confirmation and
+information to proceed. The primary contact specified in the
+participation form will be entered into the corresponding OVP web portal
+back-end by the program administrator and will be permitted to submit
+results for review on behalf of their organization. The OVP NFVI Portal
+can be found here (https://nfvi-verified.lfnetworking.org) and the OVP
+VNF Portal can be found here (https://vnf-verified.lfnetworking.org).
+
+There is no fee at this time for participation in the OVP for LF
+Networking (LFN) members or non-profit organizations. If not a member of
+LFN, please contact `LF Networking`_ for participation fee information.
Step 2: Testing
===============
-The following documents guide testers to prepare the test environment and run tests:
+NFVI Testing
+------------
+
+The following documents guide testers to prepare the NFVI test environment and run tests:
- :ref:`dovetail-system_preparation_guide`
- :ref:`dovetail-test_case_specification`
@@ -67,23 +72,33 @@ The following documents guide testers to prepare the test environment and run te
A unique Test ID is generated by the Dovetail tool for each test run and can only be
submitted to the OVP web portal once.
+VNF Testing
+-----------
+
+The following document guide testers to prepare the environment and run the VNF tests:
+
+- `Deploy ONAP via OOM <https://logs.onap.org/production/vex-yul-ecomp-jenkins-1/doc-elalto-verify-rtd/214/html/submodules/oom.git/docs/oom_quickstart_guide.html>`_
+- `ONAP VNF Test Specifications <https://docs.onap.org/en/elalto/submodules/vnfrqts/testcases.git/docs/index.html>`_
+- :ref:`dovetail-vnf_testers_guide`
+
Step 3: Submitting Test Results
===============================
Users/testers other than the primary contact may use the OVP web portal as a resource to upload,
evaluate and share results in a private manner. Testers can upload the test results to the
-`OVP web portal`_. By default, the results are visible only to the tester who uploaded the data.
+OVP web portal (either NFVI or VNF). By default, the results are visible only to the tester who uploaded the data.
Testers can self-review the test results through the portal until they are ready to ask
for OVP review. They may also add new test results as needed.
Once the tester is satisfied with the test result, the primary contact grants access to the test
-result for OVP review using a 'submit for review' operation via the portal. The test result is
-identified by the unique Test ID and becomes visible to a review group comprised of OPNFV
-community members.
+result for OVP review using a 'submit for review' operation via the portal. During this step,
+a new window may be appeared to ask primary contact to complete the application form online.
+Then the test result is identified by the unique Test ID and becomes visible to a review group
+comprised of OPNFV community members.
When a test result is made visible to the reviewers, the program administrator will ask for
-volunteers from the review group using the verified@opnfv.org email and CC the primary contact
+volunteers from the review group using the ovp-support@lfnetworking.org email and CC the primary contact
email that a review request has been made. The program administrator will supply the Test ID
and owner field (primary contact user ID) to the reviewers to identify the results.
@@ -97,28 +112,27 @@ organizations will not be part of the reviewers.
The primary contact may be asked via email for any missing information or clarification of the
test results. The reviewers will make a determination and recommend compliance or non-compliance
-to the C&C Committee. A positive review requires a minimum of two approvals from two distinct
-organizations without any negative reviews. The program administrator sends an email to OVP/C&C
+to the C&V Committee. A positive review requires a minimum of two approvals from two distinct
+organizations without any negative reviews. The program administrator sends an email to OVP/C&V
emails announcing a positive review. A one week limit is given for issues to be raised. If no
-issue is raised, the C&C Committee approves the result and the program administrator sends an
-email to OVP/C&C emails stating the result is approved.
+issue is raised, the C&V Committee approves the result and the program administrator sends an
+email to OVP/C&V emails stating the result is approved.
Normally, the outcome of the review should be communicated to the primary contact within 10
business days after all required information is in order.
-If a test result is denied, an appeal can be made to the C&C Committee for arbitration.
+If a test result is denied, an appeal can be made to the C&V Committee for arbitration.
Step 5: Grant of Use of Program Marks
=====================================
If an application is approved, further information will be communicated to the primary contact
-on the guidelines of using OVP Program Marks (including OVP logo) and the status of compliance
-for promotional purposes.
+that includes badges and on the guidelines of using OVP Program Marks for marketing and promotional purposes.
.. toctree::
:maxdepth: 2
.. References
-.. _`OVP web portal`: https://verified.opnfv.org
-.. _`OVP participation form`: https://na3.docusign.net/Member/PowerFormSigning.aspx?PowerFormId=e03e78d7-c32e-47d3-8292-350b747a0105
+.. _`OVP landing page`: https://www.lfnetworking.org/ovp/
+.. _`LF Networking`: https://www.lfnetworking.org/membership/join/
diff --git a/docs/testing/user/instructionsperinstaller/index.rst b/docs/testing/user/instructionsperinstaller/index.rst
new file mode 100644
index 00000000..adc20f71
--- /dev/null
+++ b/docs/testing/user/instructionsperinstaller/index.rst
@@ -0,0 +1,14 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV
+
+.. _dovetail-testing_user_guide_per_installer:
+
+======================================
+OVP Testing User Guide per Installer
+======================================
+
+.. toctree::
+ :maxdepth: 2
+
+ testing_instructions_apex.rst
diff --git a/docs/testing/user/instructionsperinstaller/testing_instructions_apex.rst b/docs/testing/user/instructionsperinstaller/testing_instructions_apex.rst
new file mode 100644
index 00000000..fb2e2bdd
--- /dev/null
+++ b/docs/testing/user/instructionsperinstaller/testing_instructions_apex.rst
@@ -0,0 +1,158 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV, Huawei Technologies Co.,Ltd and others.
+
+===========================================================
+Conducting OVP Testing with Dovetail using APEX installer
+===========================================================
+
+Overview
+------------------------------
+The purpose of this document is to give tips for the dovetail deployment
+on APEX installer.
+The general structure of the document is remaining according to the user guide
+document and the APEX related tips will be added under of the respective
+chapter's name.
+
+
+Installing Dovetail
+--------------------
+
+
+Checking the Test Host Readiness
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+Installing Prerequisite Packages on the Test Host
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+Configuring the Test Host Environment
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+In order to run the test scenarios properly and having access to all OS components
+that each scenario needs, the undercloud credentials should be used and copied in the
+docker container along with ssh key.
+
+The environment preparation should be applied on the Test Host environment.
+Therefore, the containers which are going to be used as part of this configuration,
+fetch the information, the files and the rest input from Test Host environment directly
+as part of the Docker command.
+
+Setting up Primary Configuration File
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+Two new environment variables could be introduced in the ``env_config.sh`` file.
+
+
+.. code-block:: bash
+
+ # Set the name of the installer type as environment variable (e.g. apex, fuel, etc)
+ # Optional parameter
+ export INSTALLER_TYPE=xxxx
+
+ # Set the deployed scenario name (e.g. os-sdn-nofeature-noha)
+ # Optional parameter
+ export DEPLOY_SCENARIO=xxxx
+
+For the OS_PASSWORD, OpenStack password from undercloud environment should be used.
+
+
+Configuration for Running Tempest Test Cases (Mandatory)
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+Configuration for Running HA Test Cases (Mandatory)
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+Below is a sample of ``${DOVETAIL_HOME}/pre_config/pod.yaml`` file with
+the required syntax when key_filename is used instead of password is employed
+by the controller.
+Moreover, the 'heat-admin' should be used as user.
+
+.. code-block:: bash
+
+ nodes:
+ -
+ # This can not be changed and must be node0.
+ name: node0
+
+ # This must be Jumpserver.
+ role: Jumpserver
+
+ # This is the instance IP of a node which has ipmitool installed.
+ ip: xx.xx.xx.xx
+
+ # User name of the user of this node. This user **must** have sudo privileges.
+ user: heat-admin
+
+ # Password of the user.
+ #password: root
+ key_filename: /root/.ssh/id_rsa
+
+ -
+ # This can not be changed and must be node1.
+ name: node1
+
+ # This must be controller.
+ role: Controller
+
+ # This is the instance IP of a controller node, which is the haproxy primary node
+ ip: xx.xx.xx.xx
+
+ # User name of the user of this node. This user **must** have sudo privileges.
+ user: heat-admin
+
+ # Password of the user.
+ #password: root
+ key_filename: /root/.ssh/id_rsa
+
+ process_info:
+ -
+ # The default attack process of yardstick.ha.rabbitmq is 'rabbitmq-server'.
+ # Here can be reset to 'rabbitmq'.
+ testcase_name: yardstick.ha.rabbitmq
+ attack_process: rabbitmq
+
+ -
+ # The default attack host for all HA test cases is 'node1'.
+ # Here can be reset to any other node given in the section 'nodes'.
+ testcase_name: yardstick.ha.glance_api
+ attack_host: node2
+
+
+
+Configuration of Hosts File (Optional)
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+Installing Dovetail on the Test Host
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+Online Test Host
+""""""""""""""""
+
+
+Offline Test Host
+"""""""""""""""""
+
+
+Starting Dovetail Docker
+------------------------
+
+
+Running the OVP Test Suite
+----------------------------
+
+
+Making Sense of OVP Test Results
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+OVP Portal Web Interface
+------------------------
+
+
+Updating Dovetail or a Test Suite
+---------------------------------
diff --git a/docs/testing/user/instructionsperinstaller/testing_instructions_xci.rst b/docs/testing/user/instructionsperinstaller/testing_instructions_xci.rst
new file mode 100644
index 00000000..017441f2
--- /dev/null
+++ b/docs/testing/user/instructionsperinstaller/testing_instructions_xci.rst
@@ -0,0 +1,220 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV, Huawei Technologies Co.,Ltd and others.
+
+===========================================================
+Conducting OVP Testing with Dovetail using XCI installer
+===========================================================
+
+Overview
+------------------------------
+The purpose of this document is to give tips for the dovetail deployment
+on XCI installer.
+The general structure of the document is remaining according to the user guide
+document and the XCI related tips will be added under of the respective
+chapter's name.
+
+In order to deploy properly the XCI installer the below steps should be followed:
+
+1- The prerequisites of chapter 2.4.1 of XCI User Guide [1] should be applied.
+
+2- If you don’t have one already, generate an SSH key in $HOME/.ssh
+ ssh-keygen -t rsa
+
+3- Clone OPNFV releng-xci repository
+
+4- Change into directory where the sandbox script is located:
+ cd releng-xci/xci
+
+5- Set the sandbox flavor, OPNFV scenario, openstack version, VM size and releng_xci and bifrost versions:
+
+ export INSTALLER_TYPE=osa
+ export XCI_FLAVOR=xxx ,chapter 2.3 of XCI User Guide [1]
+ (e.g. export XCI_FLAVOR=mini)
+ export DEPLOY_SCENARIO=yyy
+ (e.g. export DEPLOY_SCENARIO=os-nosdn-nofeature)
+
+6- Execute the sandbox script
+ ./xci-deploy.sh
+
+Once the deployement is successfully completed, the instructions below should be completed:
+
+1- You should access the OPNFV VM using ssh (ssh root@192.168.122.2)
+
+2- export DEPLOY_SCENARIO=yyy
+
+3- export PDF=/root/releng-xci/xci/var/pdf.yml
+ export IDF=/root/releng-xci/xci/var/idf.yml
+ source openrc
+
+4- Run the following ansible playbook script:
+ ansible-playbook -i releng-xci/xci/playbooks/dynamic_inventory.py releng-xci/xci/playbooks/prepare-tests.yml
+
+5- Run the following bash script:
+ ./prepare-tests.sh
+
+
+
+[1] https://docs.opnfv.org/en/latest/infrastructure/xci.html
+
+
+Installing Dovetail
+--------------------
+
+
+Checking the Test Host Readiness
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+Installing Prerequisite Packages on the Test Host
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+Configuring the Test Host Environment
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+In order to run the test scenarios properly and having access to all OS components
+that each scenario needs, the undercloud credentials should be used and copied in the
+docker container along with ssh key.
+
+The environment preparation should be applied on the Test Host environment.
+Therefore, the containers which are going to be used as part of this configuration,
+fetch the information, the files and the rest input from Test Host environment directly
+as part of the Docker command.
+
+Setting up Primary Configuration File
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+Two new environment variables could be introduced in the ``env_config.sh`` file.
+
+
+.. code-block:: bash
+
+ # For XCI installer the following environment parameters should be added in
+ # this file. Otherwise, those parameters could be ignored.
+ export INSTALLER_TYPE=osa
+ export DEPLOY_SCENARIO=os-nosdn-nofeature
+ export XCI_FLAVOR=noha
+
+The OS_PASSWORD and the rest credential details could be retrieved directly by openrc file in the OPNFV VM.
+
+Configuration for Running Tempest Test Cases (Mandatory)
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+In order for Tempest Test cases to run properly, the ``$DOVETAIL_HOME/pre_config/tempest_conf.yaml``
+file should be updated, introducing the following configuration.
+
+service_available:
+ cinder: True
+
+Configuration for Running HA Test Cases (Mandatory)
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+Below is a sample of ``${DOVETAIL_HOME}/pre_config/pod.yaml`` file with
+the required syntax when key_filename is used instead of password is employed
+by the controller.
+Moreover, the 'heat-admin' should be used as user.
+
+.. code-block:: bash
+
+ nodes:
+ -
+ # This can not be changed and must be node0.
+ name: node0
+
+ # This must be Jumpserver.
+ role: Jumpserver
+
+ # This is the instance IP of a node which has ipmitool installed.
+ ip: xx.xx.xx.xx
+
+ # User name of the user of this node. This user **must** have sudo privileges.
+ user: root
+
+ # Password of the user.
+ #password: root
+ key_filename: /root/.ssh/id_rsa
+
+ -
+ # This can not be changed and must be node1.
+ name: node1
+
+ # This must be controller.
+ role: Controller
+
+ # This is the instance IP of a controller node, which is the haproxy primary node
+ ip: xx.xx.xx.xx
+
+ # User name of the user of this node. This user **must** have sudo privileges.
+ user: root
+
+ # Password of the user.
+ #password: root
+ key_filename: /root/.ssh/id_rsa
+
+ -
+ # This can not be changed and must be node2.
+ name: node2
+
+ # This must be Compute.
+ role: Compute
+
+ # This is the instance IP of a controller node, which is the haproxy primary node
+ ip: xx.xx.xx.xx
+
+ # User name of the user of this node. This user **must** have sudo privileges.
+ user: root
+
+ # Password of the user.
+ #password: root
+ key_filename: /root/.ssh/id_rsa
+
+ process_info:
+ -
+ # The default attack process of yardstick.ha.rabbitmq is 'rabbitmq-server'.
+ # Here can be reset to 'rabbitmq'.
+ testcase_name: yardstick.ha.rabbitmq
+ attack_process: rabbitmq
+
+ -
+ # The default attack host for all HA test cases is 'node1'.
+ # Here can be reset to any other node given in the section 'nodes'.
+ testcase_name: yardstick.ha.glance_api
+ attack_host: node2
+
+Note: The definition for each active controller and computer should be done in this file.
+
+Configuration of Hosts File (Optional)
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+Installing Dovetail on the Test Host
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+Online Test Host
+""""""""""""""""
+
+
+Offline Test Host
+"""""""""""""""""
+
+
+Starting Dovetail Docker
+------------------------
+
+
+Running the OVP Test Suite
+----------------------------
+
+
+Making Sense of OVP Test Results
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+
+OVP Portal Web Interface
+------------------------
+
+
+Updating Dovetail or a Test Suite
+---------------------------------
diff --git a/docs/testing/user/ovpaddendum/exemption-strict-API-validation.rst b/docs/testing/user/ovpaddendum/exemption-strict-API-validation.rst
new file mode 100644
index 00000000..27173ed9
--- /dev/null
+++ b/docs/testing/user/ovpaddendum/exemption-strict-API-validation.rst
@@ -0,0 +1,185 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International
+.. License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) Ericsson and others
+
+.. _dovetail-exemption_process_api_response_validation:
+
+==========================================
+Disabling Strict API Validation in Tempest
+==========================================
+
+.. toctree::
+ :maxdepth: 2
+
+
+Introduction
+============
+
+In 2015, the OpenStack QA team introduced a validation mechanism for Nova API
+responses in Tempest [1]_ with the goal of enforcing Nova API micro-versions.
+The API validation mechanism verifies that API responses only contain data
+elements (properties) as explicitly defined in API response schemas [2]_. In
+case additional data elements are found in Nova API responses, the
+corresponding tests fail immediately without asserting whether or not the
+particular API operation actually succeeded or not.
+
+Independently, cloud vendors have extended their commercial OpenStack cloud
+implementations with additional functionality which requires API extensions.
+Consequently, such cloud implementations do not pass Tempest tests which
+validate API responses despite actually implementing and providing the tested
+functionality.
+
+This document describes an exemption process for use within the OPNFV Verification
+Program which
+
+i) allows vendors to pass Tempest tests if the tested functionality is
+ fully supported despite the presence of additional data elements in API
+ responses, and
+
+ii) makes the application of the exemption process transparently visible in
+ test results.
+
+
+Background and benefits for OVP
+===============================
+
+Vendors of commercial NFV products have extended OpenStack to provide
+additional (NFV) functionality to their customers and to fill functional gaps
+in OpenStack. These add-ons potentially extend the OpenStack API in two ways:
+
+i) new API endpoints and
+
+ii) additional attributes returned by existing API endpoints.
+
+New API endpoints typically go unnoticed by OpenStack Tempest tests and hence
+do not interfere with existing tests. In contrast, (Nova) Tempest tests
+actively validate the responses returned by existing API endpoints against
+pre-defined schemas. An API response is considered invalid if additional
+attributes are present (see example below). Hence, this particular type of
+functional extension of OpenStack causes existing Tempest tests to fail,
+irrespective of whether or not the functionality which is supposed to be tested
+is actually available. As a result, a Tempest test failing due to extended API
+responses does not provide information about whether the tested functionality
+is available or not.
+
+The OPNFV Verification Program has inherited the policy to strictly validate API
+responses from OpenStack by including a selection of Tempest tests in its
+compliance test suite. However, it was never discussed if OVP should adopt this
+policy as well. It turns out that this policy causes challenges for vendors of
+commercial NFV offerings to pass the OVP test suite. The exemption process
+outlined in this document aims at allowing to selectively disable strict API
+response validation in order to enable vendors to adopt OVP **if** the tested
+functionality is supported.
+
+It must be clearly understood that this exemption targets **only** the scenario
+in which additional attributes are included in API responses. It does not
+provide a loophole for passing OVP tests if the OpenStack APIs have been
+altered significantly as this is in conflict with the objective of OVP to
+create industry alignment.
+
+In conclusion, the exemption process described here is deemed beneficial for
+both the broader industry as well as for OVP: Enabling adoption of OVP by
+vendors which extended OpenStack API responses facilitates adoption of OVP in
+the industry. The limited validity period of an exemption incentivizes eventual
+alignment within the industry around a clearly specified set of APIs.
+
+
+Example: additional attributes per VM for HA policy
+---------------------------------------------------
+
+This fictional example showcases the presence of an additional attribute in an
+API response. The example shows that the 'server details' [3]_, i.e. the VM
+metadata, includes an additional attribute 'ha-policy' which is used to
+associate high-availability policies with a VM instance. This attribute is
+utilized by a proprietary add-on component to manage VM migration and recovery
+in case of compute host failures::
+
+ {
+ "server": {
+ "accessIPv4": "1.2.3.4",
+ "config_drive": "",
+ "flavor": {...},
+ "image": {...},
+ "ha_policy": "migrate" <-- additional attribute
+ "name": "new-server-test",
+ "status": "ACTIVE"
+ }
+ }
+
+
+
+Precedent in OpenStack
+======================
+
+In the OpenStack community, the OpenStack Interoperability Working Group
+(Interop WG) [4]_ is maintaining multiple API interoperability compliance
+programs [5]_. These programs utilize Tempest-based tests to determine if a
+given commercial cloud is compliant to a selected set of OpenStack APIs. After
+introduction of the strict API response validation, various cloud products
+which previously passed the compliance program failed validation because of the
+reasons outlined above.
+
+In order to mitigate this situation, the Interop WG consulted with the broader
+OpenStack community [6]_ and eventually introduced an "additional properties
+waiver" for its API compliance programs in July 2016. The waiver was
+created with a clearly defined validity period, covering roughly one year -
+equivalent to three iterations of interoperability guidelines (2015.07,
+2016.01, and 2016.08). The limited lifetime of the waiver was intended to give
+cloud product vendors a transition period for adapting their products to
+achieve full API compliance by the end of the exemption period. All details of
+the waiver are listed in [7]_. Finally, the waiver was officially canceled in
+October 2017 [8]_ after about 15 months.
+
+
+Exemption process for additional properties in API responses in the OVP
+=======================================================================
+
+The details of the exemption process for disabling strict validation of API
+responses is as follows:
+
+#. The Dovetail tool provides a new command line option "--no-api-validation" for
+ disabling strict API validation. This option needs to be explicitly given on
+ the command line to disable strict API validation. If this command line
+ option is omitted, the default behavior (i.e., strict API validation) is
+ applied.
+
+#. The test results created by the Dovetail tool includes an explicit print-out
+ stating if strict API validation was disabled during the test run or not.
+
+#. The OVP portal reads the uploaded result files and indicates for all
+ uploaded test results if strict API validation was disabled or not.
+
+#. Together with the application for certification, a company can request an
+ exemption from the requirement for strict API response checking. A rationale
+ for requesting the exemption has to be provided. The request is either
+ granted or rejected by the C&C committee. The rationale provided must
+ establish that the need for exemption is not in violation of the OVP's
+ objectives.
+
+#. Compliance badges obtained under exemption are valid for one year.
+
+#. OPNFV expects OVP participants to aim for full compliance without requiring
+ exemptions as soon as possible. Hence, an exemption can only be requested
+ twice for the same product (addressing new versions of OVP or new versions
+ of the product).
+
+#. The same logo will be used regardless of being obtained under exemption or
+ not.
+
+#. The exemption will be made available to participants of OVP as part of a
+ service release of OVP 2018.01, 2018.09 and 2019.12.
+
+#. The C&C committee will monitor the situation around exemptions and may
+ decide changes to the above process at any time, including the possibility
+ to stop issuing exemptions.
+
+
+.. [1] https://review.opendev.org/gitweb?p=openstack%2Ftempest.git;a=commitdiff;h=f0c30bc241e5160e3fe7402e738ea8f56a8b1315
+.. [2] https://github.com/openstack/tempest/tree/master/tempest/lib/api_schema/response/compute
+.. [3] https://docs.openstack.org/api-ref/compute/#show-server-details
+.. [4] https://wiki.openstack.org/wiki/Governance/InteropWG
+.. [5] https://www.openstack.org/brand/interop/
+.. [6] http://lists.openstack.org/pipermail/openstack-dev/2016-June/097349.html
+.. [7] https://review.opendev.org/gitweb?p=openstack%2Finterop.git;a=commitdiff;h=c38e18b343505f16a74a97b748362fa7f1a01e57
+.. [8] https://review.opendev.org/gitweb?p=openstack%2Finterop.git;a=commitdiff;h=5748c296a658cf5efebc16ad9d7644ca1125b073
diff --git a/docs/testing/user/ovpaddendum/index.rst b/docs/testing/user/ovpaddendum/index.rst
index a78d1997..811c2bcd 100644
--- a/docs/testing/user/ovpaddendum/index.rst
+++ b/docs/testing/user/ovpaddendum/index.rst
@@ -3,9 +3,11 @@
.. http://creativecommons.org/licenses/by/4.0
.. (c) Intel and others
-====================================================================
-OPNFV Verified Program - 2018.01 Guidelines Addendum for Danube
-====================================================================
+.. _dovetail-ovp-addendum:
+
+=======================================
+Guidelines Addendum for 2019.12 release
+=======================================
.. toctree::
:maxdepth: 2
@@ -15,56 +17,66 @@ Introduction
============
This addendum provides a high-level description of the testing scope and
-pass/fail criteria used in the OPNFV Verified Program (OVP) for the
-OPNFV Danube release. This information is intended as an overview for OVP
-testers and for the Dovetail Project to help guide test-tool and test-case
-development for the OPNFV Danube release. The Dovetail project is responsible for documenting
-test-case specifications as well as implementing the OVP tool-chain through collaboration
-with the OPNFV testing community. OVP testing focuses on establishing the
-ability of the System Under Test (SUT) to perform NFVI and VIM operations and support
-Service Provider oriented features that ensure manageable, resilient and secure
-networks.
+pass/fail criteria used in the OPNFV Verification Program (OVP) for the 2019.12
+release. This information is intended as an overview for OVP testers and for
+the Dovetail Project to help guide test-tool and test-case development for the
+OVP 2019.12 release. The Dovetail project is responsible for documenting
+test-case specifications as well as implementing the OVP tool-chain through
+collaboration with the OPNFV and ONAP testing communities. OVP testing focuses on
+establishing the ability of the System Under Test (SUT) to perform NFVI and VIM
+operations and support Service Provider oriented features that ensure
+manageable, resilient and secure networks.
Meaning of Compliance
=====================
-OPNFV Compliance indicates adherence of an NFV platform to behaviors defined
-through specific platform capabilities, allowing to prepare, instantiate,
-operate and remove VNFs running on the NFVI. Danube compliance evaluates the
-ability of a platform to support Service Provider network capabilities and
-workloads that are supported in the OPNFV platform as of this release.
-Compliance test cases are designated as compulsory or optional based on the
-maturity of OPNFV capabilities as well as industry expectations. Compulsory
-test cases may for example include NFVI management capabilities whereas tests
-for certain high-availability features may be deemed as optional.
+OPNFV Compliance indicates adherence of an NFV platform and VNF to behaviors
+defined through specific platform capabilities, allowing to prepare, instantiate,
+operate and remove VNFs running on the NFVI. OVP 2019.12 compliance evaluates
+the ability of a platform to support Service Provider network capabilities and
+workloads that are supported in the OPNFV and ONAP platforms as of this release.
+Test cases are designated as compulsory or optional based on the maturity
+of capabilities as well as industry expectations. Compulsory test cases may for
+example include NFVI management capabilities whereas tests for certain
+high-availability features may be deemed as optional.
-Test coverage and pass/fail criteria are
-designed to ensure an acceptable level of compliance but not be so restrictive
-as to disqualify variations in platform implementations, capabilities and features.
+Test coverage and pass/fail criteria are designed to ensure an acceptable level
+of compliance but not be so restrictive as to disqualify variations in platform
+implementations, capabilities and features.
SUT Assumptions
===============
-Assumptions about the System Under Test (SUT) include ...
+Assumptions about the NFVI System Under Test (SUT) for the OVP Infrastructure
+badge include ...
- The minimal specification of physical infrastructure, including controller
- nodes, compute nodes and networks, is defined by the `Pharos specification`_.
+ nodes, compute nodes and networks, is defined for the NFVI by the
+ `Pharos specification`_.
- The SUT is fully deployed and operational, i.e. SUT deployment tools are
out of scope of testing.
+Assumptions about the VNF System Under Test (SUT) for the OVP VNF
+badge include ...
+
+- The VNF templates and disk image(s) file are available, and the disk image(s)
+ have been deployed to the ONAP Cloud Site.
+
+- The required value for the VNF pre-load files are available for the selected
+ ONAP Cloud Site.
Scope of Testing
================
-The `OVP Governance Guidelines`_, as approved by the Board of Directors, outlines
-the key objectives of the OVP as follows:
+The `OVP Governance Guidelines`_, as approved by the Board of Directors,
+outlines the key objectives of the OVP as follows:
- Help build the market for
- - OPNFV based infrastructure
+ - LFN based infrastructure
- applications designed to run on that infrastructure
@@ -76,69 +88,75 @@ the key objectives of the OVP as follows:
- Enhance interoperability
The guidelines further directs the scope to be constrained to "features,
-capabilities, components, and interfaces included in an OPNFV release that are
-generally available in the industry (e.g., through adoption by an upstream
-community)", and that compliance verification is evaluated using "functional tests
-that focus on defined interfaces and/or behaviors without regard to the
-the implementation of the underlying system under test".
-
-OPNFV provides a broad range of capabilities, including the reference platform itself
-as well as tools-chains and methodologies for building infrastructures, and
-deploying and testing the platform.
-Not all these aspects are in scope for OVP and not all functions and
-components are tested in the initial version of OVP. For example, the deployment tools
-for the SUT and CI/CD toolchain are currently out of scope.
-Similarly, performance benchmarking related testing is also out of scope or
-for further study. Newer functional areas such as MANO (outside of APIs in the NFVI and
-VIM) are still developing and are for future considerations.
-
+capabilities, components, and interfaces included in an OPNFV and ONAP releases
+that are generally available in the industry (e.g., through adoption by an upstream
+community)", and that compliance verification is evaluated using "functional
+tests that focus on defined interfaces and/or behaviors without regard to the
+implementation of the underlying system under test".
+
+OPNFV provides a broad range of capabilities, including the reference platform
+itself as well as tools-chains and methodologies for building infrastructures,
+and deploying and testing the platform. Not all these aspects are in scope for
+OVP and not all functions and components are tested in the initial versions of
+OVP. For example, the deployment tools for the SUT and CI/CD toolchain are
+currently out of scope. Similarly, performance benchmarking related testing is
+also out of scope or for further study. Newer functional areas such as MANO
+(outside of APIs in the NFVI and VIM) are still developing and are for future
+considerations.
+
+ONAP provides a comprehensive platform for real-time, policy-driven orchestration
+and automation of physical and virtual network functions that will enable software,
+network, IT and cloud providers and developers to rapidly automate new services and
+support complete lifecycle management. By unifying member resources, ONAP is
+accelerating the development of a vibrant ecosystem around a globally shared
+architecture and implementation for network automation–with an open standards focus–
+faster than any one product could on its own.
General Approach
----------------
-In order to meet the above objectives for OVP, we aim to follow a general approach
-by first identifying the overall requirements for all stake-holders,
-then analyzing what OPNFV and the upstream communities can effectively test and verify
-presently to derive an initial working scope for OVP, and to recommend what the
-community should strive to achieve in future releases.
+In order to meet the above objectives for OVP, we aim to follow a general
+approach by first identifying the overall requirements for all stake-holders,
+then analyzing what OPNFV and the upstream communities can effectively test and
+verify presently to derive an initial working scope for OVP, and to recommend
+what the community should strive to achieve in future releases.
The overall requirements for OVP can be categorized by the basic cloud
-capabilities representing common operations needed by basic VNFs, and additional
-requirements for VNFs that go beyond the common cloud capabilities including
-functional extensions, operational capabilities and additional carrier grade
-requirements.
+capabilities representing common operations needed by basic VNFs, and
+additional requirements for VNFs that go beyond the common cloud capabilities
+including functional extensions, operational capabilities and additional
+carrier grade requirements.
For the basic NFV requirements, we will analyze the required test cases,
-leverage or improve upon existing test cases in OPNFV projects
-and upstream projects whenever we can, and bridge the gaps when we must, to meet
-these basic requirements.
+leverage or improve upon existing test cases in OPNFV projects and upstream
+projects whenever we can, and bridge the gaps when we must, to meet these basic
+requirements.
We are not yet ready to include compliance requirements for capabilities such
as hardware portability, carrier grade performance, fault management and other
-operational features, security, MANO and VNF verification. These areas are
+operational features, security, MANO and VNF verification. These areas are
being studied for consideration in future OVP releases.
-In some areas, we will start with a limited level of verification
-initially, constrained by what community resources are able to support at this
-time, but still serve a basic need that is not being fulfilled elsewhere.
-In these areas, we bring significant value to the community we
-serve by starting a new area of verification, breaking new ground and
-expanding it in the future.
+In some areas, we will start with a limited level of verification initially,
+constrained by what community resources are able to support at this time, but
+still serve a basic need that is not being fulfilled elsewhere. In these
+areas, we bring significant value to the community we serve by starting a new
+area of verification, breaking new ground and expanding it in the future.
-In other areas, the functions being verified have yet to reach
-wide adoption but are seen as important requirements in NFV,
-or features are only needed for specific NFV use cases but
-an industry consensus about the APIs and behaviors is still deemed beneficial. In such
-cases, we plan to incorporate the test areas as optional. An optional test
-area will not have to be run or passed in order to achieve compliance.
-Optional tests provide an opportunity for vendors to demonstrate compliance with specific OPNFV
-features beyond the mandatory test scope.
+In other areas, the functions being verified have yet to reach wide adoption
+but are seen as important requirements in NFV, or features are only needed for
+specific NFV use cases but an industry consensus about the APIs and behaviors
+is still deemed beneficial. In such cases, we plan to incorporate the test
+areas as optional. An optional test area will not have to be run or passed in
+order to achieve compliance. Optional tests provide an opportunity for vendors
+to demonstrate compliance with specific OPNFV features beyond the mandatory
+test scope.
Analysis of Scope
-----------------
-In order to define the scope of the Danube-release of the compliance and
+In order to define the scope of the 2019.12 release of the compliance and
verification program, this section analyzes NFV-focused platform capabilities
with respect to the high-level objectives and the general approach outlined in
the previous section. The analysis determines which capabilities are suitable
@@ -148,20 +166,20 @@ addressed in future releases.
1. Basic Cloud Capabilities
The intent of these tests is to verify that the SUT has the required
-capabilities that a basic VNF needs, and these capabilities are implemented
-in a way that enables this basic VNF to run on any OPNFV compliant
-deployment.
-
-A basic VNF can be thought of as a single virtual machine that is networked
-and can perform the simplest network functions, for example, a simple forwarding
-gateway, or a set of such virtual machines connected only by simple virtual network
-services. Running such basic VNF leads to a set of common requirements, including:
-
-- image management (Refstack testing Glance API)
-- identity management (Refstack testing Keystone Identity API)
-- virtual compute (Refstack testing Nova Compute API)
-- virtual storage (Refstack testing Cinder API)
-- virtual networks (Refstack testing Neutron Network API)
+capabilities that a basic VNF needs, and these capabilities are implemented in
+a way that enables this basic VNF to run on any OPNFV compliant deployment.
+
+A basic VNF can be thought of as a single virtual machine that is networked and
+can perform the simplest network functions, for example, a simple forwarding
+gateway, or a set of such virtual machines connected only by simple virtual
+network services. Running such basic VNF leads to a set of common requirements,
+including:
+
+- image management (testing Glance API)
+- identity management (testing Keystone Identity API)
+- virtual compute (testing Nova Compute API)
+- virtual storage (testing Cinder API)
+- virtual networks (testing Neutron Network API)
- forwarding packets through virtual networks in data path
- filtering packets based on security rules and port security in data path
- dynamic network runtime operations through the life of a VNF (e.g. attach/detach,
@@ -170,49 +188,45 @@ services. Running such basic VNF leads to a set of common requirements, includin
suspend/resume, reboot, migrate)
- simple virtual machine resource scheduling on multiple nodes
-OPNFV mainly supports OpenStack as the VIM up to the Danube release. The
-VNFs used in the OVP program, and features in scope for the program which are
+OPNFV mainly supports OpenStack as the VIM up to the 2019.12 release. The VNFs
+used in the OVP NFVI program, and features in scope for the program which are
considered to be basic to all VNFs, require commercial OpenStack distributions
-to support a common basic level of cloud capabilities, and to be compliant
-to a common specification for these capabilities. This requirement significantly
-overlaps with OpenStack community's Interop working group's goals, but they are not
-identical. The OVP runs the OpenStack Refstack-Compute test cases to verify
-compliance to the basic common API requirements of cloud
-management functions and VNF (as a VM) management for OPNFV.
-Additional NFV specific requirements are added in network data path validation,
-packet filtering by security group rules and port security, life cycle runtime events of
-virtual networks, multiple networks in a topology, validation
-of VNF's functional state after common life-cycle events including reboot, pause,
-suspense, stop/start and cold migration. In addition, the
-basic requirement also verifies that the SUT can allocate VNF resources based
-on simple anti-affinity rules.
+to support a common basic level of cloud capabilities, and to be compliant to a
+common specification for these capabilities. This requirement significantly
+overlaps with OpenStack community's Interop working group's goals, but they are
+not identical. The OVP runs the OpenStack Refstack-Compute test cases to verify
+compliance to the basic common API requirements of cloud management functions
+and VNF (as a VM) management for OPNFV. Additional NFV specific requirements
+are added in network data path validation, packet filtering by security group
+rules and port security, life cycle runtime events of virtual networks,
+multiple networks in a topology, validation of VNF's functional state after
+common life-cycle events including reboot, pause, suspense, stop/start and cold
+migration. In addition, the basic requirement also verifies that the SUT can
+allocate VNF resources based on simple anti-affinity rules.
The combined test cases help to ensure that these basic operations are always
-supported by a compliant platform and they adhere to
-a common standard to enable portability across OPNFV compliant platforms.
+supported by a compliant platform and they adhere to a common standard to
+enable portability across OPNFV compliant platforms.
2. NFV specific functional requirements
-NFV has functional requirements beyond the basic common cloud
-capabilities, esp. in the networking area. Examples like SDNVPN, IPv6, SFC may
-be considered additional NFV requirements beyond general purpose cloud
-computing. These feature requirements expand beyond common OpenStack (or other
-VIM) requirements. OPNFV OVP will incorporate test cases to verify
-compliance in these areas as they become mature. Because these extensions
-may impose new API demands, maturity and industry adoption is a prerequisite for
-making them a mandatory requirement for OPNFV compliance. At the time of Danube,
-we have not identified a new functional area that is mandatory for OVP.
-In the meantime, OVP
-intends to offer tests in some of these areas as an optional extension of the test
-report to be submitted for review, noting that passing these tests will not be
+NFV has functional requirements beyond the basic common cloud capabilities,
+esp. in the networking area. Examples like BGPVPN, IPv6, SFC may be considered
+additional NFV requirements beyond general purpose cloud computing. These
+feature requirements expand beyond common OpenStack (or other VIM)
+requirements. OPNFV OVP will incorporate test cases to verify compliance in
+these areas as they become mature. Because these extensions may impose new API
+demands, maturity and industry adoption is a prerequisite for making them a
+mandatory requirement for OPNFV compliance. At the time of the 2019.12 release,
+we have promoted tests of the OpenStack IPv6 API from optional to mandatory
+while keeping BGPVPN as optional test area. Passing optional tests will not be
required to pass OPNFV compliance verification.
-SDNVPN is relevant due to the wide adoption of MPLS/BGP based VPNs in wide area
-networks, which makes it necessary for data centers hosting VNFs to be able to
-seamlessly interconnect with such networks. IPv6 is a high priority service provider
-requirement to ease IP addressing and operational issues. SFC is also an important
-NFV requirement, however its implementation has not yet been accepted or adopted
-in the upstream at the time of Danube.
+BGPVPNs are relevant due to the wide adoption of MPLS/BGP based VPNs in wide
+area networks, which makes it necessary for data centers hosting VNFs to be
+able to seamlessly interconnect with such networks. SFC is also an important
+NFV requirement, however its implementation has not yet been accepted or
+adopted in the upstream at the time of the 2019.12 release.
3. High availability
@@ -223,83 +237,77 @@ configurations. The current OPNFV high availability verification focuses on
OpenStack control service failures and resource overloads, and verifies service
continuity when the system encounters such failures or resource overloads, and
also verifies the system heals after a failure episode within a reasonable time
-window. These service HA capabilities are commonly adopted in the industry
-and should be a mandatory requirement.
+window. These service HA capabilities are commonly adopted in the industry and
+should be a mandatory requirement.
The current test cases in HA cover the basic area of failure and resource
overload conditions for a cloud platform's service availability, including all
-of the basic cloud capability services, and basic compute and storage loads,
-so it is a meaningful first step for OVP. We expect additional high availability
+of the basic cloud capability services, and basic compute and storage loads, so
+it is a meaningful first step for OVP. We expect additional high availability
scenarios be extended in future releases.
-4. Resiliency
+4. Stress Testing
-Resiliency testing involves stressing the SUT and verifying its ability
-to absorb stress conditions and still provide an acceptable level of service.
+Resiliency testing involves stressing the SUT and verifying its ability to
+absorb stress conditions and still provide an acceptable level of service.
Resiliency is an important requirement for end-users.
-The OPNFV testing projects have started testing
-OPNFV system resiliency in
-the Danube release that can be used to provide limited coverage in this area.
-However, this is a relatively new test methodology in OPNFV, additional study
-and testing experiences are still needed. We defer the resiliency testing to
-future OVP releases.
+The 2019.12 release of OVP includes a load test which spins up a number of VMs
+pairs in parallel to assert that the system under test can process the workload
+spike in a stable and deterministic fashion.
5. Security
Security is among the top priorities as a carrier grade requirement by the
-end-users. Some of the basic common functions, including virtual network isolation,
-security groups, port security and role based access control are already covered as
-part of the basic cloud capabilities that are verified in OVP. These test cases
-however do not yet cover the basic required security capabilities expected of an end-user
-deployment. It is an area that we should address in the near future, to define
-a common set of requirements and develop test cases for verifying those
-requirements.
-
-Another common requirement is security vulnerability scanning.
-While the OPNFV security project integrated tools for security vulnerability
-scanning, this has not been fully analyzed or exercised in Danube release.
-This area needs further work to identify the required level of security for the
-purpose of OPNFV in order to be integrated into the OVP. End-user inputs on
-specific requirements in security is needed.
+end-users. Some of the basic common functions, including virtual network
+isolation, security groups, port security and role based access control are
+already covered as part of the basic cloud capabilities that are verified in
+OVP. These test cases however do not yet cover the basic required security
+capabilities expected of an end-user deployment. It is an area that we should
+address in the near future, to define a common set of requirements and develop
+test cases for verifying those requirements.
+
+The 2019.12 release includes new test cases which verify that the role-based
+access control (RBAC) functionality of the VIM is behaving as expected.
+
+Another common requirement is security vulnerability scanning. While the OPNFV
+security project integrated tools for security vulnerability scanning, this has
+not been fully analyzed or exercised in 2019.12 release. This area needs
+further work to identify the required level of security for the purpose of
+OPNFV in order to be integrated into the OVP. End-user inputs on specific
+requirements in security is needed.
6. Service assurance
-Service assurance (SA) is a broad area of concern for reliability of the NFVI/VIM
-and VNFs, and depends upon multiple subsystems of an NFV platform for essential
-information and control mechanisms. These subsystems include telemetry, fault management
-(e.g. alarms), performance management, audits, and control mechanisms such as security
-and configuration policies.
+Service assurance (SA) is a broad area of concern for reliability of the
+NFVI/VIM and VNFs, and depends upon multiple subsystems of an NFV platform for
+essential information and control mechanisms. These subsystems include
+telemetry, fault management (e.g. alarms), performance management, audits, and
+control mechanisms such as security and configuration policies.
-The current Danube release implements some enabling capabilities in NFVI/VIM
-such as telemetry, policy, and fault management. However, the specification of expected
-system components, behavior and the test cases to verify them have not yet
-been adequately developed. We will therefore not be testing this area at this time
-but defer to future study.
+The current 2019.12 release implements some enabling capabilities in NFVI/VIM
+such as telemetry, policy, and fault management. However, the specification of
+expected system components, behavior and the test cases to verify them have not
+yet been adequately developed. We will therefore not be testing this area at
+this time but defer to future study.
7. Use case testing
Use-case test cases exercise multiple functional capabilities of a platform in
-order to realize a larger end-to-end scenario. Such end-to-end use cases do
+order to realize a larger end-to-end scenario. Such end-to-end use cases do
not necessarily add new API requirements to the SUT per se, but exercise
-aspects of the SUT's functional capabilities in more complex ways. For
+aspects of the SUT's functional capabilities in more complex ways. For
instance, they allow for verifying the complex interactions among multiple VNFs
-and between VNFs and the cloud platform in a more realistic fashion. End-users
+and between VNFs and the cloud platform in a more realistic fashion. End-users
consider use-case-level testing as a significant tool in verifying OPNFV
compliance because it validates design patterns and support for the types of
NFVI features that users care about.
-There are a lot of projects in OPNFV developing use cases and sample VNFs,
-however most are still in early phase and require further enhancements to
-become useful additions to the OVP. Examples such as vIMS, or those which are
-not yet available in Danube release, e.g. vCPE, will be valuable additions to
-the OVP. These use cases need to be widely accepted, and since they are more
-complex, using these VNFs for OVP demands a higher level of community resources
-to implement, analyze and document these VNFs. Hence, use case testing is not
-ready for OVP at the time of Danube, but can be incorporated in Euphrates or as
-a future roadmap area.
+There are a lot of projects in OPNFV developing use cases and sample VNFs. The
+2019.12 release of OVP features two such use-case tests, spawning and verifying
+a vIMS and a vEPC, correspondingly.
-8. Additional capabilities
+8. Additional NFVI capabilities
In addition to the capabilities analyzed above, there are further system
aspects which are of importance for the OVP. These comprise operational and
@@ -316,55 +324,71 @@ considered widely available in commercial systems in order to include them in
the OVP. Hence, these aspects are left for inclusion in future releases of the
OVP.
-
-
-Scope of the Danube-release of the OVP
---------------------------------------
-
-Summarizing the results of the analysis above, the scope of the Danube-release
-of the OVP is as follows:
-
-- Test Area: Basic cloud capabilities
-
- - **OpenStack interoperability test cases excluding object storage**\*
- - **OPNFV-Functest/vPing, including both user data and ssh**
- - *Port security and security groups*
- - *VM life-cycle events*
- - *VM networking*
- - *VM resource scheduling*
- - *Forwarding packets in the data path*
+9. VNF Compliance
+
+VNF Compliance verifies the VNF template files conform to the requirements documented
+in by ONAP VNFRQTS project.
+
+10. VNF Validation
+
+VNF Validation verifies the VNF is able to onbroad within ONAP and ONAP is able to
+perform basic orchestration operations with the VNF, including instantiating the
+VNF on the Cloud Site.
+
+Scope of the 2019.12 release of the OVP
+---------------------------------------
+
+Summarizing the results of the analysis above, the scope of the 2019.12 release
+of OVP is as follows:
+
+- Mandatory NFVI test scope:
+
+ - functest.vping.userdata
+ - functest.vping.ssh
+ - functest.tempest.osinterop\*
+ - functest.tempest.compute
+ - functest.tempest.identity_v3
+ - functest.tempest.image
+ - functest.tempest.network_api
+ - functest.tempest.volume
+ - functest.tempest.neutron_trunk_ports
+ - functest.tempest.ipv6_api
+ - functest.security.patrole
+ - yardstick.ha.nova_api
+ - yardstick.ha.neutron_server
+ - yardstick.ha.keystone
+ - yardstick.ha.glance_api
+ - yardstick.ha.cinder_api
+ - yardstick.ha.cpu_load
+ - yardstick.ha.disk_load
+ - yardstick.ha.haproxy
+ - yardstick.ha.rabbitmq
+ - yardstick.ha.database
+ - bottlenecks.stress.ping
+
+- Optional NFVI test scope:
+
+ - functest.tempest.ipv6_scenario
+ - functest.tempest.multi_node_scheduling
+ - functest.tempest.network_security
+ - functest.tempest.vm_lifecycle
+ - functest.tempest.network_scenario
+ - functest.tempest.bgpvpn
+ - yardstick.ha.neutron_l3_agent
+ - yardstick.ha.controller_restart
+ - functest.vnf.vims
+ - functest.vnf.vepc
+
+- Mandatory VNF test scope:
+
+ - Refer to `ONAP VNF Test Case Descriptions <https://docs.onap.org/en/elalto/submodules/vnfrqts/testcases.git/docs/index.html>`_
\* The OPNFV OVP utilizes the same set of test cases as the OpenStack
interoperability program *OpenStack Powered Compute*. Passing the OPNFV OVP
-does **not** imply that the SUT is certified according to the *OpenStack Powered
-Compute* program. *OpenStack Powered Compute* is a trademark of the OpenStack
-foundation and the corresponding certification label can only be awarded by the
-OpenStack foundation.
-
-
-
-- Test Area: SDNVPN
-
- - *OPNFV-SDNVPN*
-
-
-- Test Area: IPv6
-
- - *OPNFV-IPv6
- (Limited to overlay tests, v6Ping)*
-
-
-- Test Area: High Availability
-
- - **OPNFV-Yardstick/HA**
- (Limited to service continuity verification of control services)
-
-[Highlighting: **Mandatory test cases**, *Optional test cases*]
-
-
-These tested areas represent significant advancement in the direction to meet
-the OVP's objectives and end-user expectations, and is a good basis for the
-initial phase of OVP.
+does **not** imply that the SUT is certified according to the *OpenStack
+Powered Compute* program. *OpenStack Powered Compute* is a trademark of the
+OpenStack foundation and the corresponding certification label can only be
+awarded by the OpenStack foundation.
Note: The SUT is limited to NFVI and VIM functions. While testing MANO
component capabilities is out of scope, certain APIs exposed towards MANO are
@@ -377,7 +401,7 @@ Scope considerations for future OVP releases
--------------------------------------------
Based on the previous analysis, the following items are outside the scope of
-the Danube release of the CV but are being considered for inclusion in future
+the 2019.12 release of OVP but are being considered for inclusion in future
releases:
- service assurance
@@ -414,6 +438,18 @@ documented and accepted by the reviewers.
Applicants who choose to run the optional test cases can include the results
of the optional test cases to highlight the additional compliance.
+
+Exemption from strict API response validation
+=============================================
+
+Vendors of commercial NFVI products may have extended the Nova API to support
+proprietary add-on features. These additions can cause Nova Tempest API tests
+to fail due to unexpected data in API responses. In order to resolve this
+transparently in the context of OVP, a temporary exemption process has been
+created. More information on the exemption can be found in section
+:ref:`dovetail-exemption_process_api_response_validation`.
+
+
.. References
.. _`OVP Governance Guidelines`: https://www.opnfv.org/wp-content/uploads/sites/12/2018/01/OVP-Governance-Guidelines-1.0.1-012218.pdf
.. _`Pharos specification`: https://wiki.opnfv.org/display/pharos/Pharos+Specification
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_log_files.png b/docs/testing/user/reviewerguide/danube/images/ovp_log_files.png
deleted file mode 100644
index d5af068b..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_log_files.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_log_setup.png b/docs/testing/user/reviewerguide/danube/images/ovp_log_setup.png
deleted file mode 100644
index 395a2c3e..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_log_setup.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_log_test_count.png b/docs/testing/user/reviewerguide/danube/images/ovp_log_test_count.png
deleted file mode 100644
index 6c363441..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_log_test_count.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_missing_ha.png b/docs/testing/user/reviewerguide/danube/images/ovp_missing_ha.png
deleted file mode 100644
index 7e2c6982..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_missing_ha.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_pass_fraction.png b/docs/testing/user/reviewerguide/danube/images/ovp_pass_fraction.png
deleted file mode 100644
index 74c99f37..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_pass_fraction.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_pass_percentage.png b/docs/testing/user/reviewerguide/danube/images/ovp_pass_percentage.png
deleted file mode 100644
index ad79f4ec..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_pass_percentage.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_result_overview.png b/docs/testing/user/reviewerguide/danube/images/ovp_result_overview.png
deleted file mode 100644
index fa12ddc7..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_result_overview.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_result_review.png b/docs/testing/user/reviewerguide/danube/images/ovp_result_review.png
deleted file mode 100644
index a8f988fc..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_result_review.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_run_results.png b/docs/testing/user/reviewerguide/danube/images/ovp_run_results.png
deleted file mode 100644
index 4b845df2..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_run_results.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_test_count.png b/docs/testing/user/reviewerguide/danube/images/ovp_test_count.png
deleted file mode 100644
index 1c9bcd30..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_test_count.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_top_nav.png b/docs/testing/user/reviewerguide/danube/images/ovp_top_nav.png
deleted file mode 100644
index 280d0c5f..00000000
--- a/docs/testing/user/reviewerguide/danube/images/ovp_top_nav.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/sut_endpoints.png b/docs/testing/user/reviewerguide/danube/images/sut_endpoints.png
deleted file mode 100644
index f11d531f..00000000
--- a/docs/testing/user/reviewerguide/danube/images/sut_endpoints.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/sut_info.png b/docs/testing/user/reviewerguide/danube/images/sut_info.png
deleted file mode 100644
index 395f5f41..00000000
--- a/docs/testing/user/reviewerguide/danube/images/sut_info.png
+++ /dev/null
Binary files differ
diff --git a/docs/testing/user/reviewerguide/images/ovp_log_files_functest_image.png b/docs/testing/user/reviewerguide/images/ovp_log_files_functest_image.png
new file mode 100644
index 00000000..66d3fd73
--- /dev/null
+++ b/docs/testing/user/reviewerguide/images/ovp_log_files_functest_image.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/images/ovp_log_setup.png b/docs/testing/user/reviewerguide/images/ovp_log_setup.png
new file mode 100644
index 00000000..f53b94d9
--- /dev/null
+++ b/docs/testing/user/reviewerguide/images/ovp_log_setup.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/images/ovp_pass_fraction.png b/docs/testing/user/reviewerguide/images/ovp_pass_fraction.png
new file mode 100644
index 00000000..30672e02
--- /dev/null
+++ b/docs/testing/user/reviewerguide/images/ovp_pass_fraction.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/images/ovp_pass_percentage.png b/docs/testing/user/reviewerguide/images/ovp_pass_percentage.png
new file mode 100644
index 00000000..1a61f7b4
--- /dev/null
+++ b/docs/testing/user/reviewerguide/images/ovp_pass_percentage.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/images/ovp_result_review.png b/docs/testing/user/reviewerguide/images/ovp_result_review.png
new file mode 100644
index 00000000..56633447
--- /dev/null
+++ b/docs/testing/user/reviewerguide/images/ovp_result_review.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/images/ovp_top_nav.png b/docs/testing/user/reviewerguide/images/ovp_top_nav.png
new file mode 100644
index 00000000..a1c261f8
--- /dev/null
+++ b/docs/testing/user/reviewerguide/images/ovp_top_nav.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_vping_ssh.png b/docs/testing/user/reviewerguide/images/ovp_vping_ssh.png
index 01384d02..01384d02 100644
--- a/docs/testing/user/reviewerguide/danube/images/ovp_vping_ssh.png
+++ b/docs/testing/user/reviewerguide/images/ovp_vping_ssh.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/danube/images/ovp_vping_user.png b/docs/testing/user/reviewerguide/images/ovp_vping_user.png
index a82e1121..a82e1121 100644
--- a/docs/testing/user/reviewerguide/danube/images/ovp_vping_user.png
+++ b/docs/testing/user/reviewerguide/images/ovp_vping_user.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/images/review_status.png b/docs/testing/user/reviewerguide/images/review_status.png
new file mode 100644
index 00000000..911b06fd
--- /dev/null
+++ b/docs/testing/user/reviewerguide/images/review_status.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/images/sut_endpoints.png b/docs/testing/user/reviewerguide/images/sut_endpoints.png
new file mode 100644
index 00000000..d5ad6239
--- /dev/null
+++ b/docs/testing/user/reviewerguide/images/sut_endpoints.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/images/sut_info.png b/docs/testing/user/reviewerguide/images/sut_info.png
new file mode 100644
index 00000000..29c249b2
--- /dev/null
+++ b/docs/testing/user/reviewerguide/images/sut_info.png
Binary files differ
diff --git a/docs/testing/user/reviewerguide/index.rst b/docs/testing/user/reviewerguide/index.rst
index bd615f19..f08ae784 100644
--- a/docs/testing/user/reviewerguide/index.rst
+++ b/docs/testing/user/reviewerguide/index.rst
@@ -2,9 +2,9 @@
.. http://creativecommons.org/licenses/by/4.0
.. (c) Ericsson AB
-=============================================
-OPNFV Verified Program 2018.01 Reviewer Guide
-=============================================
+==================
+OVP Reviewer Guide
+==================
.. toctree::
:maxdepth: 2
@@ -13,194 +13,215 @@ OPNFV Verified Program 2018.01 Reviewer Guide
Introduction
============
-This reviewer guide provides detailed guidance for reviewers on how to handle the result review
-process. Reviewers must follow the checklist below to ensure review consistency for the OPNFV
-Verified Program (OVP) 2018.01 (Danube) release at a minimum.
-
-#. **Mandatory Test Area Results** - Validate that results for all mandatory test areas are present.
-#. **Test-Case Count within Mandatory Test Area** - Check that the total number of test-cases are present in each mandatory test area.
-#. **Test-Case Pass Percentage** - Ensure all tests have passed (100% pass rate).
-#. **Log File Verification** - Inspect the log file for each test area (osinterop, ha, vping).
-#. **SUT Info Verification** - Validate the system under test (SUT) hardware and software endpoint info is present.
+This document provides detailed guidance for reviewers on how to handle the result review
+process.
+The OPNFV Verification Program (OVP) provides the ability for users to upload test results in
+`OVP portal <https://nfvi-verified.lfnetworking.org>`_ and request from OVP community to review them.
-1. Mandatory Test Area Results
-==============================
+OVP administrator will ask for review volunteers using the ovp-support@lfnetworking.org email alias.
+The incoming results for review will be identified by the administrator with particular **Test ID**
+and **Owner** values.
-Validate that results for all mandatory test areas are included in the overall test suite. The
-required mandatory test areas are:
- - **osinterop**
- - **vping**
- - **ha**
+Volunteers that will accept the review request can access the test results by login to the
+`OVP portal <https://nfvi-verified.lfnetworking.org>`_ and then click on the **Incoming Reviews**
+tab in top-level navigation bar.
-Login to the OVP portal at:
+.. image:: images/ovp_top_nav.png
+ :align: center
+ :scale: 100%
-*https://verified.opnfv.org*
-Click on the 'My Results' tab in top-level navigation bar.
+After the user submit for review the test results **Status** is changed from 'private' to 'review'.
+Reviewers can find that the corresponding OVP portal result will have a status of 'review'.
+Also there are **Application** information list here for review. All the application information
+is submitted by users at the same time they submit their results for review. Reviewers can also
+find who has already approve/not approve the test results by clicking on the **View Reviews**.
-.. image:: danube/images/ovp_top_nav.png
+.. image:: images/ovp_result_review.png
:align: center
:scale: 100%
-The OVP administrator will ask for review volunteers using the verified@opnfv.org email alias. The
-incoming results for review will be identified by the administrator with particular 'Test ID'
-and 'Owner' values. The corresponding OVP portal result will have a status of 'review'.
-.. image:: danube/images/ovp_result_review.png
+Reviewers must follow the checklist below to ensure review consistency for the OPNFV
+Verification Program (OVP) 2019.12 (Hunter) release at a minimum.
+
+#. **Test Case Pass Percentage** - Ensure all mandatory tests have passed (100% pass rate).
+#. **Mandatory Test Case Results** - Validate that results for all mandatory test cases are present.
+#. **Log File Verification** - Inspect the log file for each test case.
+#. **SUT Info Verification** - Validate the system under test (SUT) hardware and software endpoint info is present.
+
+
+Test Case Pass Percentage
+=========================
+
+All mandatory test cases have to run successfully. The below figure of the **Test Run Results**
+is one method and shows that 96.71% of the mandatory test cases have passed.
+This value must not be lower than 100%.
+
+.. image:: images/ovp_pass_percentage.png
:align: center
- :scale: 100%
+ :width: 350 px
-In the example above, this information will be provided as:
-- Test ID: a00c47e8
-- Owner: jtaylor
-Click on the hyperlink within the 'Test ID' column.
+Mandatory Test Case Results
+===========================
+
+Test results can be displayed by clicking on the hyperlink under the **Test ID** column.
+Reviewers should validate that results for all mandatory test cases are included in the overall
+test suite. The required mandatory test cases are:
+
+- bottlenecks.stress.ping
+- functest.security.patrole
+- functest.tempest.compute
+- functest.tempest.identity_v3
+- functest.tempest.image
+- functest.tempest.ipv6_api
+- functest.tempest.network_api
+- functest.tempest.neutron_trunk_ports
+- functest.tempest.osinterop
+- functest.tempest.volume
+- functest.vping.ssh
+- functest.vping.userdata
+- yardstick.ha.cinder_api
+- yardstick.ha.cpu_load
+- yardstick.ha.database
+- yardstick.ha.disk_load
+- yardstick.ha.glance_api
+- yardstick.ha.haproxy
+- yardstick.ha.keystone
+- yardstick.ha.neutron_server
+- yardstick.ha.nova_api
+- yardstick.ha.rabbitmq
*Note, that the 'Test ID' column in this view condenses the UUID used for 'Test ID' to
eight characters even though the 'Test ID' is a longer UUID in the back-end.*
-.. image:: danube/images/ovp_result_overview.png
- :align: center
- :scale: 100%
+Failed test cases can be easy identified by the color of pass/total number:
-The 'Test ID' hyperlink toggles the view to a top-level listing of the results displayed above.
-Validate that osinterop, vping and ha test area results are all present within the view.
+- **Green** when all test cases pass
+- **Orange** when at least one fails/skips
+- **Red** when all test cases fail/skip
+.. image:: images/ovp_pass_fraction.png
+ :align: center
+ :width: 350 px
-2. Test-Case Count within Mandatory Test Area
-=============================================
-Validate the test-case count within each test area. For the OVP 2018.01 release, this must break
-down as outlined in the table below.
+Log File Verification
+=====================
-.. image:: danube/images/ovp_test_count.png
- :align: center
- :scale: 100%
+Each log file of the mandatory test cases have to be verified for content.
-In the diagram above (from section 1), these counts can be gleaned from the numbers to the
-right of the test-cases. The total number is given for the osinterop (dovetail.osinterop.tc001)
-test area at 205. The vping (dovetail.vping.tc00x) and ha (dovetail.ha.tc00x) test-cases are
-broken down separately with a line for each test-case. Directly above the 'Test Result Overview'
-listing there's a summary labelled 'Test Run Results' shown below. For OVP 2018.01, a mandatory
-total of **215** test-cases must be present (205 osinterop + 8 ha + 2 vping).
+Log files can be displayed by clicking on the setup icon to the right of the results,
+as shown in the figure below.
-.. image:: danube/images/ovp_missing_ha.png
+.. image:: images/ovp_log_setup.png
:align: center
:scale: 100%
-An example of a listing that should flag a negative review is shown above. The mandatory ha test
-area is missing one test case (dovetail.ha.tc008).
-3. Test-Case Pass Percentage
-============================
+*Note, all log files can be found at results/ directory as shown at the following table.*
-All mandatory test-cases must pass. This can be validated in multiple ways. The below diagram of
-the 'Test Run Results' is one method and shows that 100% of the mandatory test-cases have passed.
-This value must not be lower than 100%.
++------------------------+--------------------------+
+| **Mandatory Test Case**| **Location** |
++------------------------+--------------------------+
+| bottlenecks | results/stress_logs/ |
++------------------------+--------------------------+
+| functest.vping | results/vping_logs/ |
++------------------------+--------------------------+
+| functest.tempest | results/tempest_logs/ |
++------------------------+--------------------------+
+| functest.security | results/security_logs/ |
++------------------------+--------------------------+
+| yardstick | results/ha_logs/ |
++------------------------+--------------------------+
-.. image:: danube/images/ovp_pass_percentage.png
- :align: center
- :width: 350 px
-Another method to check that all mandatory test-cases have passed is shown in the diagram below.
-The pass/total is given as a fraction and highlighted here in yellow. For the osinterop test area,
-the result must display [205/205] and for each of the test-cases under the vping and ha test areas
-[1/1] must be displayed.
+Bottlenecks Logs
+----------------
-.. image:: danube/images/ovp_pass_fraction.png
- :align: center
- :width: 270 px
+It must contain the 'SUCCESS' result at the end of Bottlenecks log as shown in following example:
-4. Log File Verification
-========================
+ 2019-12-03 07:35:14,630 [INFO] yardstick.benchmark.core.task task.py:129 Testcase: "ping_bottlenecks" SUCCESS!!!
-Three log files must be verified for content within each mandatory test area. The log files for
-each of the test areas is noted in the table below.
-.. image:: danube/images/ovp_log_files.png
- :align: center
- :scale: 100%
+Functest Logs
+-------------
+
+There are 2 different types of Functest logs, one is plain text for **vping** test cases and the other
+is html file for **tempest** and **security** test cases.
-The three log files can be displayed by clicking on the setup icon to the right of the results,
-as shown in the diagram below.
+For **vping** test cases, two entries displayed in the tables below must be present in log files.
-*Note, while the vping and ha test areas list multiple test-cases in the below diagram, there is
-a single log file for all test-cases within these test areas.*
+**functest.vping.ssh**
-.. image:: danube/images/ovp_log_setup.png
+.. image:: images/ovp_vping_ssh.png
:align: center
:scale: 100%
-Within the osinterop log (dovetail.osinterop.tc001.log), scroll down to the area of the log that
-begins to list the results of each test-case executed. This can be located by looking for lines
-prefaced with '**tempest.api**' and ending with '**... ok**'.
-.. image:: danube/images/ovp_log_test_count.png
+**functest.vping.userdata**
+
+.. image:: images/ovp_vping_user.png
:align: center
:scale: 100%
-The number of lines within the osinterop log for test-cases must add up according to the table
-above, where test-cases are broken down according to compute, identity, image, network and volume,
-with respective counts given in the table. The ha log (yardstick.log) must contain the 'PASS'
-result for each of the eight test-cases within this test area. This can be verified by searching
-the log for the keyword 'PASS'.
+For **tempest** and **security** test cases, it opens an html page that lists all test cases as shown
+below. All test cases must have run successfully.
+.. image:: images/ovp_log_files_functest_image.png
+ :align: center
+ :scale: 100%
-The eight lines to validate are listed below:
- - 017-10-16 05:07:49,158 yardstick.benchmark.scenarios.availability.serviceha serviceha.py:81
- INFO The HA test case PASS the SLA
- - 2017-10-16 05:08:31,387 yardstick.benchmark.scenarios.availability.serviceha serviceha.py:81
- INFO The HA test case PASS the SLA
- - 2017-10-16 05:09:13,669 yardstick.benchmark.scenarios.availability.serviceha serviceha.py:81
- INFO The HA test case PASS the SLA
- - 2017-10-16 05:09:55,967 yardstick.benchmark.scenarios.availability.serviceha serviceha.py:81
- INFO The HA test case PASS the SLA
- - 2017-10-16 05:10:38,407 yardstick.benchmark.scenarios.availability.serviceha serviceha.py:81
- INFO The HA test case PASS the SLA
- - 2017-10-16 05:11:00,030 yardstick.benchmark.scenarios.availability.scenario_general
- scenario_general.py:71 INFO [92m Congratulations, the HA test case PASS! [0m
- - 2017-10-16 05:11:22,536 yardstick.benchmark.scenarios.availability.scenario_general
- scenario_general.py:71 INFO [92m Congratulations, the HA test case PASS! [0m
- - 2017-10-16 05:12:07,880 yardstick.benchmark.scenarios.availability.scenario_general
- scenario_general.py:71 INFO [92m Congratulations, the HA test case PASS! [0m
+Yardstick Logs
+--------------
+The yardstick log must contain the 'SUCCESS' result for each of the test-cases within this
+test area. This can be verified by searching the log for the keyword 'SUCCESS'.
-The final validation is for the vping test area log file (functest.log). The two entries
-displayed in the diagrams below must be present in this log file.
+An example of a FAILED and a SUCCESS test case are listed below:
- - vping_userdata
- - vping_ssh
+ 2018-08-28 10:25:09,946 [ERROR] yardstick.benchmark.scenarios.availability.monitor.monitor_multi monitor_multi.py:78 SLA **failure**: 14.015082 > 5.000000
-.. image:: danube/images/ovp_vping_user.png
- :align: center
- :scale: 100%
+ 2018-08-28 10:23:41,907 [INFO] yardstick.benchmark.core.task task.py:127 Testcase: "opnfv_yardstick_tc052" **SUCCESS**!!!
-.. image:: danube/images/ovp_vping_ssh.png
- :align: center
- :scale: 100%
-5. SUT Info Verification
-========================
+SUT Info Verification
+=====================
SUT information must be present in the results to validate that all required endpoint services
and at least two controllers were present during test execution. For the results shown below,
-click the '**info**' hyperlink in the **SUT** column to navigate to the SUT information page.
+click the **info** hyperlink in the **SUT** column to navigate to the SUT information page.
-.. image:: danube/images/sut_info.png
+.. image:: images/sut_info.png
:align: center
:scale: 100%
-In the '**Endpoints**' listing shown below for the SUT VIM component, ensure that services are
+
+In the **Endpoints** listing shown below for the SUT VIM component, ensure that services are
present for identify, compute, image, volume and network at a minimum by inspecting the
-'**Service Type**' column.
+**Service Type** column.
-.. image:: danube/images/sut_endpoints.png
+.. image:: images/sut_endpoints.png
:align: center
:scale: 100%
-Inspect the '**Hosts**' listing found below the Endpoints secion of the SUT info page and ensure
+
+Inspect the **Hosts** listing found below the Endpoints secion of the SUT info page and ensure
at least two hosts are present, as two controllers are required the for the mandatory HA
-test-cases.
+test cases.
+
+
+Approve or Not Approve Results
+==============================
+
+When you decide to approve or not approve this test, you can click the **Operation** and choose
+**approve** or **not approve**. Once you have approved or not approved the test, you can click
+**View Reviews** to find the review status as shown below.
+
+.. image:: images/review_status.png
+ :align: center
+ :scale: 100%
diff --git a/docs/testing/user/systempreparation/index.rst b/docs/testing/user/systempreparation/index.rst
index afc5dc47..71916736 100644
--- a/docs/testing/user/systempreparation/index.rst
+++ b/docs/testing/user/systempreparation/index.rst
@@ -5,9 +5,9 @@
.. _dovetail-system_preparation_guide:
-===============================================
-OPNFV Verified Program system preparation guide
-===============================================
+=================================
+OVP NFVI System Preparation Guide
+=================================
This document provides a general guide to hardware system prerequisites
and expectations for running OPNFV OVP testing. For detailed guide of
@@ -52,7 +52,7 @@ The SUT is expected to be connected with high performance networks. These networ
are expected in the SUT:
- A management network by which the Test Node can reach all identity, image, network,
-and compute services in the SUT
+ and compute services in the SUT
- A data network that supports the virtual network capabilities and data path testing
Additional networks, such as Light Out Management or storage networks, may be
diff --git a/docs/testing/user/testspecification/faultmanagement/index.rst b/docs/testing/user/testspecification/faultmanagement/index.rst
deleted file mode 100644
index e69de29b..00000000
--- a/docs/testing/user/testspecification/faultmanagement/index.rst
+++ /dev/null
diff --git a/docs/testing/user/testspecification/forwardingpackets/index.rst b/docs/testing/user/testspecification/forwardingpackets/index.rst
deleted file mode 100644
index 4a4ffea9..00000000
--- a/docs/testing/user/testspecification/forwardingpackets/index.rst
+++ /dev/null
@@ -1,147 +0,0 @@
-.. This work is licensed under a Creative Commons Attribution 4.0 International License.
-.. http://creativecommons.org/licenses/by/4.0
-.. (c) Huawei Technologies Co.,Ltd
-
-==================================================
-Forwarding Packets in Data Path test specification
-==================================================
-
-.. toctree::
- :maxdepth: 2
-
-Scope
-=====
-
-This test area evaluates the ability of the system under test to support basic packet forwarding.
-The test in this test area will evaluate basic packet forwarding through virtual IPv4 networks
-in data path, including creating server and verifying network connectivity to the created server
-with ping operation using MTU sized packets.
-
-References
-==========
-
-N/A
-
-Definitions and abbreviations
-=============================
-
-The following terms and abbreviations are used in conjunction with this test
-area
-
-- API - Application Programming Interface
-- ICMP - Internet Control Message Protocol
-- MTU - Maximum Transmission Unit
-- NFVi - Network Functions Virtualization infrastructure
-- SSH - Secure Shell
-- VIM - Virtual Infrastructure Manager
-- VM - Virtual Machine
-
-System Under Test (SUT)
-=======================
-
-The system under test is assumed to be the NFVi and VIM in operation on a
-Pharos compliant infrastructure.
-
-Test Area Structure
-===================
-
-The test area is structured based on the basic operations of forwarding packets
-in data path through virtual networks. Specifically, the test performs
-clean-up operations which return the system to the same state as before the test.
-
-This test case is included in the test case dovetail.tempest.tc001 of OVP test suite.
-
-Test Descriptions
-=================
-
-API Used and Reference
-----------------------
-
-Security Groups: https://developer.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
-
-- create security group
-- delete security group
-
-Networks: https://developer.openstack.org/api-ref/networking/v2/index.html#networks
-
-- create network
-- delete network
-
-Routers and interface: https://developer.openstack.org/api-ref/networking/v2/index.html#routers-routers
-
-- create router
-- delete router
-- add interface to router
-
-Subnets: https://developer.openstack.org/api-ref/networking/v2/index.html#subnets
-
-- create subnet
-- delete subnet
-
-Servers: https://developer.openstack.org/api-ref/compute/
-
-- create keypair
-- create server
-- delete server
-- add/assign floating ip
-
-Ports: https://developer.openstack.org/api-ref/networking/v2/index.html#ports
-
-- create port
-- delete port
-
-Floating IPs: https://developer.openstack.org/api-ref/networking/v2/index.html#floating-ips-floatingips
-
-- create floating ip
-- delete floating ip
-
-----------------------------
-MTU Sized Frames Fit Through
-----------------------------
-
-Test case specification
------------------------
-
-tempest.scenario.test_network_basic_ops.TestNetworkBasicOps.test_mtu_sized_frames
-
-Test preconditions
-------------------
-
-* Nova has been configured to boot VMs with Neutron-managed networking
-* Neutron net-mtu extension API
-* One public network
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-''''''''''''''
-
-* Test action 1: Create a security group SG1, which has rules for allowing
- incoming and outgoing SSH and ICMP traffic
-* Test action 2: Create a neutron network NET1
-* Test action 3: Create a tenant router R1 which routes traffic to public network
-* Test action 4: Create a subnet SUBNET1 and add it as router interface
-* Test action 5: Create a server VM1 with SG1 and NET1, and assign a floating
- ip FIP1 (via R1) to VM1
-* Test action 6: Set MTU size to be the default MTU size of the SUT's network
-* Test action 7: Host sends MTU sized ICMP packets to VM1 using ``ping``
-* **Test assertion 1:** Ping FIP1 using MTU sized packets successfully
-* Test action 8: SSH to VM1 with FIP1
-* **Test assertion 2:** SSH to VM1 with FIP1 successfully
-* Test action 9: Delete SG1, NET1, SUBNET1, R1, VM1 and FIP1
-
-Pass / fail criteria
-''''''''''''''''''''
-
-This test evaluates the network connectivity using MTU sized frames.
-Specifically, the test verifies that:
-
-* With Neutron net-mtu extension configured, MTU sized packets can fit through network.
-
-In order to pass this test, all test assertions listed in the test execution above need to pass.
-
-Post conditions
----------------
-
-N/A
diff --git a/docs/testing/user/testspecification/highavailability/index.rst b/docs/testing/user/testspecification/highavailability/index.rst
index 1dd99d41..e489894f 100644
--- a/docs/testing/user/testspecification/highavailability/index.rst
+++ b/docs/testing/user/testspecification/highavailability/index.rst
@@ -31,7 +31,7 @@ This test area references the following specifications:
- ETSI GS NFV-REL 001
- - http://www.etsi.org/deliver/etsi_gs/NFV-REL/001_099/001/01.01.01_60/gs_nfv-rel001v010101p.pdf
+ - https://www.etsi.org/deliver/etsi_gs/NFV-REL/001_099/001/01.01.01_60/gs_nfv-rel001v010101p.pdf
- OpenStack High Availability Guide
@@ -84,7 +84,9 @@ Test Case 1 - Controller node OpenStack service down - nova-api
Short name
----------
-dovetail.ha.tc001.nova-api_service_down
+yardstick.ha.nova_api
+
+Yardstick test case: opnfv_yardstick_tc019.yaml
Use case specification
----------------------
@@ -102,6 +104,7 @@ Test preconditions
There is more than one controller node, which is providing the "nova-api"
service for API end-point.
+
Denoted a controller node as Node1 in the following configuration.
@@ -169,8 +172,10 @@ Post conditions
---------------
Restart the process of "nova-api" if they are not running.
-Delete image with "openstack image delete test-cirros"
-Delete flavor with "openstack flavor delete m1.test"
+
+Delete image with "openstack image delete test-cirros".
+
+Delete flavor with "openstack flavor delete m1.test".
---------------------------------------------------------------------
@@ -180,7 +185,9 @@ Test Case 2 - Controller node OpenStack service down - neutron-server
Short name
----------
-dovetail.ha.tc002.neutron-server_service_down
+yardstick.ha.neutron_server
+
+Yardstick test case: opnfv_yardstick_tc045.yaml
Use case specification
----------------------
@@ -196,6 +203,7 @@ Test preconditions
There is more than one controller node, which is providing the "neutron-server"
service for API end-point.
+
Denoted a controller node as Node1 in the following configuration.
Basic test flow execution description and pass/fail criteria
@@ -264,7 +272,9 @@ Test Case 3 - Controller node OpenStack service down - keystone
Short name
----------
-dovetail.ha.tc003.keystone_service_down
+yardstick.ha.keystone
+
+Yardstick test case: opnfv_yardstick_tc046.yaml
Use case specification
----------------------
@@ -280,6 +290,7 @@ Test preconditions
There is more than one controller node, which is providing the "keystone"
service for API end-point.
+
Denoted a controller node as Node1 in the following configuration.
Basic test flow execution description and pass/fail criteria
@@ -342,7 +353,9 @@ Test Case 4 - Controller node OpenStack service down - glance-api
Short name
----------
-dovetail.ha.tc004.glance-api_service_down
+yardstick.ha.glance_api
+
+Yardstick test case: opnfv_yardstick_tc047.yaml
Use case specification
----------------------
@@ -358,6 +371,7 @@ Test preconditions
There is more than one controller node, which is providing the "glance-api"
service for API end-point.
+
Denoted a controller node as Node1 in the following configuration.
@@ -430,7 +444,9 @@ Test Case 5 - Controller node OpenStack service down - cinder-api
Short name
----------
-dovetail.ha.tc005.cinder-api_service_down
+yardstick.ha.cinder_api
+
+Yardstick test case: opnfv_yardstick_tc048.yaml
Use case specification
----------------------
@@ -446,6 +462,7 @@ Test preconditions
There is more than one controller node, which is providing the "cinder-api"
service for API end-point.
+
Denoted a controller node as Node1 in the following configuration.
Basic test flow execution description and pass/fail criteria
@@ -509,7 +526,9 @@ Test Case 6 - Controller Node CPU Overload High Availability
Short name
----------
-dovetail.ha.tc006.cpu_overload
+yardstick.ha.cpu_load
+
+Yardstick test case: opnfv_yardstick_tc051.yaml
Use case specification
----------------------
@@ -526,6 +545,7 @@ Test preconditions
There is more than one controller node, which is providing the "cinder-api",
"neutron-server", "glance-api" and "keystone" services for API end-point.
+
Denoted a controller node as Node1 in the following configuration.
Basic test flow execution description and pass/fail criteria
@@ -594,7 +614,9 @@ Test Case 7 - Controller Node Disk I/O Overload High Availability
Short name
----------
-dovetail.ha.tc007.disk_I/O_overload
+yardstick.ha.disk_load
+
+Yardstick test case: opnfv_yardstick_tc052.yaml
Use case specification
----------------------
@@ -668,16 +690,18 @@ Test Case 8 - Controller Load Balance as a Service High Availability
Short name
----------
-dovetail.ha.tc008.load_balance_service_down
+yardstick.ha.haproxy
+
+Yardstick test case: opnfv_yardstick_tc053.yaml
Use case specification
----------------------
-This test verifies the high availability of "load balancer" service. When
-the "load balancer" service of a specified controller node is killed, whether
-"load balancer" service on other controller nodes will work, and whether the
-controller node will restart the "load balancer" service are checked. This
-test case kills the processes of "load balancer" service on the selected
+This test verifies the high availability of "haproxy" service. When
+the "haproxy" service of a specified controller node is killed, whether
+"haproxy" service on other controller nodes will work, and whether the
+controller node will restart the "haproxy" service are checked. This
+test case kills the processes of "haproxy" service on the selected
controller node, then checks whether the request of the related OpenStack
command is processed with no failure and whether the killed processes are
recovered.
@@ -685,8 +709,10 @@ recovered.
Test preconditions
------------------
-There is more than one controller node, which is providing the "load balancer"
-service for rest-api. Denoted as Node1 in the following configuration.
+There is more than one controller node, which is providing the "haproxy"
+service for rest-api.
+
+Denoted as Node1 in the following configuration.
Basic test flow execution description and pass/fail criteria
------------------------------------------------------------
@@ -694,33 +720,32 @@ Basic test flow execution description and pass/fail criteria
Methodology for monitoring high availability
''''''''''''''''''''''''''''''''''''''''''''
-The high availability of "load balancer" service is evaluated by monitoring
+The high availability of "haproxy" service is evaluated by monitoring
service outage time and process outage time
Service outage time is tested by continuously executing "openstack image list"
command in loop and checking if the response of the command request is returned
with no failure.
-When the response fails, the "load balancer" service is considered in outage.
+When the response fails, the "haproxy" service is considered in outage.
The time between the first response failure and the last response failure is
considered as service outage time.
-Process outage time is tested by checking the status of processes of "load
-balancer" service on the selected controller node. The time of those processes
+Process outage time is tested by checking the status of processes of "haproxy"
+service on the selected controller node. The time of those processes
being killed to the time of those processes being recovered is the process
outage time.
-Process recovery is verified by checking the existence of processes of "load
-balancer" service.
+Process recovery is verified by checking the existence of processes of "haproxy" service.
Test execution
''''''''''''''
* Test action 1: Connect to Node1 through SSH, and check that processes of
- "load balancer" service are running on Node1
-* Test action 2: Start two monitors: one for processes of "load balancer"
+ "haproxy" service are running on Node1
+* Test action 2: Start two monitors: one for processes of "haproxy"
service and the other for "openstack image list" command. Each monitor will
run as an independent process
* Test action 3: Connect to Node1 through SSH, and then kill the processes of
- "load balancer" service
+ "haproxy" service
* Test action 4: Continuously measure service outage time from the monitor until
the service outage time is more than 5s
* Test action 5: Continuously measure process outage time from the monitor until
@@ -737,7 +762,322 @@ A negative result will be generated if the above is not met in completion.
Post conditions
---------------
-Restart the processes of "load balancer" if they are not running.
+Restart the processes of "haproxy" if they are not running.
+
+----------------------------------------------------------------
+Test Case 9 - Controller node OpenStack service down - Database
+----------------------------------------------------------------
+
+Short name
+----------
+
+yardstick.ha.database
+
+Yardstick test case: opnfv_yardstick_tc090.yaml
+
+Use case specification
+----------------------
+
+This test case verifies that the high availability of the data base instances
+used by OpenStack (mysql) on control node is working properly.
+Specifically, this test case kills the processes of database service on a
+selected control node, then checks whether the request of the related
+OpenStack command is OK and the killed processes are recovered.
+
+Test preconditions
+------------------
+
+In this test case, an attacker called "kill-process" is needed.
+This attacker includes three parameters: fault_type, process_name and host.
+
+The purpose of this attacker is to kill any process with a specific process
+name which is run on the host node. In case that multiple processes use the
+same name on the host node, all of them are going to be killed by this attacker.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Methodology for verifying service continuity and recovery
+'''''''''''''''''''''''''''''''''''''''''''''''''''''''''
+
+In order to verify this service two different monitors are going to be used.
+
+As first monitor is used a OpenStack command and acts as watcher for
+database connection of different OpenStack components.
+
+For second monitor is used a process monitor and the main purpose is to watch
+whether the database processes on the host node are killed properly.
+
+Therefore, in this test case, there are two metrics:
+
+* service_outage_time, which indicates the maximum outage time (seconds)
+ of the specified OpenStack command request
+* process_recover_time, which indicates the maximum time (seconds) from the
+ process being killed to recovered
+
+Test execution
+''''''''''''''
+* Test action 1: Connect to Node1 through SSH, and check that "database"
+ processes are running on Node1
+* Test action 2: Start two monitors: one for "database" processes on the host
+ node and the other for connection toward database from OpenStack
+ components, verifying the results of openstack image list, openstack router list,
+ openstack stack list and openstack volume list.
+ Each monitor will run as an independent process
+* Test action 3: Connect to Node1 through SSH, and then kill the "mysql"
+ process(es)
+* Test action 4: Stop monitors after a period of time specified by "waiting_time".
+ The monitor info will be aggregated.
+* Test action 5: Verify the SLA and set the verdict of the test case to pass or fail.
+
+
+Pass / fail criteria
+''''''''''''''''''''
+
+Check whether the SLA is passed:
+- The process outage time is less than 30s.
+- The service outage time is less than 5s.
+
+The database operations are carried out in above order and no errors occur.
+
+A negative result will be generated if the above is not met in completion.
+
+Post conditions
+---------------
+
+The database service is up and running again.
+If the database service did not recover successfully by itself,
+the test explicitly restarts the database service.
+
+------------------------------------------------------------------------
+Test Case 10 - Controller Messaging Queue as a Service High Availability
+------------------------------------------------------------------------
+
+Short name
+----------
+
+yardstick.ha.rabbitmq
+
+Yardstick test case: opnfv_yardstick_tc056.yaml
+
+Use case specification
+----------------------
+
+This test case will verify the high availability of the messaging queue
+service (RabbitMQ) that supports OpenStack on controller node. This
+test case expects that message bus service implementation is RabbitMQ.
+If the SUT uses a different message bus implementations, the Dovetail
+configuration (pod.yaml) can be changed accordingly. When messaging
+queue service (which is active) of a specified controller node
+is killed, the test case will check whether messaging queue services
+(which are standby) on other controller nodes will be switched active,
+and whether the cluster manager on the attacked controller node will
+restart the stopped messaging queue.
+
+Test preconditions
+------------------
+
+There is more than one controller node, which is providing the "messaging queue"
+service. Denoted as Node1 in the following configuration.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Methodology for verifying service continuity and recovery
+'''''''''''''''''''''''''''''''''''''''''''''''''''''''''
+
+The high availability of "messaging queue" service is evaluated by monitoring
+service outage time and process outage time.
+
+Service outage time is tested by continuously executing "openstack image list",
+"openstack network list", "openstack volume list" and "openstack stack list"
+commands in loop and checking if the responses of the command requests are
+returned with no failure.
+When the response fails, the "messaging queue" service is considered in outage.
+The time between the first response failure and the last response failure is
+considered as service outage time.
+
+Process outage time is tested by checking the status of processes of "messaging
+queue" service on the selected controller node. The time of those processes
+being killed to the time of those processes being recovered is the process
+outage time.
+Process recovery is verified by checking the existence of processes of
+"messaging queue" service.
+
+Test execution
+''''''''''''''
+
+* Test action 1: Start five monitors: one for processes of "messaging queue"
+ service and the others for "openstack image list", "openstack network list",
+ "openstack stack list" and "openstack volume list" command. Each monitor
+ will run as an independent process
+* Test action 2: Connect to Node1 through SSH, and then kill all the processes of
+ "messaging queue" service
+* Test action 3: Continuously measure service outage time from the monitors until
+ the service outage time is more than 5s
+* Test action 4: Continuously measure process outage time from the monitor until
+ the process outage time is more than 30s
+
+Pass / fail criteria
+''''''''''''''''''''
+
+Test passes if the process outage time is no more than 30s and
+the service outage time is no more than 5s.
+
+A negative result will be generated if the above is not met in completion.
+
+Post conditions
+---------------
+Restart the processes of "messaging queue" if they are not running.
+
+---------------------------------------------------------------------------
+Test Case 11 - Controller node OpenStack service down - Controller Restart
+---------------------------------------------------------------------------
+
+Short name
+----------
+
+yardstick.ha.controller_restart
+
+Yardstick test case: opnfv_yardstick_tc025.yaml
+
+Use case specification
+----------------------
+
+This test case verifies that the high availability of controller node is working
+properly.
+Specifically, this test case shutdowns a specified controller node via IPMI,
+then checks whether all services provided by the controller node are OK with
+some monitor tools.
+
+Test preconditions
+------------------
+
+In this test case, an attacker called "host-shutdown" is needed.
+This attacker includes two parameters: fault_type and host.
+
+The purpose of this attacker is to shutdown a controller and check whether the
+services are handled by this controller are still working normally.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Methodology for verifying service continuity and recovery
+'''''''''''''''''''''''''''''''''''''''''''''''''''''''''
+
+In order to verify this service one monitor is going to be used.
+
+This monitor is using an OpenStack command and the respective command name of
+the OpenStack component that we want to verify that the respective service is
+still running normally.
+
+In this test case, there is one metric: 1)service_outage_time: which indicates
+the maximum outage time (seconds) of the specified OpenStack command request.
+
+Test execution
+''''''''''''''
+* Test action 1: Connect to Node1 through SSH, and check that controller services
+ are running normally
+* Test action 2: Start monitors: each monitor will run as independently
+ process, monitoring the image list, router list, stack list and volume list accordingly.
+ The monitor info will be collected.
+* Test action 3: Using the IPMI component, the Node1 is shut-down remotely.
+* Test action 4: Stop monitors after a period of time specified by "waiting_time".
+ The monitor info will be aggregated.
+* Test action 5: Verify the SLA and set the verdict of the test case to pass or fail.
+
+
+Pass / fail criteria
+''''''''''''''''''''
+
+Check whether the SLA is passed:
+- The process outage time is less than 30s.
+- The service outage time is less than 5s.
+
+The controller operations are carried out in above order and no errors occur.
+
+A negative result will be generated if the above is not met in completion.
+
+Post conditions
+---------------
+
+The controller has been restarted
+
+----------------------------------------------------------------------------
+Test Case 12 - OpenStack Controller Virtual Router Service High Availability
+----------------------------------------------------------------------------
+
+Short name
+----------
+
+yardstick.ha.neutron_l3_agent
+
+Yardstick test case: opnfv_yardstick_tc058.yaml
+
+Use case specification
+----------------------
+
+This test case will verify the high availability of virtual routers(L3 agent)
+on controller node. When a virtual router service on a specified controller
+node is shut down, this test case will check whether the network of virtual
+machines will be affected, and whether the attacked virtual router service
+will be recovered.
+
+Test preconditions
+------------------
+
+There is more than one controller node, which is providing the Neutron API
+extension called "neutron-l3-agent" virtual router service API.
+
+Denoted as Node1 in the following configuration.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Methodology for verifying service continuity and recovery
+'''''''''''''''''''''''''''''''''''''''''''''''''''''''''
+
+The high availability of "neutrol-l3-agent" virtual router service is evaluated
+by monitoring service outage time and process outage time.
+
+Service outage is tested using ping to virtual machines. Ping tests that
+the network routing of virtual machines is ok.
+When the response fails, the virtual router service is considered in outage.
+The time between the first response failure and the last response failure is
+considered as service outage time.
+
+Process outage time is tested by checking the status of processes of "neutron-l3-agent"
+service on the selected controller node. The time of those processes being
+killed to the time of those processes being recovered is the process outage time.
+
+Process recovery is verified by checking the existence of processes of
+"neutron-l3-agent" service.
+
+Test execution
+''''''''''''''
+* Test action 1: Two host VMs are booted, these two hosts are in two different
+ networks, the networks are connected by a virtual router.
+* Test action 2: Start monitors: each monitor will run with independently process.
+ The monitor info will be collected.
+* Test action 3: Do attacker: Connect the host through SSH, and then execute the kill
+ process script with param value specified by “process_name”
+* Test action 4: Stop monitors after a period of time specified by “waiting_time”
+ The monitor info will be aggregated.
+* Test action 5: Verify the SLA and set the verdict of the test case to pass or fail.
+
+Pass / fail criteria
+''''''''''''''''''''
+
+Check whether the SLA is passed:
+- The process outage time is less than 30s.
+- The service outage time is less than 5s.
+
+A negative result will be generated if the above is not met in completion.
+
+Post conditions
+---------------
+Delete image with "openstack image delete neutron-l3-agent_ha_image".
+Delete flavor with "openstack flavor delete neutron-l3-agent_ha_flavor".
diff --git a/docs/testing/user/testspecification/index.rst b/docs/testing/user/testspecification/index.rst
index 2a8298c9..69df2a01 100644
--- a/docs/testing/user/testspecification/index.rst
+++ b/docs/testing/user/testspecification/index.rst
@@ -4,9 +4,9 @@
.. _dovetail-test_case_specification:
-==================================================
-OPNFV Verified Program test specification
-==================================================
+=======================
+OVP Test Specifications
+=======================
Introduction
============
@@ -18,26 +18,31 @@ associated test specification.
All tests in the OVP are required to fulfill a specific set of criteria in
order that the OVP is able to provide a fair assessment of the system under
-test. Test requirements are described in the 'Test Case Requirements'_
+test. Test requirements are described in the :ref:dovetail-test_case_requirements
document.
All tests areas addressed in the OVP are covered in the following test
specification documents.
.. toctree::
- :maxdepth: 1
-
- ./highavailability/index.rst
- ./vimoperationscompute/index.rst
- ./vimoperationsidentity/index.rst
- ./vimoperationsimage/index.rst
- ./vimoperationsnetwork/index.rst
- ./vimoperationsvolume/index.rst
- ./vping/index.rst
- ./ipv6/index.rst
- ./forwardingpackets/index.rst
- ./securitygroup/index.rst
- ./dynamicnetwork/index.rst
- ./machinelifecycle/index.rst
- ./multiplenodes/index.rst
- ./vpn/index.rst
+ :maxdepth: 2
+
+ highavailability/index
+ security_patrole/index
+ security_patrole_vxlan_dependent/index
+ stress/index
+ tempest_compute/index
+ tempest_identity_v3/index
+ tempest_image/index
+ tempest_ipv6/index
+ tempest_multi_node_scheduling/index
+ tempest_network_api/index
+ tempest_network_scenario/index
+ tempest_network_security/index
+ tempest_osinterop/index
+ tempest_trunk_ports/index
+ tempest_vm_lifecycle/index
+ tempest_volume/index
+ vnf/index
+ vping/index
+ vpn/index
diff --git a/docs/testing/user/testspecification/security_patrole/index.rst b/docs/testing/user/testspecification/security_patrole/index.rst
new file mode 100644
index 00000000..250d8bbd
--- /dev/null
+++ b/docs/testing/user/testspecification/security_patrole/index.rst
@@ -0,0 +1,327 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV
+
+=====================
+Patrole Tempest Tests
+=====================
+
+Scope
+=====
+
+This test area evaluates the ability of a system under test to support the
+role-based access control (RBAC) implementation.
+The test area specifically validates services image and networking.
+
+
+References
+================
+
+- `OpenStack image service API reference <https://docs.openstack.org/api-ref/image/v2/index.html>`_
+- `OpenStack metadata definitions service API reference <https://docs.openstack.org/api-ref/image/v2/metadefs-index.html>`_
+- `OpenStack layer 2 networking service API reference <https://docs.openstack.org/api-ref/network/v2/index.html#layer-2-networking>`_
+- `OpenStack layer 3 networking service API reference <https://docs.openstack.org/api-ref/network/v2/index.html#layer-3-networking>`_
+- `OpenStack network security API reference <https://docs.openstack.org/api-ref/network/v2/index.html#security>`_
+- `OpenStack resource management API reference <https://docs.openstack.org/api-ref/network/v2/index.html#resource-management>`_
+
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVI and VIM deployed on a Pharos
+compliant infrastructure.
+
+
+Test Area Structure
+====================
+
+The test area is structured in individual tests as listed below. Each test case
+is able to run independently, i.e. irrelevant of the state created by a previous
+test. For detailed information on the individual steps and assertions performed
+by the tests, review the Python source code accessible via the following links.
+
+**Image basic RBAC test:**
+
+These tests cover the RBAC tests of image basic operations.
+
+Implementation:
+`BasicOperationsImagesRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/image/test_images_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_create_image
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_create_image_tag
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_deactivate_image
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_delete_image
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_delete_image_tag
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_download_image
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_list_images
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_publicize_image
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_reactivate_image
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_show_image
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_update_image
+- patrole_tempest_plugin.tests.api.image.test_images_rbac.BasicOperationsImagesRbacTest.test_upload_image
+
+
+**Image namespaces RBAC test:**
+
+These tests cover the RBAC tests of image namespaces.
+
+Implementation:
+`ImageNamespacesRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/image/test_image_namespace_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_rbac.ImageNamespacesRbacTest.test_create_metadef_namespace
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_rbac.ImageNamespacesRbacTest.test_modify_metadef_namespace
+
+
+**Image namespaces objects RBAC test:**
+
+These tests cover the RBAC tests of image namespaces objects.
+
+Implementation:
+`ImageNamespacesObjectsRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/image/test_image_namespace_objects_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_objects_rbac.ImageNamespacesObjectsRbacTest.test_create_metadef_object_in_namespace
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_objects_rbac.ImageNamespacesObjectsRbacTest.test_list_metadef_objects_in_namespace
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_objects_rbac.ImageNamespacesObjectsRbacTest.test_show_metadef_object_in_namespace
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_objects_rbac.ImageNamespacesObjectsRbacTest.test_update_metadef_object_in_namespace
+
+
+**Image namespaces property RBAC test:**
+
+These tests cover the RBAC tests of image namespaces property.
+
+Implementation:
+`NamespacesPropertyRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/image/test_image_namespace_property_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_property_rbac.NamespacesPropertyRbacTest.test_add_md_properties
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_property_rbac.NamespacesPropertyRbacTest.test_get_md_properties
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_property_rbac.NamespacesPropertyRbacTest.test_get_md_property
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_property_rbac.NamespacesPropertyRbacTest.test_modify_md_properties
+
+
+**Image namespaces tags RBAC test:**
+
+These tests cover the RBAC tests of image namespaces tags.
+
+Implementation:
+`NamespaceTagsRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/image/test_image_namespace_tags_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_tags_rbac.NamespaceTagsRbacTest.test_create_namespace_tag
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_tags_rbac.NamespaceTagsRbacTest.test_create_namespace_tags
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_tags_rbac.NamespaceTagsRbacTest.test_list_namespace_tags
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_tags_rbac.NamespaceTagsRbacTest.test_show_namespace_tag
+- patrole_tempest_plugin.tests.api.image.test_image_namespace_tags_rbac.NamespaceTagsRbacTest.test_update_namespace_tag
+
+
+**Image resource types RBAC test:**
+
+These tests cover the RBAC tests of image resource types.
+
+Implementation:
+`ImageResourceTypesRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/image/test_image_resource_types_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.image.test_image_resource_types_rbac.ImageResourceTypesRbacTest.test_add_metadef_resource_type
+- patrole_tempest_plugin.tests.api.image.test_image_resource_types_rbac.ImageResourceTypesRbacTest.test_get_metadef_resource_type
+- patrole_tempest_plugin.tests.api.image.test_image_resource_types_rbac.ImageResourceTypesRbacTest.test_list_metadef_resource_types
+
+
+**Image member RBAC test:**
+
+These tests cover the RBAC tests of image member.
+
+Implementation:
+`ImagesMemberRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/image/test_images_member_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.image.test_images_member_rbac.ImagesMemberRbacTest.test_add_image_member
+- patrole_tempest_plugin.tests.api.image.test_images_member_rbac.ImagesMemberRbacTest.test_delete_image_member
+- patrole_tempest_plugin.tests.api.image.test_images_member_rbac.ImagesMemberRbacTest.test_list_image_members
+- patrole_tempest_plugin.tests.api.image.test_images_member_rbac.ImagesMemberRbacTest.test_show_image_member
+
+
+**Network agents RBAC test:**
+
+These tests cover the RBAC tests of network agents.
+
+Implementation:
+`AgentsRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_agents_rbac.py#L24>`_ and
+`DHCPAgentSchedulersRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_agents_rbac.py#L147>`_.
+
+- patrole_tempest_plugin.tests.api.network.test_agents_rbac.AgentsRbacTest.test_show_agent
+- patrole_tempest_plugin.tests.api.network.test_agents_rbac.AgentsRbacTest.test_update_agent
+- patrole_tempest_plugin.tests.api.network.test_agents_rbac.DHCPAgentSchedulersRbacTest.test_add_dhcp_agent_to_network
+- patrole_tempest_plugin.tests.api.network.test_agents_rbac.DHCPAgentSchedulersRbacTest.test_delete_network_from_dhcp_agent
+- patrole_tempest_plugin.tests.api.network.test_agents_rbac.DHCPAgentSchedulersRbacTest.test_list_networks_hosted_by_one_dhcp_agent
+
+
+**Network floating ips RBAC test:**
+
+These tests cover the RBAC tests of network floating ips.
+
+Implementation:
+`FloatingIpsRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_floating_ips_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_floating_ips_rbac.FloatingIpsRbacTest.test_create_floating_ip
+- patrole_tempest_plugin.tests.api.network.test_floating_ips_rbac.FloatingIpsRbacTest.test_create_floating_ip_floatingip_address
+- patrole_tempest_plugin.tests.api.network.test_floating_ips_rbac.FloatingIpsRbacTest.test_delete_floating_ip
+- patrole_tempest_plugin.tests.api.network.test_floating_ips_rbac.FloatingIpsRbacTest.test_show_floating_ip
+- patrole_tempest_plugin.tests.api.network.test_floating_ips_rbac.FloatingIpsRbacTest.test_update_floating_ip
+
+
+**Network basic RBAC test:**
+
+These tests cover the RBAC tests of network basic operations.
+
+Implementation:
+`NetworksRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_networks_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_create_network
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_create_network_is_default
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_create_network_router_external
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_create_network_shared
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_delete_network
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_list_dhcp_agents_on_hosting_network
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_show_network
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_show_network_provider_network_type
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_show_network_provider_physical_network
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_show_network_provider_segmentation_id
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_show_network_router_external
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_update_network
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_update_network_provider_physical_network
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_update_network_provider_segmentation_id
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_update_network_router_external
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_update_network_shared
+
+
+**Network ports RBAC test:**
+
+These tests cover the RBAC tests of network ports.
+
+Implementation:
+`PortsRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_ports_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_create_port
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_create_port_allowed_address_pairs
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_create_port_binding_host_id
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_create_port_binding_profile
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_create_port_device_owner
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_create_port_fixed_ips_ip_address
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_create_port_mac_address
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_create_port_security_enabled
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_delete_port
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_show_port
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_show_port_binding_host_id
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_show_port_binding_profile
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_show_port_binding_vif_details
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_show_port_binding_vif_type
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_update_port
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_update_port_allowed_address_pairs
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_update_port_binding_host_id
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_update_port_binding_profile
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_update_port_device_owner
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_update_port_fixed_ips_ip_address
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_update_port_mac_address
+- patrole_tempest_plugin.tests.api.network.test_ports_rbac.PortsRbacTest.test_update_port_security_enabled
+
+
+**Network routers RBAC test:**
+
+These tests cover the RBAC tests of network routers.
+
+Implementation:
+`RouterRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_routers_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_add_router_interface
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_create_router
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_create_router_enable_snat
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_create_router_external_fixed_ips
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_delete_router
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_remove_router_interface
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_show_high_availability_router
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_show_router
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_update_router
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_update_router_enable_snat
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_update_router_external_fixed_ips
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_update_router_external_gateway_info
+- patrole_tempest_plugin.tests.api.network.test_routers_rbac.RouterRbacTest.test_update_router_external_gateway_info_network_id
+
+
+**Network security groups RBAC test:**
+
+These tests cover the RBAC tests of network security groups.
+
+Implementation:
+`SecGroupRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_security_groups_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_security_groups_rbac.SecGroupRbacTest.test_create_security_group
+- patrole_tempest_plugin.tests.api.network.test_security_groups_rbac.SecGroupRbacTest.test_create_security_group_rule
+- patrole_tempest_plugin.tests.api.network.test_security_groups_rbac.SecGroupRbacTest.test_delete_security_group
+- patrole_tempest_plugin.tests.api.network.test_security_groups_rbac.SecGroupRbacTest.test_delete_security_group_rule
+- patrole_tempest_plugin.tests.api.network.test_security_groups_rbac.SecGroupRbacTest.test_list_security_group_rules
+- patrole_tempest_plugin.tests.api.network.test_security_groups_rbac.SecGroupRbacTest.test_list_security_groups
+- patrole_tempest_plugin.tests.api.network.test_security_groups_rbac.SecGroupRbacTest.test_show_security_group_rule
+- patrole_tempest_plugin.tests.api.network.test_security_groups_rbac.SecGroupRbacTest.test_show_security_group
+- patrole_tempest_plugin.tests.api.network.test_security_groups_rbac.SecGroupRbacTest.test_update_security_group
+
+
+**Network service providers RBAC test:**
+
+These tests cover the RBAC tests of network service providers.
+
+Implementation:
+`ServiceProvidersRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_service_providers_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_service_providers_rbac.ServiceProvidersRbacTest.test_list_service_providers
+
+
+**Network subnetpools RBAC test:**
+
+These tests cover the RBAC tests of network subnetpools.
+
+Implementation:
+`SubnetPoolsRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_subnetpools_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_subnetpools_rbac.SubnetPoolsRbacTest.test_create_subnetpool
+- patrole_tempest_plugin.tests.api.network.test_subnetpools_rbac.SubnetPoolsRbacTest.test_create_subnetpool_shared
+- patrole_tempest_plugin.tests.api.network.test_subnetpools_rbac.SubnetPoolsRbacTest.test_delete_subnetpool
+- patrole_tempest_plugin.tests.api.network.test_subnetpools_rbac.SubnetPoolsRbacTest.test_show_subnetpool
+- patrole_tempest_plugin.tests.api.network.test_subnetpools_rbac.SubnetPoolsRbacTest.test_update_subnetpool
+- patrole_tempest_plugin.tests.api.network.test_subnetpools_rbac.SubnetPoolsRbacTest.test_update_subnetpool_is_default
+
+
+**Network subnets RBAC test:**
+
+These tests cover the RBAC tests of network subnets.
+
+Implementation:
+`SubnetsRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_subnets_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_subnets_rbac.SubnetsRbacTest.test_create_subnet
+- patrole_tempest_plugin.tests.api.network.test_subnets_rbac.SubnetsRbacTest.test_delete_subnet
+- patrole_tempest_plugin.tests.api.network.test_subnets_rbac.SubnetsRbacTest.test_list_subnets
+- patrole_tempest_plugin.tests.api.network.test_subnets_rbac.SubnetsRbacTest.test_show_subnet
+- patrole_tempest_plugin.tests.api.network.test_subnets_rbac.SubnetsRbacTest.test_update_subnet
+
+
+**Network flavors RBAC test:**
+
+These tests cover the RBAC tests of network flavors.
+
+Implementation:
+`FlavorsRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_flavors_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_flavors_rbac.FlavorsPluginRbacTest.test_create_flavor
+- patrole_tempest_plugin.tests.api.network.test_flavors_rbac.FlavorsPluginRbacTest.test_delete_flavor
+- patrole_tempest_plugin.tests.api.network.test_flavors_rbac.FlavorsPluginRbacTest.test_list_flavors
+- patrole_tempest_plugin.tests.api.network.test_flavors_rbac.FlavorsPluginRbacTest.test_show_flavor
+- patrole_tempest_plugin.tests.api.network.test_flavors_rbac.FlavorsPluginRbacTest.test_update_flavor
+
+
+**Network segments RBAC test:**
+
+These tests cover the RBAC tests of network segments.
+
+Implementation:
+`SegmentsRbacTest <https://github.com/openstack/patrole/blob/0.4.0/patrole_tempest_plugin/tests/api/network/test_network_segments_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_network_segments_rbac.NetworkSegmentsRbacTest.test_create_network_segments
+- patrole_tempest_plugin.tests.api.network.test_network_segments_rbac.NetworkSegmentsRbacTest.test_show_network_segments
+- patrole_tempest_plugin.tests.api.network.test_network_segments_rbac.NetworkSegmentsRbacTest.test_update_network_segments
diff --git a/docs/testing/user/testspecification/security_patrole_vxlan_dependent/index.rst b/docs/testing/user/testspecification/security_patrole_vxlan_dependent/index.rst
new file mode 100644
index 00000000..646cb8b6
--- /dev/null
+++ b/docs/testing/user/testspecification/security_patrole_vxlan_dependent/index.rst
@@ -0,0 +1,51 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV
+
+=====================================
+Patrole Tempest Tests Depend on Vxlan
+=====================================
+
+Scope
+=====
+
+This test area includes some tempest role-based access control (RBAC) tests
+which depend on vxlan physical networks.
+
+
+References
+================
+
+- `OpenStack image service API reference <https://docs.openstack.org/api-ref/image/v2/index.html>`_
+- `OpenStack metadata definitions service API reference <https://docs.openstack.org/api-ref/image/v2/metadefs-index.html>`_
+- `OpenStack layer 2 networking service API reference <https://docs.openstack.org/api-ref/network/v2/index.html#layer-2-networking>`_
+- `OpenStack layer 3 networking service API reference <https://docs.openstack.org/api-ref/network/v2/index.html#layer-3-networking>`_
+- `OpenStack network security API reference <https://docs.openstack.org/api-ref/network/v2/index.html#security>`_
+- `OpenStack resource management API reference <https://docs.openstack.org/api-ref/network/v2/index.html#resource-management>`_
+
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVI and VIM deployed on a Pharos
+compliant infrastructure.
+
+
+Test Area Structure
+====================
+
+The test area is structured in individual tests as listed below. Each test case
+is able to run independently, i.e. irrelevant of the state created by a previous
+test. For detailed information on the individual steps and assertions performed
+by the tests, review the Python source code accessible via the following links.
+
+
+**Network basic RBAC test:**
+
+These tests cover the RBAC tests of network basic operations by creating a vxlan provider network.
+
+Implementation:
+`NetworksRbacTest <https://github.com/openstack/patrole/blob/0.2.0/patrole_tempest_plugin/tests/api/network/test_networks_rbac.py>`_
+
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_create_network_provider_network_type
+- patrole_tempest_plugin.tests.api.network.test_networks_rbac.NetworksRbacTest.test_create_network_provider_segmentation_id
diff --git a/docs/testing/user/testspecification/stress/index.rst b/docs/testing/user/testspecification/stress/index.rst
new file mode 100644
index 00000000..5483fc93
--- /dev/null
+++ b/docs/testing/user/testspecification/stress/index.rst
@@ -0,0 +1,159 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV
+
+=========================
+Stress Test Specification
+=========================
+
+.. toctree::
+ :maxdepth: 2
+
+Scope
+=====
+
+The stress test involves testing and verifying the ability of the SUT to withstand
+stress and other challenging factors. Main purpose behind the testing is to make sure
+the SUT is able to absorb failures while providing an acceptable level of service.
+
+
+References
+==========
+
+This test area references the following specifications, definitions and reviews:
+
+- Upstream OpenStack NOVA Resiliency
+
+ - https://wiki.openstack.org/wiki/NovaResiliency
+
+- Stress Testing over OPNFV Platform
+
+ - https://wiki.opnfv.org/display/bottlenecks/Stress+Testing+over+OPNFV+Platform
+
+
+Definitions and Abbreviations
+=============================
+
+The following terms and abbreviations are used in conjunction with this test area
+
+- iff - if and only if
+- NFVI - Network Functions Virtualization Infrastructure
+- NaN - Not a Number
+- Service Level - Measurable terms that describe the quality of the service provided by the SUT within a given time period
+- SUT - System Under Test
+- VM - Virtual Machine
+
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVI and VIM in operation on a
+Pharos compliant infrastructure.
+
+
+Test Area Structure
+====================
+
+According to the testing goals stated in the test scope section,
+preceding test will not affect the subsequent test
+as long as the SUT is able to sustain the given stress
+while providing an acceptable level of service.
+Any FAIL result from a single test case will cause the SUT failing the whole test.
+
+
+Test Descriptions
+=================
+
+----------------------------------------------------------------
+Test Case 1 - Concurrent capacity based on life-cycle ping test
+----------------------------------------------------------------
+
+Short name
+----------
+
+bottlenecks.stress.ping
+
+Use case specification
+----------------------
+
+This test case verifies the ability of the SUT concurrently setting up VM pairs
+for different tenants (through different OpenStack related components) and
+providing acceptable capacity under stressful conditions. The connectivity between
+VMs in a VM pair for a tenant is validated through Ping test. A life-cycle event
+in this test case is particularly referred to a VM pair life-cycle consisting of
+spawning, pinging and destroying.
+
+Test preconditions
+------------------
+
+* heat_template_version: 2013-05-23
+* ElasticSearch Port: 9200
+* LogStash Port: 5044
+* Kibana Port: 5601
+* Yardstick Port: 5000
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Methodology for validating capacity of the SUT
+''''''''''''''''''''''''''''''''''''''''''''''
+Validating capacity of the SUT based on life-cycle ping test generally involves
+2 subtests which provides secondary validation for the SUT furnishing users with
+reliable capacity without being crushed.
+
+Let *N1*, *N2*, *N3* and *P1* be certain preset numbers, respectively.
+In subtest 1, the SUT concurrently setting up *N1* VM pairs with each VM pair
+belonging to a different tenant. Then VM1 in a VM pair pings VM2 for *P1* times
+with *P1* packets. The connectivity could be validated iff VM1 successfully pings
+VM2 with these *P1* packets.
+Subtest 1 is finished iff all the concurrent (*N1*) requests for creating VM pairs
+are fulfilled with returned values that indicate the statuses of the VM pairs creations.
+
+Subtest 2 is executed after subtest 1 as secondary validation of the capacity.
+It follows the same workflow as subtest 1 does to set up *N2* VM pairs.
+
+Assume *S1* and *S2* be the numbers of VM pairs that are successfully created in
+subtest 1 and subtest 2, respectively. If *min(S1,S2)>=N3*, then the SUT is considered as PASS.
+Otherwise, we denote the SUT with FAIL.
+
+Note that for subtest 1, if the number of successfully created VM pairs, i.e., *S1*,
+is smaller than *N3*. Subtest 2 will not be executed and SUT will be marked with FAIL.
+
+Test execution
+''''''''''''''
+* Test action 1: Install the testing tools by pulling and running the Bottlenecks Docker container
+* Test action 2: Prepare the test by sourcing openstack credential file,
+ eliminating the environment constraints, i.e., Quota setting, setting up
+ Yardstick docker, pulling and registering OS images and VM flavor
+* Test action 3: Call Yardstick to concurrently creating *N1* VM pairs for N1 tenants
+* Test action 4: In each VM pair, VM1 pings VM2 for *P1* times with *P1* packets while recording the successful numbers
+* Test action 5: Mark the VM pairs with *P1* successful pings as PASS and record the total number of PASS VM pairs as *S1*
+* Test action 6: Destroy all the VM pairs
+* Test action 7: If *S1<N3*, the SUT is marked with FAIL and the test return. Otherwise go to *Test action 8*
+* Test action 8: Go to *Test action 3* and do the test again to create *N2* VM pairs with PASS VM pairs counted as *S2*
+* Test action 9: If *S2<N3*, the SUT is marked with FAIL. Otherwise marked with PASS.
+
+Pass / Fail criteria
+''''''''''''''''''''
+
+Typical setting of *(N1, N2, N3, P1)* is *(5, 5, 5, 10)*.
+The reference setting above is acquired based on the results from OPNFV CI jobs
+and testing over commercial products.
+
+The connectivity within a VM pair is validated iff:
+
+* VM1 successfully pings VM2 for *P1* times with *P1* packets
+
+The SUT is considered passing the test iff:
+
+* *min(S1,S2)>=N3*
+
+Note that after each subtest, the program will check if the successfully created number of VM pairs
+is smaller than *N3*. If true, the program will return and the SUT will be marked with FAIL.
+Then the passing criteria is equal to the equation above. When subtest 1 returns, *S2* here is denoted
+by NaN.
+
+Post conditions
+---------------
+
+N/A
diff --git a/docs/testing/user/testspecification/tempest_compute/index.rst b/docs/testing/user/testspecification/tempest_compute/index.rst
new file mode 100644
index 00000000..ee86ae77
--- /dev/null
+++ b/docs/testing/user/testspecification/tempest_compute/index.rst
@@ -0,0 +1,132 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) Ericsson AB
+
+==================================
+Tempest Compute test specification
+==================================
+
+
+Scope
+=====
+
+The Tempest Compute test area evaluates the ability of the System Under Test (SUT)
+to support dynamic network runtime operations through the life of a VNF.
+The tests in this test area will evaluate IPv4 network runtime operations
+functionality.
+
+These runtime operations includes:
+
+- Create, list and show flavors
+- Create and list security group rules
+- Create, delete and list security groups
+- Create, delete, show and list interfaces; attach and deattach ports to servers
+- List server addresses
+- Individual version endpoints info works
+- Servers Test Boot From Volume
+
+
+References
+==========
+
+`Security Groups: <https://docs.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups>`_
+
+- create security group
+- delete security group
+
+`Networks: <https://docs.openstack.org/api-ref/network/v2/index.html#networks>`_
+
+- create network
+- delete network
+
+`Routers and interface: <https://docs.openstack.org/api-ref/network/v2/index.html#routers-routers>`_
+
+- create router
+- update router
+- delete router
+- add interface to router
+
+`Subnets: <https://docs.openstack.org/api-ref/network/v2/index.html#subnets>`_
+
+- create subnet
+- update subnet
+- delete subnet
+
+`Servers: <https://docs.openstack.org/api-ref/compute/>`_
+
+- create keypair
+- create server
+- delete server
+- add/assign floating IP
+- disassociate floating IP
+
+`Ports: <https://docs.openstack.org/api-ref/network/v2/index.html#ports>`_
+
+- create port
+- update port
+- delete port
+
+`Floating IPs: <https://docs.openstack.org/api-ref/network/v2/index.html#floating-ips-floatingips>`_
+
+- create floating IP
+- delete floating IP
+
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVi and VIM in operation on a
+Pharos compliant infrastructure.
+
+
+Test Area Structure
+===================
+
+The test area is structured in individual tests as listed below.
+For detailed information on the individual steps and assertions performed
+by the tests, review the Python source code accessible via the following links:
+
+All these test cases are included in the test case functest.tempest.compute of
+OVP test suite.
+
+
+Test Area Structure
+===================
+
+The test area is structured in individual tests as listed below.
+For detailed information on the individual steps and assertions performed
+by the tests, review the Python source code accessible via the following links:
+
+
+
+- `Flavor V2 test <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/compute/flavors/test_flavors.py#L20>`_
+ - tempest.api.compute.flavors.test_flavors.FlavorsV2TestJSON.test_get_flavor
+ - tempest.api.compute.flavors.test_flavors.FlavorsV2TestJSON.test_list_flavors
+
+- `Security Group Rules test <https://github.com/openstack/tempest/blob/master/tempest/api/compute/security_groups/test_security_group_rules.py#L20>`_
+ - tempest.api.compute.security_groups.test_security_group_rules.SecurityGroupRulesTestJSON.test_security_group_rules_create
+ - tempest.api.compute.security_groups.test_security_group_rules.SecurityGroupRulesTestJSON.test_security_group_rules_list
+
+- `Security Groups test <https://github.com/openstack/tempest/blob/master/tempest/api/compute/security_groups/test_security_groups.py#L23>`_
+ - tempest.api.compute.security_groups.test_security_groups.SecurityGroupsTestJSON.test_security_groups_create_list_delete
+
+- `Attach Interfaces test <https://github.com/openstack/tempest/blob/master/tempest/api/compute/servers/test_attach_interfaces.py#L347>`_
+ - tempest.api.compute.servers.test_attach_interfaces.AttachInterfacesUnderV243Test.test_add_remove_fixed_ip
+
+
+- `Server Addresses test <https://github.com/openstack/tempest/blob/master/tempest/api/compute/servers/test_server_addresses.py#L21>`_
+ - tempest.api.compute.servers.test_server_addresses.ServerAddressesTestJSON.test_list_server_addresses
+ - tempest.api.compute.servers.test_server_addresses.ServerAddressesTestJSON.test_list_server_addresses_by_network
+
+
+- `Test Versions <https://github.com/openstack/tempest/blob/master/tempest/api/compute/test_versions.py#L19>`_
+ - tempest.api.compute.test_versions.TestVersions.test_get_version_details
+
+
+- `Servers Test Boot From Volume <https://github.com/openstack/tempest/blob/master/tempest/api/compute/servers/test_create_server.py#L158>`_
+ - tempest.api.compute.servers.test_create_server.ServersTestBootFromVolume.test_verify_server_details
+ - tempest.api.compute.servers.test_create_server.ServersTestBootFromVolume.test_list_servers
+
+
+- `Server Basic Operations test <https://github.com/openstack/tempest/blob/master/tempest/scenario/test_server_basic_ops.py#L30>`_
+ - tempest.scenario.test_server_basic_ops.TestServerBasicOps.test_server_basic_ops
diff --git a/docs/testing/user/testspecification/tempest_identity_v3/index.rst b/docs/testing/user/testspecification/tempest_identity_v3/index.rst
new file mode 100644
index 00000000..599427a3
--- /dev/null
+++ b/docs/testing/user/testspecification/tempest_identity_v3/index.rst
@@ -0,0 +1,83 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) Ericsson AB
+
+======================================
+Tempest Identity v3 test specification
+======================================
+
+
+Scope
+=====
+
+The Tempest Identity v3 test area evaluates the ability of the System Under Test
+(SUT) to create, list, delete and verify users through the life of a VNF.
+The tests in this test area will evaluate IPv4 network runtime operations
+functionality.
+
+These runtime operations may include that create, list, verify and delete:
+
+- credentials
+- domains
+- endpoints
+- user groups
+- policies
+- regions
+- roles
+- services
+- identities
+- API versions
+
+References
+==========
+
+`Identity API v3.0 <https://docs.openstack.org/api-ref/identity/v3/index.html>`_
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVi and VIM in operation on a
+Pharos compliant infrastructure.
+
+Test Area Structure
+===================
+
+The test area is structured in individual tests as listed below.
+For detailed information on the individual steps and assertions performed
+by the tests, review the Python source code accessible via the following links:
+
+All these test cases are included in the test case functest.tempest.identity_v3 of
+OVP test suite.
+
+- `Create, Get, Update and Delete Credentials <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_credentials.py#L21>`_
+ - tempest.api.identity.admin.v3.test_credentials.CredentialsTestJSON.test_credentials_create_get_update_delete
+
+- `Create and Verify Domain <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_domains.py#L159>`_
+ - tempest.api.identity.v3.test_domains.DefaultDomainTestJSON.test_default_domain_exists
+
+- `Create, Update and Delete Domain <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_domains.py>`_
+ - tempest.api.identity.admin.v3.test_domains.DomainsTestJSON.test_create_update_delete_domain
+
+- `Create and Update endpoint <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_endpoints.py>`_
+ - tempest.api.identity.admin.v3.test_endpoints.EndPointsTestJSON.test_update_endpoint
+
+- `Create, List and Delete Group Users <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_groups.py>`_
+ - tempest.api.identity.admin.v3.test_groups.GroupsV3TestJSON.test_group_users_add_list_delete
+
+- `Update Policy <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_policies.py>`_
+ - tempest.api.identity.admin.v3.test_policies.PoliciesTestJSON.test_create_update_delete_policy
+
+- `Create a Region with a Specific Id <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_regions.py>`_
+ - tempest.api.identity.admin.v3.test_regions.RegionsTestJSON.test_create_region_with_specific_id
+
+- `Create, Update and Show Role List <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_roles.py>`_
+ - tempest.api.identity.admin.v3.test_roles.RolesV3TestJSON.test_role_create_update_show_list
+
+- `Create a Service <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_services.py>`_
+ - tempest.api.identity.admin.v3.test_services.ServicesTestJSON.test_create_update_get_service
+
+- `Create and List Trusts <https://github.com/openstack/tempest/blob/12.2.0/tempest/api/identity/admin/v3/test_trusts.py#L193>`_
+ - tempest.api.identity.admin.v3.test_trusts.TrustsV3TestJSON.test_get_trusts_all
+
+- `List API Versions <https://github.com/openstack/tempest/blob/18.0.0/tempest/api/identity/v3/test_api_discovery.py>`_
+ - tempest.api.identity.v3.test_api_discovery.TestApiDiscovery.test_list_api_versions
diff --git a/docs/testing/user/testspecification/tempest_image/index.rst b/docs/testing/user/testspecification/tempest_image/index.rst
new file mode 100644
index 00000000..121b38e6
--- /dev/null
+++ b/docs/testing/user/testspecification/tempest_image/index.rst
@@ -0,0 +1,42 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) Ericsson AB
+
+================================
+Tempest Image test specification
+================================
+
+
+Scope
+=====
+
+The Tempest Image test area tests the basic operations of Images of the System Under
+Test (SUT) through the life of a VNF. The tests in this test area will evaluate IPv4
+network runtime operations functionality.
+
+References
+==========
+
+`Image Service API v2 <https://docs.openstack.org/api-ref/image/v2/index.html>`_
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVi and VIM in operation on a
+Pharos compliant infrastructure.
+
+Test Area Structure
+===================
+
+The test area is structured in individual tests as listed below.
+For detailed information on the individual steps and assertions performed
+by the tests, review the Python source code accessible via the following links:
+
+All these test cases are included in the test case functest.tempest.image of
+OVP test suite.
+
+- `Register, Upload, Get Image and Get Image File API's <https://github.com/openstack/tempest/blob/18.0.0/tempest/api/image/v2/test_images.py#L32>`_
+ - tempest.api.image.v2.test_images.BasicOperationsImagesTest.test_register_upload_get_image_file
+
+- `List Versions <https://github.com/openstack/tempest/blob/18.0.0/tempest/api/image/v2/test_versions.py>`_
+ - tempest.api.image.v2.test_versions.VersionsTest.test_list_versions
diff --git a/docs/testing/user/testspecification/tempest_ipv6/index.rst b/docs/testing/user/testspecification/tempest_ipv6/index.rst
new file mode 100644
index 00000000..3da23c99
--- /dev/null
+++ b/docs/testing/user/testspecification/tempest_ipv6/index.rst
@@ -0,0 +1,140 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV
+
+========================
+IPv6 test specification
+========================
+
+Scope
+=====
+
+The IPv6 test area will evaluate the ability for a SUT to support IPv6
+Tenant Network features and functionality. The tests in this test area will
+evaluate,
+
+- network, subnet, port, router API CRUD operations
+- interface add and remove operations
+- security group and security group rule API CRUD operations
+- IPv6 address assignment with dual stack, dual net, multiprefix in mode DHCPv6 stateless or SLAAC
+
+References
+================
+
+- upstream openstack API reference
+
+ - https://docs.openstack.org/api-ref/
+
+- upstream openstack IPv6 reference
+
+ - https://docs.openstack.org/newton/networking-guide/config-ipv6.html
+
+Definitions and abbreviations
+=============================
+
+The following terms and abbreviations are used in conjunction with this test area
+
+- API - Application Programming Interface
+- CIDR - Classless Inter-Domain Routing
+- CRUD - Create, Read, Update, and Delete
+- DHCP - Dynamic Host Configuration Protocol
+- DHCPv6 - Dynamic Host Configuration Protocol version 6
+- ICMP - Internet Control Message Protocol
+- NFVI - Network Functions Virtualization Infrastructure
+- NIC - Network Interface Controller
+- RA - Router Advertisements
+- radvd - The Router Advertisement Daemon
+- SDN - Software Defined Network
+- SLAAC - Stateless Address Auto Configuration
+- TCP - Transmission Control Protocol
+- UDP - User Datagram Protocol
+- VM - Virtual Machine
+- vNIC - virtual Network Interface Card
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVI and VIM deployed with a Pharos compliant infrastructure.
+
+Test Area Structure
+====================
+
+The test area is structured based on network, port and subnet operations. Each test case
+is able to run independently, i.e. irrelevant of the state created by a previous test.
+
+Test Descriptions
+=================
+
+API Used and Reference
+----------------------
+
+Networks: https://docs.openstack.org/api-ref/network/v2/index.html#networks
+
+- show network details
+- update network
+- delete network
+- list networks
+- create netowrk
+- bulk create networks
+
+Subnets: https://docs.openstack.org/api-ref/network/v2/index.html#subnets
+
+- list subnets
+- create subnet
+- bulk create subnet
+- show subnet details
+- update subnet
+- delete subnet
+
+Routers and interface: https://docs.openstack.org/api-ref/network/v2/index.html#routers-routers
+
+- list routers
+- create router
+- show router details
+- update router
+- delete router
+- add interface to router
+- remove interface from router
+
+Ports: https://docs.openstack.org/api-ref/network/v2/index.html#ports
+
+- show port details
+- update port
+- delete port
+- list port
+- create port
+- bulk create ports
+
+Security groups: https://docs.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
+
+- list security groups
+- create security groups
+- show security group
+- update security group
+- delete security group
+
+Security groups rules: https://docs.openstack.org/api-ref/network/v2/index.html#security-group-rules-security-group-rules
+
+- list security group rules
+- create security group rule
+- show security group rule
+- delete security group rule
+
+Servers: https://docs.openstack.org/api-ref/compute/
+
+- list servers
+- create server
+- create multiple servers
+- list servers detailed
+- show server details
+- update server
+- delete server
+
+All IPv6 api and scenario test cases addressed in OVP are covered in the
+following test specification documents.
+
+.. toctree::
+ :maxdepth: 2
+
+ ipv6_api.rst
+ ipv6_scenario.rst
diff --git a/docs/testing/user/testspecification/ipv6/index.rst b/docs/testing/user/testspecification/tempest_ipv6/ipv6_api.rst
index 0ede0a3d..b3d2c9dc 100644
--- a/docs/testing/user/testspecification/ipv6/index.rst
+++ b/docs/testing/user/testspecification/tempest_ipv6/ipv6_api.rst
@@ -2,137 +2,6 @@
.. http://creativecommons.org/licenses/by/4.0
.. (c) OPNFV
-========================
-IPv6 test specification
-========================
-
-.. toctree::
- :maxdepth: 2
-
-Scope
-=====
-
-The IPv6 test area will evaluate the ability for a SUT to support IPv6
-Tenant Network features and functionality. The tests in this test area will
-evaluate,
-
-- network, subnet, port, router API CRUD operations
-- interface add and remove operations
-- security group and security group rule API CRUD operations
-- IPv6 address assignment with dual stack, dual net, multiprefix in mode DHCPv6 stateless or SLAAC
-
-References
-================
-
-- upstream openstack API reference
-
- - http://developer.openstack.org/api-ref
-
-- upstream openstack IPv6 reference
-
- - https://docs.openstack.org/newton/networking-guide/config-ipv6.html
-
-Definitions and abbreviations
-=============================
-
-The following terms and abbreviations are used in conjunction with this test area
-
-- API - Application Programming Interface
-- CIDR - Classless Inter-Domain Routing
-- CRUD - Create, Read, Update, and Delete
-- DHCP - Dynamic Host Configuration Protocol
-- DHCPv6 - Dynamic Host Configuration Protocol version 6
-- ICMP - Internet Control Message Protocol
-- NFVI - Network Functions Virtualization Infrastructure
-- NIC - Network Interface Controller
-- RA - Router Advertisements
-- radvd - The Router Advertisement Daemon
-- SDN - Software Defined Network
-- SLAAC - Stateless Address Auto Configuration
-- TCP - Transmission Control Protocol
-- UDP - User Datagram Protocol
-- VM - Virtual Machine
-- vNIC - virtual Network Interface Card
-
-System Under Test (SUT)
-=======================
-
-The system under test is assumed to be the NFVI and VIM deployed with a Pharos compliant infrastructure.
-
-Test Area Structure
-====================
-
-The test area is structured based on network, port and subnet operations. Each test case
-is able to run independently, i.e. irrelevant of the state created by a previous test.
-
-Test Descriptions
-=================
-
-API Used and Reference
-----------------------
-
-Networks: https://developer.openstack.org/api-ref/networking/v2/index.html#networks
-
-- show network details
-- update network
-- delete network
-- list networks
-- create netowrk
-- bulk create networks
-
-Subnets: https://developer.openstack.org/api-ref/networking/v2/index.html#subnets
-
-- list subnets
-- create subnet
-- bulk create subnet
-- show subnet details
-- update subnet
-- delete subnet
-
-Routers and interface: https://developer.openstack.org/api-ref/networking/v2/index.html#routers-routers
-
-- list routers
-- create router
-- show router details
-- update router
-- delete router
-- add interface to router
-- remove interface from router
-
-Ports: https://developer.openstack.org/api-ref/networking/v2/index.html#ports
-
-- show port details
-- update port
-- delete port
-- list port
-- create port
-- bulk create ports
-
-Security groups: https://developer.openstack.org/api-ref/networking/v2/index.html#security-groups-security-groups
-
-- list security groups
-- create security groups
-- show security group
-- update security group
-- delete security group
-
-Security groups rules: https://developer.openstack.org/api-ref/networking/v2/index.html#security-group-rules-security-group-rules
-
-- list security group rules
-- create security group rule
-- show security group rule
-- delete security group rule
-
-Servers: https://developer.openstack.org/api-ref/compute/
-
-- list servers
-- create server
-- create multiple servers
-- list servers detailed
-- show server details
-- update server
-- delete server
-
------------------------------------------------------------------
Test Case 1 - Create and Delete Bulk Network, IPv6 Subnet and Port
------------------------------------------------------------------
@@ -140,7 +9,7 @@ Test Case 1 - Create and Delete Bulk Network, IPv6 Subnet and Port
Short name
----------
-dovetail.ipv6.tc001.bulk_network_subnet_port_create_delete
+functest.tempest.ipv6_api.bulk_network_subnet_port_create_delete
Use case specification
----------------------
@@ -192,7 +61,7 @@ Test execution
* Test action 18: List all networks, verifying the network ids are no longer present
* **Test assertion 8:** The two "id" parameters are not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to use bulk create commands to create networks, IPv6 subnets and ports on
@@ -215,7 +84,7 @@ Test Case 2 - Create, Update and Delete an IPv6 Network and Subnet
Short name
-----------
-dovetail.ipv6.tc002.network_subnet_create_update_delete
+functest.tempest.ipv6_api.network_subnet_create_update_delete
Use case specification
----------------------
@@ -256,7 +125,7 @@ Test execution
* **Test assertion 5:** The network "id" is not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to create, update, delete network, IPv6 subnet on the
@@ -279,7 +148,7 @@ Test Case 3 - Check External Network Visibility
Short name
-----------
-dovetail.ipv6.tc003.external_network_visibility
+functest.tempest.ipv6_api.external_network_visibility
Use case specification
----------------------
@@ -292,10 +161,12 @@ Test preconditions
------------------
1. The SUT has at least one external network.
-2. In the external network list, there is no network without external router, i.e.,
-all networks in this list are with external router.
+
+2. In the external network list, there is no network without external router,
+ i.e., all networks in this list are with external router.
+
3. There is one external network with configured public network id and there is
-no subnet on this network
+ no subnet on this network
Basic test flow execution description and pass/fail criteria
------------------------------------------------------------
@@ -318,7 +189,7 @@ Test execution
* **Test assertion 4:** There is no subnet of the external network with the configured
public network id
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to use list commands to list external networks, pre-configured
@@ -340,7 +211,7 @@ Test Case 4 - List IPv6 Networks and Subnets
Short name
-----------
-dovetail.ipv6.tc004.network_subnet_list
+functest.tempest.ipv6_api.network_subnet_list
Use case specification
----------------------
@@ -377,7 +248,7 @@ Test execution
* Test action 8: List all networks, verifying the network id is no longer present
* **Test assertion 4:** The network "id" parameter is not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to use create commands to create network, IPv6 subnet, list
@@ -399,7 +270,7 @@ Test Case 5 - Show Details of an IPv6 Network and Subnet
Short name
----------
-dovetail.ipv6.tc005.network_subnet_show
+functest.tempest.ipv6_api.network_subnet_show
Use case specification
----------------------
@@ -437,7 +308,7 @@ Test execution
* Test action 8: List all networks, verifying the network id is no longer present
* **Test assertion 4:** The "id" parameter is not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to use create commands to create network, IPv6 subnet and show
@@ -459,7 +330,7 @@ Test Case 6 - Create an IPv6 Port in Allowed Allocation Pools
Short name
----------
-dovetail.ipv6.tc006.port_create_in_allocation_pool
+functest.tempest.ipv6_api.port_create_in_allocation_pool
Use case specification
----------------------
@@ -502,7 +373,7 @@ Test execution
* Test action 12: List all networks, verifying the network id is no longer present
* **Test assertion 5:** The "id" parameter is not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to use create commands to create an IPv6 subnet within allowed
@@ -524,7 +395,7 @@ Test Case 7 - Create an IPv6 Port with Empty Security Groups
Short name
-----------
-dovetail.ipv6.tc007.port_create_empty_security_group
+functest.tempest.ipv6_api.port_create_empty_security_group
Use case specification
----------------------
@@ -560,7 +431,7 @@ Test execution
* Test action 10: List all networks, verifying the network id is no longer present
* **Test assertion 4:** The "id" parameter is not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to use create commands to create port with
@@ -581,7 +452,7 @@ Test Case 8 - Create, Update and Delete an IPv6 Port
Short name
----------
-dovetail.ipv6.tc008.port_create_update_delete
+functest.tempest.ipv6_api.port_create_update_delete
Use case specification
----------------------
@@ -618,7 +489,7 @@ Test execution
* Test action 9: List all networks, verifying the network id is no longer present
* **Test assertion 4:** The "id" parameter is not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to use create/update/delete commands to create/update/delete port
@@ -640,7 +511,7 @@ Test Case 9 - List IPv6 Ports
Short name
----------
-dovetail.ipv6.tc009.port_list
+functest.tempest.ipv6_api.port_list
Use case specification
----------------------
@@ -672,7 +543,7 @@ Test execution
* Test action 7: List all networks, verifying the network id is no longer present
* **Test assertion 3:** The "id" parameter is not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to use list commands to list the networks and ports on
@@ -693,7 +564,7 @@ Test Case 10 - Show Key/Valus Details of an IPv6 Port
Short name
----------
-dovetail.ipv6.tc010.port_show_details
+functest.tempest.ipv6_api.port_show_details
Use case specification
----------------------
@@ -731,7 +602,7 @@ Test execution
* Test action 8: List all networks, verifying the network id is no longer present
* **Test assertion 4:** The "id" parameter is not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to use show commands to show port details on the SUT API.
@@ -754,7 +625,7 @@ Test Case 11 - Add Multiple Interfaces for an IPv6 Router
Short name
-----------
-dovetail.ipv6.tc011.router_add_multiple_interface
+functest.tempest.ipv6_api.router_add_multiple_interface
Use case specification
----------------------
@@ -796,7 +667,7 @@ Test execution
* **Test assertion 3:** The interfaces, router, IPv6 subnets and networks ids are not present in the lists
after deleting
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to use bulk create commands to create networks, IPv6 subnets and ports on
@@ -819,7 +690,7 @@ Test Case 12 - Add and Remove an IPv6 Router Interface with port_id
Short name
----------
-dovetail.ipv6.tc012.router_interface_add_remove_with_port
+functest.tempest.ipv6_api.router_interface_add_remove_with_port
Use case specification
----------------------
@@ -857,7 +728,7 @@ Test execution
ones are not found in the list.
* **Test assertion 3:** interfaces, ports, routers, subnets and networks are not found in the lists after deleting
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to use add/remove commands to add/remove router interface to the port,
@@ -879,7 +750,7 @@ Test Case 13 - Add and Remove an IPv6 Router Interface with subnet_id
Short name
----------
-dovetail.ipv6.tc013.router_interface_add_remove
+functest.tempest.ipv6_api.router_interface_add_remove
Use case specification
----------------------
@@ -923,7 +794,7 @@ Test execution
* Test action 13: List all networks, verifying the network id is no longer present
* **Test assertion 7:** The network "id" parameter is not present in the network list
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to add and remove router interface with the subnet id on the
@@ -947,7 +818,7 @@ Test Case 14 - Create, Show, List, Update and Delete an IPv6 router
Short name
----------
-dovetail.ipv6.tc014.router_create_show_list_update_delete
+functest.tempest.ipv6_api.router_create_show_list_update_delete
Use case specification
----------------------
@@ -985,7 +856,7 @@ Test execution
* Test action 7: List all routers, verifying the router id is no longer present
* **Test assertion 8:** The "id" parameter is not present in the router list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to create, show, list, update and delete router on
@@ -1011,7 +882,7 @@ Test Case 15 - Create, List, Update, Show and Delete an IPv6 security group
Short name
----------
-dovetail.ipv6.tc015.security_group_create_list_update_show_delete
+functest.tempest.ipv6_api.security_group_create_list_update_show_delete
Use case specification
----------------------
@@ -1046,7 +917,7 @@ Test execution
* Test action 7: List all security groups, verifying the security group's id is no longer present
* **Test assertion 6:** The "id" parameter is not present in the security group list
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to create list, update, show and delete security group on
@@ -1069,7 +940,7 @@ Test Case 16 - Create, Show and Delete IPv6 security group rule
Short name
----------
-dovetail.ipv6.tc016.security_group_rule_create_show_delete
+functest.tempest.ipv6_api.security_group_rule_create_show_delete
Use case specification
----------------------
@@ -1106,7 +977,7 @@ Test execution
* Test action 8: List all security groups, verifying the security group's id is no longer present
* **Test assertion 4:** The security group "id" parameter is not present in the list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to create, show, list and delete security group rules on
@@ -1127,7 +998,7 @@ Test Case 17 - List IPv6 Security Groups
Short name
----------
-dovetail.ipv6.tc017.security_group_list
+functest.tempest.ipv6_api.security_group_list
Use case specification
----------------------
@@ -1152,7 +1023,7 @@ Test execution
if the default security group exists
* **Test assertion 1:** The default security group is in the list
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test evaluates the ability to list security groups on the SUT API.
@@ -1164,624 +1035,3 @@ Post conditions
---------------
None
-
-----------------------------------------------------------------------------
-Test Case 18 - IPv6 Address Assignment - Dual Stack, SLAAC, DHCPv6 Stateless
-----------------------------------------------------------------------------
-
-Short name
-----------
-
-dovetail.ipv6.tc018.dhcpv6_stateless
-
-Use case specification
-----------------------
-
-This test case evaluates IPv6 address assignment in ipv6_ra_mode 'dhcpv6_stateless'
-and ipv6_address_mode 'dhcpv6_stateless'.
-In this case, guest instance obtains IPv6 address from OpenStack managed radvd
-using SLAAC and optional info from dnsmasq using DHCPv6 stateless. This test case then
-verifies the ping6 available VM can ping the other VM's v4 and v6 addresses
-as well as the v6 subnet's gateway ip in the same network, the reference is
-
-tempest.scenario.test_network_v6.TestGettingAddress.test_dhcp6_stateless_from_os
-
-Test preconditions
-------------------
-
-There should exist a public router or a public network.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-'''''''''''''''
-
-* Test action 1: Create one network, storing the "id" parameter returned in the response
-* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
- parameter returned in the response
-* Test action 3: If there exists a public router, use it as the router. Otherwise,
- use the public network to create a router
-* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
-* Test action 5: Create one IPv6 subnet of the network created in test action 1 in
- ipv6_ra_mode 'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
- storing the "id" parameter returned in the response
-* Test action 6: Connect the IPv6 subnet to the router, using the stored IPv6 subnet id
-* Test action 7: Boot two VMs on this network, storing the "id" parameters returned in the response
-* **Test assertion 1:** The vNIC of each VM gets one v4 address and one v6 address actually assigned
-* **Test assertion 2:** Each VM can ping the other's v4 private address
-* **Test assertion 3:** The ping6 available VM can ping the other's v6 address
- as well as the v6 subnet's gateway ip
-* Test action 8: Delete the 2 VMs created in test action 7, using the stored ids
-* Test action 9: List all VMs, verifying the ids are no longer present
-* **Test assertion 4:** The two "id" parameters are not present in the VM list
-* Test action 10: Delete the IPv4 subnet created in test action 2, using the stored id
-* Test action 11: Delete the IPv6 subnet created in test action 5, using the stored id
-* Test action 12: List all subnets, verifying the ids are no longer present
-* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
-* Test action 13: Delete the network created in test action 1, using the stored id
-* Test action 14: List all networks, verifying the id is no longer present
-* **Test assertion 6:** The "id" parameter is not present in the network list
-
-Pass / fail criteria
-'''''''''''''''''''''
-
-This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode
-'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
-and verify the ping6 available VM can ping the other VM's v4 and v6 addresses as well as
-the v6 subnet's gateway ip in the same network. Specifically it verifies that:
-
-* The IPv6 addresses in mode 'dhcpv6_stateless' assigned successfully
-* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnet's gateway ip
-* All items created using create commands are able to be removed using the returned identifiers
-
-Post conditions
----------------
-
-None
-
---------------------------------------------------------------------------------------
-Test Case 19 - IPv6 Address Assignment - Dual Net, Dual Stack, SLAAC, DHCPv6 Stateless
---------------------------------------------------------------------------------------
-
-Short name
-----------
-
-dovetail.ipv6.tc019.dualnet_dhcpv6_stateless
-
-Use case specification
-----------------------
-
-This test case evaluates IPv6 address assignment in ipv6_ra_mode 'dhcpv6_stateless'
-and ipv6_address_mode 'dhcpv6_stateless'.
-In this case, guest instance obtains IPv6 address from OpenStack managed radvd
-using SLAAC and optional info from dnsmasq using DHCPv6 stateless. This test case then
-verifies the ping6 available VM can ping the other VM's v4 address in one network
-and v6 address in another network as well as the v6 subnet's gateway ip, the reference is
-
-tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_dhcp6_stateless_from_os
-
-Test preconditions
-------------------
-
-There should exists a public router or a public network.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-'''''''''''''''
-
-* Test action 1: Create one network, storing the "id" parameter returned in the response
-* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
- parameter returned in the response
-* Test action 3: If there exists a public router, use it as the router. Otherwise,
- use the public network to create a router
-* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
-* Test action 5: Create another network, storing the "id" parameter returned in the response
-* Test action 6: Create one IPv6 subnet of network created in test action 5 in
- ipv6_ra_mode 'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
- storing the "id" parameter returned in the response
-* Test action 7: Connect the IPv6 subnet to the router, using the stored IPv6 subnet id
-* Test action 8: Boot two VMs on these two networks, storing the "id" parameters returned in the response
-* Test action 9: Turn on 2nd NIC of each VM for the network created in test action 5
-* **Test assertion 1:** The 1st vNIC of each VM gets one v4 address assigned and
- the 2nd vNIC of each VM gets one v6 address actually assigned
-* **Test assertion 2:** Each VM can ping the other's v4 private address
-* **Test assertion 3:** The ping6 available VM can ping the other's v6 address
- as well as the v6 subnet's gateway ip
-* Test action 10: Delete the 2 VMs created in test action 8, using the stored ids
-* Test action 11: List all VMs, verifying the ids are no longer present
-* **Test assertion 4:** The two "id" parameters are not present in the VM list
-* Test action 12: Delete the IPv4 subnet created in test action 2, using the stored id
-* Test action 13: Delete the IPv6 subnet created in test action 6, using the stored id
-* Test action 14: List all subnets, verifying the ids are no longer present
-* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
-* Test action 15: Delete the 2 networks created in test action 1 and 5, using the stored ids
-* Test action 16: List all networks, verifying the ids are no longer present
-* **Test assertion 6:** The two "id" parameters are not present in the network list
-
-Pass / fail criteria
-''''''''''''''''''''
-
-This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'dhcpv6_stateless'
-and ipv6_address_mode 'dhcpv6_stateless', and verify the ping6 available VM can ping
-the other VM's v4 address in one network and v6 address in another network as well as
-the v6 subnet's gateway ip. Specifically it verifies that:
-
-* The IPv6 addresses in mode 'dhcpv6_stateless' assigned successfully
-* The VM can ping the other VM's IPv4 address in one network and IPv6 address in another
- network as well as the v6 subnet's gateway ip
-* All items created using create commands are able to be removed using the returned identifiers
-
-Post conditions
----------------
-
-None
-
------------------------------------------------------------------------------------------------
-Test Case 20 - IPv6 Address Assignment - Multiple Prefixes, Dual Stack, SLAAC, DHCPv6 Stateless
------------------------------------------------------------------------------------------------
-
-Short name
-----------
-
-dovetail.ipv6.tc020.multiple_prefixes_dhcpv6_stateless
-
-Use case specification
-----------------------
-
-This test case evaluates IPv6 address assignment in ipv6_ra_mode 'dhcpv6_stateless'
-and ipv6_address_mode 'dhcpv6_stateless'.
-In this case, guest instance obtains IPv6 addresses from OpenStack managed radvd
-using SLAAC and optional info from dnsmasq using DHCPv6 stateless. This test case then
-verifies the ping6 available VM can ping the other VM's one v4 address and two v6
-addresses with different prefixes as well as the v6 subnets' gateway ips in the
-same network, the reference is
-
-tempest.scenario.test_network_v6.TestGettingAddress.test_multi_prefix_dhcpv6_stateless
-
-Test preconditions
-------------------
-
-There should exist a public router or a public network.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-'''''''''''''''
-
-* Test action 1: Create one network, storing the "id" parameter returned in the response
-* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
- parameter returned in the response
-* Test action 3: If there exists a public router, use it as the router. Otherwise,
- use the public network to create a router
-* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
-* Test action 5: Create two IPv6 subnets of the network created in test action 1 in
- ipv6_ra_mode 'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
- storing the "id" parameters returned in the response
-* Test action 6: Connect the two IPv6 subnets to the router, using the stored IPv6 subnet ids
-* Test action 7: Boot two VMs on this network, storing the "id" parameters returned in the response
-* **Test assertion 1:** The vNIC of each VM gets one v4 address and two v6 addresses with
- different prefixes actually assigned
-* **Test assertion 2:** Each VM can ping the other's v4 private address
-* **Test assertion 3:** The ping6 available VM can ping the other's v6 addresses
- as well as the v6 subnets' gateway ips
-* Test action 8: Delete the 2 VMs created in test action 7, using the stored ids
-* Test action 9: List all VMs, verifying the ids are no longer present
-* **Test assertion 4:** The two "id" parameters are not present in the VM list
-* Test action 10: Delete the IPv4 subnet created in test action 2, using the stored id
-* Test action 11: Delete two IPv6 subnets created in test action 5, using the stored ids
-* Test action 12: List all subnets, verifying the ids are no longer present
-* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
-* Test action 13: Delete the network created in test action 1, using the stored id
-* Test action 14: List all networks, verifying the id is no longer present
-* **Test assertion 6:** The "id" parameter is not present in the network list
-
-Pass / fail criteria
-'''''''''''''''''''''
-
-This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'dhcpv6_stateless'
-and ipv6_address_mode 'dhcpv6_stateless',
-and verify the ping6 available VM can ping the other VM's v4 address and two
-v6 addresses with different prefixes as well as the v6 subnets' gateway ips in the same network.
-Specifically it verifies that:
-
-* The different prefixes IPv6 addresses in mode 'dhcpv6_stateless' assigned successfully
-* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnets' gateway ips
-* All items created using create commands are able to be removed using the returned identifiers
-
-Post conditions
----------------
-
-None
-
----------------------------------------------------------------------------------------------------------
-Test Case 21 - IPv6 Address Assignment - Dual Net, Multiple Prefixes, Dual Stack, SLAAC, DHCPv6 Stateless
----------------------------------------------------------------------------------------------------------
-
-Short name
-----------
-
-dovetail.ipv6.tc021.dualnet_multiple_prefixes_dhcpv6_stateless
-
-Use case specification
-----------------------
-
-This test case evaluates IPv6 address assignment in ipv6_ra_mode 'dhcpv6_stateless'
-and ipv6_address_mode 'dhcpv6_stateless'.
-In this case, guest instance obtains IPv6 addresses from OpenStack managed radvd
-using SLAAC and optional info from dnsmasq using DHCPv6 stateless. This test case then
-verifies the ping6 available VM can ping the other VM's v4 address in one network
-and two v6 addresses with different prefixes in another network as well as the
-v6 subnets' gateway ips, the reference is
-
-tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_multi_prefix_dhcpv6_stateless
-
-Test preconditions
-------------------
-
-There should exist a public router or a public network.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-'''''''''''''''
-
-* Test action 1: Create one network, storing the "id" parameter returned in the response
-* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
- parameter returned in the response
-* Test action 3: If there exists a public router, use it as the router. Otherwise,
- use the public network to create a router
-* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
-* Test action 5: Create another network, storing the "id" parameter returned in the response
-* Test action 6: Create two IPv6 subnets of network created in test action 5 in
- ipv6_ra_mode 'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
- storing the "id" parameters returned in the response
-* Test action 7: Connect the two IPv6 subnets to the router, using the stored IPv6 subnet ids
-* Test action 8: Boot two VMs on these two networks, storing the "id" parameters returned in the response
-* Test action 9: Turn on 2nd NIC of each VM for the network created in test action 5
-* **Test assertion 1:** The vNIC of each VM gets one v4 address and two v6 addresses
- with different prefixes actually assigned
-* **Test assertion 2:** Each VM can ping the other's v4 private address
-* **Test assertion 3:** The ping6 available VM can ping the other's v6 addresses
- as well as the v6 subnets' gateway ips
-* Test action 10: Delete the 2 VMs created in test action 8, using the stored ids
-* Test action 11: List all VMs, verifying the ids are no longer present
-* **Test assertion 4:** The two "id" parameters are not present in the VM list
-* Test action 12: Delete the IPv4 subnet created in test action 2, using the stored id
-* Test action 13: Delete two IPv6 subnets created in test action 6, using the stored ids
-* Test action 14: List all subnets, verifying the ids are no longer present
-* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
-* Test action 15: Delete the 2 networks created in test action 1 and 5, using the stored ids
-* Test action 16: List all networks, verifying the ids are no longer present
-* **Test assertion 6:** The two "id" parameters are not present in the network list
-
-Pass / fail criteria
-'''''''''''''''''''''
-
-This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'dhcpv6_stateless'
-and ipv6_address_mode 'dhcpv6_stateless',
-and verify the ping6 available VM can ping the other VM's v4 address in one network and two
-v6 addresses with different prefixes in another network as well as the v6 subnets'
-gateway ips. Specifically it verifies that:
-
-* The IPv6 addresses in mode 'dhcpv6_stateless' assigned successfully
-* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnets' gateway ips
-* All items created using create commands are able to be removed using the returned identifiers
-
-Post conditions
----------------
-
-None
-
-----------------------------------------------------------
-Test Case 22 - IPv6 Address Assignment - Dual Stack, SLAAC
-----------------------------------------------------------
-
-Short name
-----------
-
-dovetail.ipv6.tc022.slaac
-
-Use case specification
-----------------------
-
-This test case evaluates IPv6 address assignment in ipv6_ra_mode 'slaac' and
-ipv6_address_mode 'slaac'.
-In this case, guest instance obtains IPv6 address from OpenStack managed radvd
-using SLAAC. This test case then verifies the ping6 available VM can ping the other
-VM's v4 and v6 addresses as well as the v6 subnet's gateway ip in the
-same network, the reference is
-
-tempest.scenario.test_network_v6.TestGettingAddress.test_slaac_from_os
-
-Test preconditions
-------------------
-
-There should exist a public router or a public network.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-'''''''''''''''
-
-* Test action 1: Create one network, storing the "id" parameter returned in the response
-* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
- parameter returned in the response
-* Test action 3: If there exists a public router, use it as the router. Otherwise,
- use the public network to create a router
-* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
-* Test action 5: Create one IPv6 subnet of the network created in test action 1 in
- ipv6_ra_mode 'slaac' and ipv6_address_mode 'slaac', storing the "id" parameter returned in the response
-* Test action 6: Connect the IPv6 subnet to the router, using the stored IPv6 subnet id
-* Test action 7: Boot two VMs on this network, storing the "id" parameters returned in the response
-* **Test assertion 1:** The vNIC of each VM gets one v4 address and one v6 address actually assigned
-* **Test assertion 2:** Each VM can ping the other's v4 private address
-* **Test assertion 3:** The ping6 available VM can ping the other's v6 address
- as well as the v6 subnet's gateway ip
-* Test action 8: Delete the 2 VMs created in test action 7, using the stored ids
-* Test action 9: List all VMs, verifying the ids are no longer present
-* **Test assertion 4:** The two "id" parameters are not present in the VM list
-* Test action 10: Delete the IPv4 subnet created in test action 2, using the stored id
-* Test action 11: Delete the IPv6 subnet created in test action 5, using the stored id
-* Test action 12: List all subnets, verifying the ids are no longer present
-* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
-* Test action 13: Delete the network created in test action 1, using the stored id
-* Test action 14: List all networks, verifying the id is no longer present
-* **Test assertion 6:** The "id" parameter is not present in the network list
-
-Pass / fail criteria
-'''''''''''''''''''''
-
-This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'slaac'
-and ipv6_address_mode 'slaac',
-and verify the ping6 available VM can ping the other VM's v4 and v6 addresses as well as
-the v6 subnet's gateway ip in the same network. Specifically it verifies that:
-
-* The IPv6 addresses in mode 'slaac' assigned successfully
-* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnet's gateway ip
-* All items created using create commands are able to be removed using the returned identifiers
-
-Post conditions
----------------
-
-None
-
---------------------------------------------------------------------
-Test Case 23 - IPv6 Address Assignment - Dual Net, Dual Stack, SLAAC
---------------------------------------------------------------------
-
-Short name
-----------
-
-dovetail.ipv6.tc023.dualnet_slaac
-
-Use case specification
-----------------------
-
-This test case evaluates IPv6 address assignment in ipv6_ra_mode 'slaac' and
-ipv6_address_mode 'slaac'.
-In this case, guest instance obtains IPv6 address from OpenStack managed radvd
-using SLAAC. This test case then verifies the ping6 available VM can ping the other
-VM's v4 address in one network and v6 address in another network as well as the
-v6 subnet's gateway ip, the reference is
-
-tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_slaac_from_os
-
-Test preconditions
-------------------
-
-There should exist a public router or a public network.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-'''''''''''''''
-
-* Test action 1: Create one network, storing the "id" parameter returned in the response
-* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
- parameter returned in the response
-* Test action 3: If there exists a public router, use it as the router. Otherwise,
- use the public network to create a router
-* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
-* Test action 5: Create another network, storing the "id" parameter returned in the response
-* Test action 6: Create one IPv6 subnet of network created in test action 5 in
- ipv6_ra_mode 'slaac' and ipv6_address_mode 'slaac', storing the "id" parameter returned in the response
-* Test action 7: Connect the IPv6 subnet to the router, using the stored IPv6 subnet id
-* Test action 8: Boot two VMs on these two networks, storing the "id" parameters returned in the response
-* Test action 9: Turn on 2nd NIC of each VM for the network created in test action 5
-* **Test assertion 1:** The 1st vNIC of each VM gets one v4 address assigned and
- the 2nd vNIC of each VM gets one v6 address actually assigned
-* **Test assertion 2:** Each VM can ping the other's v4 private address
-* **Test assertion 3:** The ping6 available VM can ping the other's v6 address
- as well as the v6 subnet's gateway ip
-* Test action 10: Delete the 2 VMs created in test action 8, using the stored ids
-* Test action 11: List all VMs, verifying the ids are no longer present
-* **Test assertion 4:** The two "id" parameters are not present in the VM list
-* Test action 12: Delete the IPv4 subnet created in test action 2, using the stored id
-* Test action 13: Delete the IPv6 subnet created in test action 6, using the stored id
-* Test action 14: List all subnets, verifying the ids are no longer present
-* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
-* Test action 15: Delete the 2 networks created in test action 1 and 5, using the stored ids
-* Test action 16: List all networks, verifying the ids are no longer present
-* **Test assertion 6:** The two "id" parameters are not present in the network list
-
-Pass / fail criteria
-'''''''''''''''''''''
-
-This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'slaac'
-and ipv6_address_mode 'slaac',
-and verify the ping6 available VM can ping the other VM's v4 address in one network and
-v6 address in another network as well as the v6 subnet's gateway ip. Specifically it verifies that:
-
-* The IPv6 addresses in mode 'slaac' assigned successfully
-* The VM can ping the other VM's IPv4 address in one network and IPv6 address
- in another network as well as the v6 subnet's gateway ip
-* All items created using create commands are able to be removed using the returned identifiers
-
-Post conditions
----------------
-
-None
-
------------------------------------------------------------------------------
-Test Case 24 - IPv6 Address Assignment - Multiple Prefixes, Dual Stack, SLAAC
------------------------------------------------------------------------------
-
-Short name
-----------
-
-dovetail.ipv6.tc024.multiple_prefixes_slaac
-
-Use case specification
-----------------------
-
-This test case evaluates IPv6 address assignment in ipv6_ra_mode 'slaac' and
-ipv6_address_mode 'slaac'.
-In this case, guest instance obtains IPv6 addresses from OpenStack managed radvd
-using SLAAC. This test case then verifies the ping6 available VM can ping the other
-VM's one v4 address and two v6 addresses with different prefixes as well as the v6
-subnets' gateway ips in the same network, the reference is
-
-tempest.scenario.test_network_v6.TestGettingAddress.test_multi_prefix_slaac
-
-Test preconditions
-------------------
-
-There should exists a public router or a public network.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-'''''''''''''''
-
-* Test action 1: Create one network, storing the "id" parameter returned in the response
-* Test action 2: Create one IPv4 subnet of the created network, storing the "id
- parameter returned in the response
-* Test action 3: If there exists a public router, use it as the router. Otherwise,
- use the public network to create a router
-* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
-* Test action 5: Create two IPv6 subnets of the network created in test action 1 in
- ipv6_ra_mode 'slaac' and ipv6_address_mode 'slaac', storing the "id" parameters returned in the response
-* Test action 6: Connect the two IPv6 subnets to the router, using the stored IPv6 subnet ids
-* Test action 7: Boot two VMs on this network, storing the "id" parameters returned in the response
-* **Test assertion 1:** The vNIC of each VM gets one v4 address and two v6 addresses with
- different prefixes actually assigned
-* **Test assertion 2:** Each VM can ping the other's v4 private address
-* **Test assertion 3:** The ping6 available VM can ping the other's v6 addresses
- as well as the v6 subnets' gateway ips
-* Test action 8: Delete the 2 VMs created in test action 7, using the stored ids
-* Test action 9: List all VMs, verifying the ids are no longer present
-* **Test assertion 4:** The two "id" parameters are not present in the VM list
-* Test action 10: Delete the IPv4 subnet created in test action 2, using the stored id
-* Test action 11: Delete two IPv6 subnets created in test action 5, using the stored ids
-* Test action 12: List all subnets, verifying the ids are no longer present
-* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
-* Test action 13: Delete the network created in test action 1, using the stored id
-* Test action 14: List all networks, verifying the id is no longer present
-* **Test assertion 6:** The "id" parameter is not present in the network list
-
-Pass / fail criteria
-'''''''''''''''''''''
-
-This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'slaac'
-and ipv6_address_mode 'slaac',
-and verify the ping6 available VM can ping the other VM's v4 address and two
-v6 addresses with different prefixes as well as the v6 subnets' gateway ips in the same network.
-Specifically it verifies that:
-
-* The different prefixes IPv6 addresses in mode 'slaac' assigned successfully
-* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnets' gateway ips
-* All items created using create commands are able to be removed using the returned identifiers
-
-Post conditions
----------------
-
-None
-
----------------------------------------------------------------------------------------
-Test Case 25 - IPv6 Address Assignment - Dual Net, Dual Stack, Multiple Prefixes, SLAAC
----------------------------------------------------------------------------------------
-
-Short name
-----------
-
-dovetail.ipv6.tc025.dualnet_multiple_prefixes_slaac
-
-Use case specification
-----------------------
-
-This test case evaluates IPv6 address assignment in ipv6_ra_mode 'slaac' and
-ipv6_address_mode 'slaac'.
-In this case, guest instance obtains IPv6 addresses from OpenStack managed radvd
-using SLAAC. This test case then verifies the ping6 available VM can ping the other
-VM's v4 address in one network and two v6 addresses with different prefixes in another
-network as well as the v6 subnets' gateway ips, the reference is
-
-tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_multi_prefix_slaac
-
-Test preconditions
-------------------
-
-There should exist a public router or a public network.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-'''''''''''''''
-
-* Test action 1: Create one network, storing the "id" parameter returned in the response
-* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
- parameter returned in the response
-* Test action 3: If there exists a public router, use it as the router. Otherwise,
- use the public network to create a router
-* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
-* Test action 5: Create another network, storing the "id" parameter returned in the response
-* Test action 6: Create two IPv6 subnets of network created in test action 5 in
- ipv6_ra_mode 'slaac' and ipv6_address_mode 'slaac', storing the "id" parameters returned in the response
-* Test action 7: Connect the two IPv6 subnets to the router, using the stored IPv6 subnet ids
-* Test action 8: Boot two VMs on these two networks, storing the "id" parameters returned in the response
-* Test action 9: Turn on 2nd NIC of each VM for the network created in test action 5
-* **Test assertion 1:** The vNIC of each VM gets one v4 address and two v6 addresses
- with different prefixes actually assigned
-* **Test assertion 2:** Each VM can ping the other's v4 private address
-* **Test assertion 3:** The ping6 available VM can ping the other's v6 addresses
- as well as the v6 subnets' gateway ips
-* Test action 10: Delete the 2 VMs created in test action 8, using the stored ids
-* Test action 11: List all VMs, verifying the ids are no longer present
-* **Test assertion 4:** The two "id" parameters are not present in the VM list
-* Test action 12: Delete the IPv4 subnet created in test action 2, using the stored id
-* Test action 13: Delete two IPv6 subnets created in test action 6, using the stored ids
-* Test action 14: List all subnets, verifying the ids are no longer present
-* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
-* Test action 15: Delete the 2 networks created in test action 1 and 5, using the stored ids
-* Test action 16: List all networks, verifying the ids are no longer present
-* **Test assertion 6:** The two "id" parameters are not present in the network list
-
-Pass / fail criteria
-'''''''''''''''''''''
-
-This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'slaac'
-and ipv6_address_mode 'slaac',
-and verify the ping6 available VM can ping the other VM's v4 address in one network and two
-v6 addresses with different prefixes in another network as well as the v6 subnets' gateway ips.
-Specifically it verifies that:
-
-* The IPv6 addresses in mode 'slaac' assigned successfully
-* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnets' gateway ips
-* All items created using create commands are able to be removed using the returned identifiers
-
-Post conditions
----------------
-
-None
-
-
-
diff --git a/docs/testing/user/testspecification/tempest_ipv6/ipv6_scenario.rst b/docs/testing/user/testspecification/tempest_ipv6/ipv6_scenario.rst
new file mode 100644
index 00000000..5871321d
--- /dev/null
+++ b/docs/testing/user/testspecification/tempest_ipv6/ipv6_scenario.rst
@@ -0,0 +1,624 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV
+
+----------------------------------------------------------------------------
+Test Case 1 - IPv6 Address Assignment - Dual Stack, SLAAC, DHCPv6 Stateless
+----------------------------------------------------------------------------
+
+Short name
+----------
+
+functest.tempest.ipv6_scenario.dhcpv6_stateless
+
+Use case specification
+----------------------
+
+This test case evaluates IPv6 address assignment in ipv6_ra_mode 'dhcpv6_stateless'
+and ipv6_address_mode 'dhcpv6_stateless'.
+In this case, guest instance obtains IPv6 address from OpenStack managed radvd
+using SLAAC and optional info from dnsmasq using DHCPv6 stateless. This test case then
+verifies the ping6 available VM can ping the other VM's v4 and v6 addresses
+as well as the v6 subnet's gateway ip in the same network, the reference is
+
+tempest.scenario.test_network_v6.TestGettingAddress.test_dhcp6_stateless_from_os
+
+Test preconditions
+------------------
+
+There should exist a public router or a public network.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+'''''''''''''''
+
+* Test action 1: Create one network, storing the "id" parameter returned in the response
+* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
+ parameter returned in the response
+* Test action 3: If there exists a public router, use it as the router. Otherwise,
+ use the public network to create a router
+* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
+* Test action 5: Create one IPv6 subnet of the network created in test action 1 in
+ ipv6_ra_mode 'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
+ storing the "id" parameter returned in the response
+* Test action 6: Connect the IPv6 subnet to the router, using the stored IPv6 subnet id
+* Test action 7: Boot two VMs on this network, storing the "id" parameters returned in the response
+* **Test assertion 1:** The vNIC of each VM gets one v4 address and one v6 address actually assigned
+* **Test assertion 2:** Each VM can ping the other's v4 private address
+* **Test assertion 3:** The ping6 available VM can ping the other's v6 address
+ as well as the v6 subnet's gateway ip
+* Test action 8: Delete the 2 VMs created in test action 7, using the stored ids
+* Test action 9: List all VMs, verifying the ids are no longer present
+* **Test assertion 4:** The two "id" parameters are not present in the VM list
+* Test action 10: Delete the IPv4 subnet created in test action 2, using the stored id
+* Test action 11: Delete the IPv6 subnet created in test action 5, using the stored id
+* Test action 12: List all subnets, verifying the ids are no longer present
+* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
+* Test action 13: Delete the network created in test action 1, using the stored id
+* Test action 14: List all networks, verifying the id is no longer present
+* **Test assertion 6:** The "id" parameter is not present in the network list
+
+Pass / Fail criteria
+'''''''''''''''''''''
+
+This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode
+'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
+and verify the ping6 available VM can ping the other VM's v4 and v6 addresses as well as
+the v6 subnet's gateway ip in the same network. Specifically it verifies that:
+
+* The IPv6 addresses in mode 'dhcpv6_stateless' assigned successfully
+* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnet's gateway ip
+* All items created using create commands are able to be removed using the returned identifiers
+
+Post conditions
+---------------
+
+None
+
+--------------------------------------------------------------------------------------
+Test Case 2 - IPv6 Address Assignment - Dual Net, Dual Stack, SLAAC, DHCPv6 Stateless
+--------------------------------------------------------------------------------------
+
+Short name
+----------
+
+functest.tempest.ipv6_scenario.dualnet_dhcpv6_stateless
+
+Use case specification
+----------------------
+
+This test case evaluates IPv6 address assignment in ipv6_ra_mode 'dhcpv6_stateless'
+and ipv6_address_mode 'dhcpv6_stateless'.
+In this case, guest instance obtains IPv6 address from OpenStack managed radvd
+using SLAAC and optional info from dnsmasq using DHCPv6 stateless. This test case then
+verifies the ping6 available VM can ping the other VM's v4 address in one network
+and v6 address in another network as well as the v6 subnet's gateway ip, the reference is
+
+tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_dhcp6_stateless_from_os
+
+Test preconditions
+------------------
+
+There should exists a public router or a public network.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+'''''''''''''''
+
+* Test action 1: Create one network, storing the "id" parameter returned in the response
+* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
+ parameter returned in the response
+* Test action 3: If there exists a public router, use it as the router. Otherwise,
+ use the public network to create a router
+* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
+* Test action 5: Create another network, storing the "id" parameter returned in the response
+* Test action 6: Create one IPv6 subnet of network created in test action 5 in
+ ipv6_ra_mode 'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
+ storing the "id" parameter returned in the response
+* Test action 7: Connect the IPv6 subnet to the router, using the stored IPv6 subnet id
+* Test action 8: Boot two VMs on these two networks, storing the "id" parameters returned in the response
+* Test action 9: Turn on 2nd NIC of each VM for the network created in test action 5
+* **Test assertion 1:** The 1st vNIC of each VM gets one v4 address assigned and
+ the 2nd vNIC of each VM gets one v6 address actually assigned
+* **Test assertion 2:** Each VM can ping the other's v4 private address
+* **Test assertion 3:** The ping6 available VM can ping the other's v6 address
+ as well as the v6 subnet's gateway ip
+* Test action 10: Delete the 2 VMs created in test action 8, using the stored ids
+* Test action 11: List all VMs, verifying the ids are no longer present
+* **Test assertion 4:** The two "id" parameters are not present in the VM list
+* Test action 12: Delete the IPv4 subnet created in test action 2, using the stored id
+* Test action 13: Delete the IPv6 subnet created in test action 6, using the stored id
+* Test action 14: List all subnets, verifying the ids are no longer present
+* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
+* Test action 15: Delete the 2 networks created in test action 1 and 5, using the stored ids
+* Test action 16: List all networks, verifying the ids are no longer present
+* **Test assertion 6:** The two "id" parameters are not present in the network list
+
+Pass / Fail criteria
+''''''''''''''''''''
+
+This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'dhcpv6_stateless'
+and ipv6_address_mode 'dhcpv6_stateless', and verify the ping6 available VM can ping
+the other VM's v4 address in one network and v6 address in another network as well as
+the v6 subnet's gateway ip. Specifically it verifies that:
+
+* The IPv6 addresses in mode 'dhcpv6_stateless' assigned successfully
+* The VM can ping the other VM's IPv4 address in one network and IPv6 address in another
+ network as well as the v6 subnet's gateway ip
+* All items created using create commands are able to be removed using the returned identifiers
+
+Post conditions
+---------------
+
+None
+
+-----------------------------------------------------------------------------------------------
+Test Case 3 - IPv6 Address Assignment - Multiple Prefixes, Dual Stack, SLAAC, DHCPv6 Stateless
+-----------------------------------------------------------------------------------------------
+
+Short name
+----------
+
+functest.tempest.ipv6_scenario.multiple_prefixes_dhcpv6_stateless
+
+Use case specification
+----------------------
+
+This test case evaluates IPv6 address assignment in ipv6_ra_mode 'dhcpv6_stateless'
+and ipv6_address_mode 'dhcpv6_stateless'.
+In this case, guest instance obtains IPv6 addresses from OpenStack managed radvd
+using SLAAC and optional info from dnsmasq using DHCPv6 stateless. This test case then
+verifies the ping6 available VM can ping the other VM's one v4 address and two v6
+addresses with different prefixes as well as the v6 subnets' gateway ips in the
+same network, the reference is
+
+tempest.scenario.test_network_v6.TestGettingAddress.test_multi_prefix_dhcpv6_stateless
+
+Test preconditions
+------------------
+
+There should exist a public router or a public network.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+'''''''''''''''
+
+* Test action 1: Create one network, storing the "id" parameter returned in the response
+* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
+ parameter returned in the response
+* Test action 3: If there exists a public router, use it as the router. Otherwise,
+ use the public network to create a router
+* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
+* Test action 5: Create two IPv6 subnets of the network created in test action 1 in
+ ipv6_ra_mode 'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
+ storing the "id" parameters returned in the response
+* Test action 6: Connect the two IPv6 subnets to the router, using the stored IPv6 subnet ids
+* Test action 7: Boot two VMs on this network, storing the "id" parameters returned in the response
+* **Test assertion 1:** The vNIC of each VM gets one v4 address and two v6 addresses with
+ different prefixes actually assigned
+* **Test assertion 2:** Each VM can ping the other's v4 private address
+* **Test assertion 3:** The ping6 available VM can ping the other's v6 addresses
+ as well as the v6 subnets' gateway ips
+* Test action 8: Delete the 2 VMs created in test action 7, using the stored ids
+* Test action 9: List all VMs, verifying the ids are no longer present
+* **Test assertion 4:** The two "id" parameters are not present in the VM list
+* Test action 10: Delete the IPv4 subnet created in test action 2, using the stored id
+* Test action 11: Delete two IPv6 subnets created in test action 5, using the stored ids
+* Test action 12: List all subnets, verifying the ids are no longer present
+* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
+* Test action 13: Delete the network created in test action 1, using the stored id
+* Test action 14: List all networks, verifying the id is no longer present
+* **Test assertion 6:** The "id" parameter is not present in the network list
+
+Pass / Fail criteria
+'''''''''''''''''''''
+
+This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'dhcpv6_stateless'
+and ipv6_address_mode 'dhcpv6_stateless',
+and verify the ping6 available VM can ping the other VM's v4 address and two
+v6 addresses with different prefixes as well as the v6 subnets' gateway ips in the same network.
+Specifically it verifies that:
+
+* The different prefixes IPv6 addresses in mode 'dhcpv6_stateless' assigned successfully
+* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnets' gateway ips
+* All items created using create commands are able to be removed using the returned identifiers
+
+Post conditions
+---------------
+
+None
+
+---------------------------------------------------------------------------------------------------------
+Test Case 4 - IPv6 Address Assignment - Dual Net, Multiple Prefixes, Dual Stack, SLAAC, DHCPv6 Stateless
+---------------------------------------------------------------------------------------------------------
+
+Short name
+----------
+
+functest.tempest.ipv6_scenario.dualnet_multiple_prefixes_dhcpv6_stateless
+
+Use case specification
+----------------------
+
+This test case evaluates IPv6 address assignment in ipv6_ra_mode 'dhcpv6_stateless'
+and ipv6_address_mode 'dhcpv6_stateless'.
+In this case, guest instance obtains IPv6 addresses from OpenStack managed radvd
+using SLAAC and optional info from dnsmasq using DHCPv6 stateless. This test case then
+verifies the ping6 available VM can ping the other VM's v4 address in one network
+and two v6 addresses with different prefixes in another network as well as the
+v6 subnets' gateway ips, the reference is
+
+tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_multi_prefix_dhcpv6_stateless
+
+Test preconditions
+------------------
+
+There should exist a public router or a public network.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+'''''''''''''''
+
+* Test action 1: Create one network, storing the "id" parameter returned in the response
+* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
+ parameter returned in the response
+* Test action 3: If there exists a public router, use it as the router. Otherwise,
+ use the public network to create a router
+* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
+* Test action 5: Create another network, storing the "id" parameter returned in the response
+* Test action 6: Create two IPv6 subnets of network created in test action 5 in
+ ipv6_ra_mode 'dhcpv6_stateless' and ipv6_address_mode 'dhcpv6_stateless',
+ storing the "id" parameters returned in the response
+* Test action 7: Connect the two IPv6 subnets to the router, using the stored IPv6 subnet ids
+* Test action 8: Boot two VMs on these two networks, storing the "id" parameters returned in the response
+* Test action 9: Turn on 2nd NIC of each VM for the network created in test action 5
+* **Test assertion 1:** The vNIC of each VM gets one v4 address and two v6 addresses
+ with different prefixes actually assigned
+* **Test assertion 2:** Each VM can ping the other's v4 private address
+* **Test assertion 3:** The ping6 available VM can ping the other's v6 addresses
+ as well as the v6 subnets' gateway ips
+* Test action 10: Delete the 2 VMs created in test action 8, using the stored ids
+* Test action 11: List all VMs, verifying the ids are no longer present
+* **Test assertion 4:** The two "id" parameters are not present in the VM list
+* Test action 12: Delete the IPv4 subnet created in test action 2, using the stored id
+* Test action 13: Delete two IPv6 subnets created in test action 6, using the stored ids
+* Test action 14: List all subnets, verifying the ids are no longer present
+* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
+* Test action 15: Delete the 2 networks created in test action 1 and 5, using the stored ids
+* Test action 16: List all networks, verifying the ids are no longer present
+* **Test assertion 6:** The two "id" parameters are not present in the network list
+
+Pass / Fail criteria
+'''''''''''''''''''''
+
+This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'dhcpv6_stateless'
+and ipv6_address_mode 'dhcpv6_stateless',
+and verify the ping6 available VM can ping the other VM's v4 address in one network and two
+v6 addresses with different prefixes in another network as well as the v6 subnets'
+gateway ips. Specifically it verifies that:
+
+* The IPv6 addresses in mode 'dhcpv6_stateless' assigned successfully
+* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnets' gateway ips
+* All items created using create commands are able to be removed using the returned identifiers
+
+Post conditions
+---------------
+
+None
+
+----------------------------------------------------------
+Test Case 5 - IPv6 Address Assignment - Dual Stack, SLAAC
+----------------------------------------------------------
+
+Short name
+----------
+
+functest.tempest.ipv6_scenario.slaac
+
+Use case specification
+----------------------
+
+This test case evaluates IPv6 address assignment in ipv6_ra_mode 'slaac' and
+ipv6_address_mode 'slaac'.
+In this case, guest instance obtains IPv6 address from OpenStack managed radvd
+using SLAAC. This test case then verifies the ping6 available VM can ping the other
+VM's v4 and v6 addresses as well as the v6 subnet's gateway ip in the
+same network, the reference is
+
+tempest.scenario.test_network_v6.TestGettingAddress.test_slaac_from_os
+
+Test preconditions
+------------------
+
+There should exist a public router or a public network.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+'''''''''''''''
+
+* Test action 1: Create one network, storing the "id" parameter returned in the response
+* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
+ parameter returned in the response
+* Test action 3: If there exists a public router, use it as the router. Otherwise,
+ use the public network to create a router
+* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
+* Test action 5: Create one IPv6 subnet of the network created in test action 1 in
+ ipv6_ra_mode 'slaac' and ipv6_address_mode 'slaac', storing the "id" parameter returned in the response
+* Test action 6: Connect the IPv6 subnet to the router, using the stored IPv6 subnet id
+* Test action 7: Boot two VMs on this network, storing the "id" parameters returned in the response
+* **Test assertion 1:** The vNIC of each VM gets one v4 address and one v6 address actually assigned
+* **Test assertion 2:** Each VM can ping the other's v4 private address
+* **Test assertion 3:** The ping6 available VM can ping the other's v6 address
+ as well as the v6 subnet's gateway ip
+* Test action 8: Delete the 2 VMs created in test action 7, using the stored ids
+* Test action 9: List all VMs, verifying the ids are no longer present
+* **Test assertion 4:** The two "id" parameters are not present in the VM list
+* Test action 10: Delete the IPv4 subnet created in test action 2, using the stored id
+* Test action 11: Delete the IPv6 subnet created in test action 5, using the stored id
+* Test action 12: List all subnets, verifying the ids are no longer present
+* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
+* Test action 13: Delete the network created in test action 1, using the stored id
+* Test action 14: List all networks, verifying the id is no longer present
+* **Test assertion 6:** The "id" parameter is not present in the network list
+
+Pass / Fail criteria
+'''''''''''''''''''''
+
+This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'slaac'
+and ipv6_address_mode 'slaac',
+and verify the ping6 available VM can ping the other VM's v4 and v6 addresses as well as
+the v6 subnet's gateway ip in the same network. Specifically it verifies that:
+
+* The IPv6 addresses in mode 'slaac' assigned successfully
+* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnet's gateway ip
+* All items created using create commands are able to be removed using the returned identifiers
+
+Post conditions
+---------------
+
+None
+
+--------------------------------------------------------------------
+Test Case 6 - IPv6 Address Assignment - Dual Net, Dual Stack, SLAAC
+--------------------------------------------------------------------
+
+Short name
+----------
+
+functest.tempest.ipv6_scenario.dualnet_slaac
+
+Use case specification
+----------------------
+
+This test case evaluates IPv6 address assignment in ipv6_ra_mode 'slaac' and
+ipv6_address_mode 'slaac'.
+In this case, guest instance obtains IPv6 address from OpenStack managed radvd
+using SLAAC. This test case then verifies the ping6 available VM can ping the other
+VM's v4 address in one network and v6 address in another network as well as the
+v6 subnet's gateway ip, the reference is
+
+tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_slaac_from_os
+
+Test preconditions
+------------------
+
+There should exist a public router or a public network.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+'''''''''''''''
+
+* Test action 1: Create one network, storing the "id" parameter returned in the response
+* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
+ parameter returned in the response
+* Test action 3: If there exists a public router, use it as the router. Otherwise,
+ use the public network to create a router
+* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
+* Test action 5: Create another network, storing the "id" parameter returned in the response
+* Test action 6: Create one IPv6 subnet of network created in test action 5 in
+ ipv6_ra_mode 'slaac' and ipv6_address_mode 'slaac', storing the "id" parameter returned in the response
+* Test action 7: Connect the IPv6 subnet to the router, using the stored IPv6 subnet id
+* Test action 8: Boot two VMs on these two networks, storing the "id" parameters returned in the response
+* Test action 9: Turn on 2nd NIC of each VM for the network created in test action 5
+* **Test assertion 1:** The 1st vNIC of each VM gets one v4 address assigned and
+ the 2nd vNIC of each VM gets one v6 address actually assigned
+* **Test assertion 2:** Each VM can ping the other's v4 private address
+* **Test assertion 3:** The ping6 available VM can ping the other's v6 address
+ as well as the v6 subnet's gateway ip
+* Test action 10: Delete the 2 VMs created in test action 8, using the stored ids
+* Test action 11: List all VMs, verifying the ids are no longer present
+* **Test assertion 4:** The two "id" parameters are not present in the VM list
+* Test action 12: Delete the IPv4 subnet created in test action 2, using the stored id
+* Test action 13: Delete the IPv6 subnet created in test action 6, using the stored id
+* Test action 14: List all subnets, verifying the ids are no longer present
+* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
+* Test action 15: Delete the 2 networks created in test action 1 and 5, using the stored ids
+* Test action 16: List all networks, verifying the ids are no longer present
+* **Test assertion 6:** The two "id" parameters are not present in the network list
+
+Pass / Fail criteria
+'''''''''''''''''''''
+
+This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'slaac'
+and ipv6_address_mode 'slaac',
+and verify the ping6 available VM can ping the other VM's v4 address in one network and
+v6 address in another network as well as the v6 subnet's gateway ip. Specifically it verifies that:
+
+* The IPv6 addresses in mode 'slaac' assigned successfully
+* The VM can ping the other VM's IPv4 address in one network and IPv6 address
+ in another network as well as the v6 subnet's gateway ip
+* All items created using create commands are able to be removed using the returned identifiers
+
+Post conditions
+---------------
+
+None
+
+-----------------------------------------------------------------------------
+Test Case 7 - IPv6 Address Assignment - Multiple Prefixes, Dual Stack, SLAAC
+-----------------------------------------------------------------------------
+
+Short name
+----------
+
+functest.tempest.ipv6_scenario.multiple_prefixes_slaac
+
+Use case specification
+----------------------
+
+This test case evaluates IPv6 address assignment in ipv6_ra_mode 'slaac' and
+ipv6_address_mode 'slaac'.
+In this case, guest instance obtains IPv6 addresses from OpenStack managed radvd
+using SLAAC. This test case then verifies the ping6 available VM can ping the other
+VM's one v4 address and two v6 addresses with different prefixes as well as the v6
+subnets' gateway ips in the same network, the reference is
+
+tempest.scenario.test_network_v6.TestGettingAddress.test_multi_prefix_slaac
+
+Test preconditions
+------------------
+
+There should exists a public router or a public network.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+'''''''''''''''
+
+* Test action 1: Create one network, storing the "id" parameter returned in the response
+* Test action 2: Create one IPv4 subnet of the created network, storing the "id
+ parameter returned in the response
+* Test action 3: If there exists a public router, use it as the router. Otherwise,
+ use the public network to create a router
+* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
+* Test action 5: Create two IPv6 subnets of the network created in test action 1 in
+ ipv6_ra_mode 'slaac' and ipv6_address_mode 'slaac', storing the "id" parameters returned in the response
+* Test action 6: Connect the two IPv6 subnets to the router, using the stored IPv6 subnet ids
+* Test action 7: Boot two VMs on this network, storing the "id" parameters returned in the response
+* **Test assertion 1:** The vNIC of each VM gets one v4 address and two v6 addresses with
+ different prefixes actually assigned
+* **Test assertion 2:** Each VM can ping the other's v4 private address
+* **Test assertion 3:** The ping6 available VM can ping the other's v6 addresses
+ as well as the v6 subnets' gateway ips
+* Test action 8: Delete the 2 VMs created in test action 7, using the stored ids
+* Test action 9: List all VMs, verifying the ids are no longer present
+* **Test assertion 4:** The two "id" parameters are not present in the VM list
+* Test action 10: Delete the IPv4 subnet created in test action 2, using the stored id
+* Test action 11: Delete two IPv6 subnets created in test action 5, using the stored ids
+* Test action 12: List all subnets, verifying the ids are no longer present
+* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
+* Test action 13: Delete the network created in test action 1, using the stored id
+* Test action 14: List all networks, verifying the id is no longer present
+* **Test assertion 6:** The "id" parameter is not present in the network list
+
+Pass / Fail criteria
+'''''''''''''''''''''
+
+This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'slaac'
+and ipv6_address_mode 'slaac',
+and verify the ping6 available VM can ping the other VM's v4 address and two
+v6 addresses with different prefixes as well as the v6 subnets' gateway ips in the same network.
+Specifically it verifies that:
+
+* The different prefixes IPv6 addresses in mode 'slaac' assigned successfully
+* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnets' gateway ips
+* All items created using create commands are able to be removed using the returned identifiers
+
+Post conditions
+---------------
+
+None
+
+---------------------------------------------------------------------------------------
+Test Case 8 - IPv6 Address Assignment - Dual Net, Dual Stack, Multiple Prefixes, SLAAC
+---------------------------------------------------------------------------------------
+
+Short name
+----------
+
+functest.tempest.ipv6_scenario.dualnet_multiple_prefixes_slaac
+
+Use case specification
+----------------------
+
+This test case evaluates IPv6 address assignment in ipv6_ra_mode 'slaac' and
+ipv6_address_mode 'slaac'.
+In this case, guest instance obtains IPv6 addresses from OpenStack managed radvd
+using SLAAC. This test case then verifies the ping6 available VM can ping the other
+VM's v4 address in one network and two v6 addresses with different prefixes in another
+network as well as the v6 subnets' gateway ips, the reference is
+
+tempest.scenario.test_network_v6.TestGettingAddress.test_dualnet_multi_prefix_slaac
+
+Test preconditions
+------------------
+
+There should exist a public router or a public network.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+'''''''''''''''
+
+* Test action 1: Create one network, storing the "id" parameter returned in the response
+* Test action 2: Create one IPv4 subnet of the created network, storing the "id"
+ parameter returned in the response
+* Test action 3: If there exists a public router, use it as the router. Otherwise,
+ use the public network to create a router
+* Test action 4: Connect the IPv4 subnet to the router, using the stored IPv4 subnet id
+* Test action 5: Create another network, storing the "id" parameter returned in the response
+* Test action 6: Create two IPv6 subnets of network created in test action 5 in
+ ipv6_ra_mode 'slaac' and ipv6_address_mode 'slaac', storing the "id" parameters returned in the response
+* Test action 7: Connect the two IPv6 subnets to the router, using the stored IPv6 subnet ids
+* Test action 8: Boot two VMs on these two networks, storing the "id" parameters returned in the response
+* Test action 9: Turn on 2nd NIC of each VM for the network created in test action 5
+* **Test assertion 1:** The vNIC of each VM gets one v4 address and two v6 addresses
+ with different prefixes actually assigned
+* **Test assertion 2:** Each VM can ping the other's v4 private address
+* **Test assertion 3:** The ping6 available VM can ping the other's v6 addresses
+ as well as the v6 subnets' gateway ips
+* Test action 10: Delete the 2 VMs created in test action 8, using the stored ids
+* Test action 11: List all VMs, verifying the ids are no longer present
+* **Test assertion 4:** The two "id" parameters are not present in the VM list
+* Test action 12: Delete the IPv4 subnet created in test action 2, using the stored id
+* Test action 13: Delete two IPv6 subnets created in test action 6, using the stored ids
+* Test action 14: List all subnets, verifying the ids are no longer present
+* **Test assertion 5:** The "id" parameters of IPv4 and IPv6 are not present in the list
+* Test action 15: Delete the 2 networks created in test action 1 and 5, using the stored ids
+* Test action 16: List all networks, verifying the ids are no longer present
+* **Test assertion 6:** The two "id" parameters are not present in the network list
+
+Pass / Fail criteria
+'''''''''''''''''''''
+
+This test evaluates the ability to assign IPv6 addresses in ipv6_ra_mode 'slaac'
+and ipv6_address_mode 'slaac',
+and verify the ping6 available VM can ping the other VM's v4 address in one network and two
+v6 addresses with different prefixes in another network as well as the v6 subnets' gateway ips.
+Specifically it verifies that:
+
+* The IPv6 addresses in mode 'slaac' assigned successfully
+* The VM can ping the other VM's IPv4 and IPv6 private addresses as well as the v6 subnets' gateway ips
+* All items created using create commands are able to be removed using the returned identifiers
+
+Post conditions
+---------------
+
+None
+
+
+
diff --git a/docs/testing/user/testspecification/multiplenodes/index.rst b/docs/testing/user/testspecification/tempest_multi_node_scheduling/index.rst
index 4b4859a9..f414de61 100644
--- a/docs/testing/user/testspecification/multiplenodes/index.rst
+++ b/docs/testing/user/testspecification/tempest_multi_node_scheduling/index.rst
@@ -53,37 +53,38 @@ on multiple nodes. Each test case is able to run independently, i.e. irrelevant
the state created by a previous test. Specifically, every test performs clean-up
operations which return the system to the same state as before the test.
-All these test cases are included in the test case dovetail.tempest.tc005 of
+All these test cases are included in the test case functest.tempest.multi_node_scheduling of
OVP test suite.
Test Descriptions
=================
+----------------------
API Used and Reference
----------------------
-Security Groups: https://developer.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
+Security Groups: https://docs.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
- create security group
- delete security group
-Networks: https://developer.openstack.org/api-ref/networking/v2/index.html#networks
+Networks: https://docs.openstack.org/api-ref/network/v2/index.html#networks
- create network
- delete network
-Routers and interface: https://developer.openstack.org/api-ref/networking/v2/index.html#routers-routers
+Routers and interface: https://docs.openstack.org/api-ref/network/v2/index.html#routers-routers
- create router
- delete router
- add interface to router
-Subnets: https://developer.openstack.org/api-ref/networking/v2/index.html#subnets
+Subnets: https://docs.openstack.org/api-ref/network/v2/index.html#subnets
- create subnet
- delete subnet
-Servers: https://developer.openstack.org/api-ref/compute/
+Servers: https://docs.openstack.org/api-ref/compute/
- create keypair
- create server
@@ -95,17 +96,17 @@ Servers: https://developer.openstack.org/api-ref/compute/
- list server groups
- show server group details
-Ports: https://developer.openstack.org/api-ref/networking/v2/index.html#ports
+Ports: https://docs.openstack.org/api-ref/network/v2/index.html#ports
- create port
- delete port
-Floating IPs: https://developer.openstack.org/api-ref/networking/v2/index.html#floating-ips-floatingips
+Floating IPs: https://docs.openstack.org/api-ref/network/v2/index.html#floating-ips-floatingips
- create floating IP
- delete floating IP
-Availability zone: https://developer.openstack.org/api-ref/compute/
+Availability zone: https://docs.openstack.org/api-ref/compute/
- get availability zone
@@ -144,7 +145,7 @@ Test execution
that every server ended up on a different host
* Test action 6: Delete the created servers
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the functionality of VM resource scheduling.
@@ -189,7 +190,7 @@ Test execution
* Test action 5: List all server groups
* **Test assertion 4:** SERG1 and SERG2 are not in the list
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the functionality of creating and deleting server groups with the same name and policy.
@@ -232,7 +233,7 @@ Test execution
* Test action 3: Delete SERG1 and list all server groups
* **Test assertion 3:** SERG1 is not in the list
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the functionality of creating and deleting server group with affinity policy.
@@ -274,7 +275,7 @@ Test execution
* Test action 3: Delete SERG1 and list all server groups
* **Test assertion 3:** SERG1 is not in the list
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the functionality of creating and deleting server group with anti-affinity policy.
@@ -315,7 +316,7 @@ Test execution
* **Test assertion 1:** SERG1 is in the list
* Test action 4: Delete SERG1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the functionality of listing server groups.
@@ -357,7 +358,7 @@ Test execution
* **Test assertion 1:** All values in D1 are the same as the values in D2
* Test action 4: Delete SERG1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the functionality of showing server group details.
diff --git a/docs/testing/user/testspecification/tempest_network_api/index.rst b/docs/testing/user/testspecification/tempest_network_api/index.rst
new file mode 100644
index 00000000..4fc47208
--- /dev/null
+++ b/docs/testing/user/testspecification/tempest_network_api/index.rst
@@ -0,0 +1,130 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) Ericsson AB
+
+======================================
+Tempest Network API test specification
+======================================
+
+
+Scope
+=====
+
+The Tempest Network API test area tests the basic operations of the System Under
+Test (SUT) through the life of a VNF.
+The tests in this test area will evaluate IPv4 network runtime operations
+functionality.
+
+These runtime operations may include that create, list, verify or delete:
+
+- Floating IP
+- Network
+- Subnet
+- Port
+- External Network Visibility
+- Router
+- Subnetpools
+- API Version Resources
+
+References
+==========
+
+`Networks: <https://docs.openstack.org/api-ref/network/v2/#networks>`_
+
+- create network
+- delete network
+
+`Routers and interface: <https://docs.openstack.org/api-ref/network/v2/index.html#routers-routers>`_
+
+- create router
+- update router
+- delete router
+- add interface to router
+
+`Subnets: <https://docs.openstack.org/api-ref/network/v2/index.html#subnets>`_
+
+- create subnet
+- update subnet
+- delete subnet
+
+`Subnetpools: <https://docs.openstack.org/api-ref/network/v2/#subnet-pools-extension-subnetpools>`_
+
+- create subnetpool
+- update subnetpool
+- delete subnetpool
+
+`Ports: <https://docs.openstack.org/api-ref/network/v2/index.html#ports>`_
+
+- create port
+- update port
+- delete port
+
+`Floating IPs: <https://docs.openstack.org/api-ref/network/v2/index.html#floating-ips-floatingips>`_
+
+- create floating IP
+- delete floating IP
+
+`Api Versions <https://docs.openstack.org/api-ref/network/v2/#api-versions>`_
+
+- list version
+- show version
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVi and VIM in operation on a
+Pharos compliant infrastructure.
+
+Test Area Structure
+===================
+
+The test area is structured in individual tests as listed below.
+For detailed information on the individual steps and assertions performed
+by the tests, review the Python source code accessible via the following links:
+
+All these test cases are included in the test case functest.tempest.network of
+OVP test suite.
+
+
+`List, Show and Verify the Details of the Available Extensions <https://github.com/openstack/tempest/blob/master/tempest/api/network/test_extensions.py>`_
+ - tempest.api.network.test_extensions.ExtensionsTestJSON.test_list_show_extensions
+
+`Floating IP tests <https://github.com/openstack/tempest/blob/master/tempest/api/network/test_floating_ips.py>`_
+ - Create a Floating IP
+ - Update a Floating IP
+ - Delete a Floating IP
+ - List all Floating IPs
+ - Show Floating IP Details
+ - Associate a Floating IP with a Port and then Delete that Port
+ - Associate a Floating IP with a Port and then with a Port on Another Router
+
+- tempest.api.network.test_floating_ips.FloatingIPTestJSON.test_create_floating_ip_specifying_a_fixed_ip_address
+- tempest.api.network.test_floating_ips.FloatingIPTestJSON.test_create_list_show_update_delete_floating_ip
+
+`Network tests <https://github.com/openstack/tempest/blob/master/tempest/api/network/test_networks.py#L405>`_
+ - Bulk Network Creation & Deletion
+ - Bulk Subnet Create & Deletion
+ - Bulk Port Creation & Deletion
+ - List Project's Networks
+
+- tempest.api.network.test_networks.BulkNetworkOpsTest.test_bulk_create_delete_network
+- tempest.api.network.test_networks.BulkNetworkOpsTest.test_bulk_create_delete_port
+- tempest.api.network.test_networks.BulkNetworkOpsTest.test_bulk_create_delete_subnet
+
+`External Network Visibility test <https://github.com/openstack/tempest/blob/master/tempest/api/network/test_networks.py#L526>`_
+ - tempest.api.network.test_networks.NetworksTest.test_external_network_visibility
+
+`Create Port with No Security Groups test <https://github.com/openstack/tempest/blob/master/tempest/api/network/test_ports.py>`_
+ - tempest.api.network.test_ports.PortsTestJSON.test_create_port_with_no_securitygroups
+
+`Router test <https://github.com/openstack/tempest/blob/master/tempest/api/network/test_routers.py>`_
+ - tempest.api.network.test_routers.RoutersTest.test_add_multiple_router_interfaces
+ - tempest.api.network.test_routers.RoutersTest.test_add_remove_router_interface_with_port_id
+ - tempest.api.network.test_routers.RoutersTest.test_add_remove_router_interface_with_subnet_id
+ - tempest.api.network.test_routers.RoutersTest.test_create_show_list_update_delete_router
+
+`Create, List, Show, Update and Delete Subnetpools <https://github.com/openstack/tempest/blob/master/tempest/api/network/test_subnetpools_extensions.py>`_
+ - tempest.api.network.test_subnetpools_extensions.SubnetPoolsTestJSON.test_create_list_show_update_delete_subnetpools
+
+`API Version Resources test <https://github.com/openstack/tempest/blob/master/tempest/api/network/test_versions.py>`_
+ - tempest.api.network.test_versions.NetworksApiDiscovery.test_api_version_resources
diff --git a/docs/testing/user/testspecification/dynamicnetwork/index.rst b/docs/testing/user/testspecification/tempest_network_scenario/index.rst
index a25e4f66..feee105d 100644
--- a/docs/testing/user/testspecification/dynamicnetwork/index.rst
+++ b/docs/testing/user/testspecification/tempest_network_scenario/index.rst
@@ -2,9 +2,9 @@
.. http://creativecommons.org/licenses/by/4.0
.. (c) Huawei Technologies Co.,Ltd
-=====================================================
-Dynamic Network Runtime Operations test specification
-=====================================================
+===========================================
+Tempest Network Scenario test specification
+===========================================
.. toctree::
:maxdepth: 2
@@ -12,7 +12,7 @@ Dynamic Network Runtime Operations test specification
Scope
=====
-The dynamic network runtime operations test area evaluates the ability of the
+The Tempest Network scenario test area evaluates the ability of the
system under test to support dynamic network runtime operations through the
life of a VNF (e.g. attach/detach, enable/disable, read stats).
The tests in this test area will evaluate IPv4 network runtime operations
@@ -58,39 +58,40 @@ test case is able to run independently, i.e. irrelevant of the state created by
a previous test. Specifically, every test performs clean-up operations which
return the system to the same state as before the test.
-All these test cases are included in the test case dovetail.tempest.tc003 of
+All these test cases are included in the test case functest.tempest.network_scenario of
OVP test suite.
Test Descriptions
=================
+----------------------
API Used and Reference
----------------------
-Security Groups: https://developer.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
+Security Groups: https://docs.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
- create security group
- delete security group
-Networks: https://developer.openstack.org/api-ref/networking/v2/index.html#networks
+Networks: https://docs.openstack.org/api-ref/network/v2/index.html#networks
- create network
- delete network
-Routers and interface: https://developer.openstack.org/api-ref/networking/v2/index.html#routers-routers
+Routers and interface: https://docs.openstack.org/api-ref/network/v2/index.html#routers-routers
- create router
- update router
- delete router
- add interface to router
-Subnets: https://developer.openstack.org/api-ref/networking/v2/index.html#subnets
+Subnets: https://docs.openstack.org/api-ref/network/v2/index.html#subnets
- create subnet
- update subnet
- delete subnet
-Servers: https://developer.openstack.org/api-ref/compute/
+Servers: https://docs.openstack.org/api-ref/compute/
- create keypair
- create server
@@ -98,13 +99,13 @@ Servers: https://developer.openstack.org/api-ref/compute/
- add/assign floating IP
- disassociate floating IP
-Ports: https://developer.openstack.org/api-ref/networking/v2/index.html#ports
+Ports: https://docs.openstack.org/api-ref/network/v2/index.html#ports
- create port
- update port
- delete port
-Floating IPs: https://developer.openstack.org/api-ref/networking/v2/index.html#floating-ips-floatingips
+Floating IPs: https://docs.openstack.org/api-ref/network/v2/index.html#floating-ips-floatingips
- create floating IP
- delete floating IP
@@ -148,7 +149,7 @@ Test execution
* **Test assertion 5:** Ping FIP1 and SSH to VM2 via FIP1 successfully
* Test action 8: Delete SG1, NET1, SUBNET1, R1, VM1, VM2 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the functionality of basic network operations.
@@ -219,7 +220,7 @@ Test execution
* **Test assertion 2:** Ping NET2's internal gateway successfully
* Test action 8: Delete SG1, NET1, NET2, SUBNET1, SUBNET2, R1, NIC2, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the functionality of adding network to an active VM.
@@ -277,7 +278,7 @@ Test execution
retrieve the VM1's configured dns and verify it has been successfully updated
* Test action 7: Delete SG1, NET1, SUBNET1, R1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the functionality of updating subnet's configurations.
@@ -335,7 +336,7 @@ Test execution
* **Test assertion 6:** Ping FIP1 via SSHCLNT1 successfully
* Test action 10: Delete SG1, NET1, SUBNET1, R1, SSHCLNT1, VM1, VM2 and FIP1, FIP2
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the VM public and project connectivity status by changing VM port
@@ -394,7 +395,7 @@ Test execution
* **Test assertion 3:** Ping FIP1 and SSH to VM1 with FIP1 successfully
* Test action 8: Delete SG1, NET1, SUBNET1, R1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the router public connectivity status by changing
diff --git a/docs/testing/user/testspecification/securitygroup/index.rst b/docs/testing/user/testspecification/tempest_network_security/index.rst
index 61aa1c4b..6934be1f 100644
--- a/docs/testing/user/testspecification/securitygroup/index.rst
+++ b/docs/testing/user/testspecification/tempest_network_security/index.rst
@@ -53,21 +53,22 @@ port security. Each test case is able to run independently, i.e. irrelevant of
the state created by a previous test. Specifically, every test performs clean-up
operations which return the system to the same state as before the test.
-All these test cases are included in the test case dovetail.tempest.tc002 of
+All these test cases are included in the test case functest.tempest.network_security of
OVP test suite.
Test Descriptions
=================
+----------------------
API Used and Reference
----------------------
-Security Groups: https://developer.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
+Security Groups: https://docs.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
- create security group
- delete security group
-Networks: https://developer.openstack.org/api-ref/networking/v2/index.html#networks
+Networks: https://docs.openstack.org/api-ref/network/v2/index.html#networks
- create network
- delete network
@@ -75,27 +76,27 @@ Networks: https://developer.openstack.org/api-ref/networking/v2/index.html#netwo
- create floating ip
- delete floating ip
-Routers and interface: https://developer.openstack.org/api-ref/networking/v2/index.html#routers-routers
+Routers and interface: https://docs.openstack.org/api-ref/network/v2/index.html#routers-routers
- create router
- delete router
- list routers
- add interface to router
-Subnets: https://developer.openstack.org/api-ref/networking/v2/index.html#subnets
+Subnets: https://docs.openstack.org/api-ref/network/v2/index.html#subnets
- create subnet
- list subnets
- delete subnet
-Servers: https://developer.openstack.org/api-ref/compute/
+Servers: https://docs.openstack.org/api-ref/compute/
- create keypair
- create server
- delete server
- add/assign floating ip
-Ports: https://developer.openstack.org/api-ref/networking/v2/index.html#ports
+Ports: https://docs.openstack.org/api-ref/network/v2/index.html#ports
- update port
- list ports
@@ -144,7 +145,7 @@ Test execution
* **Test assertion 2:** The ping operation is successful
* Test action 16: Delete SG1, NET1, NET2, SUBNET1, SUBNET2, R1, VM1, VM2 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to prevent MAC spoofing by using port security.
@@ -205,7 +206,7 @@ Test execution
* Test action 14: Delete SG1, SG2, SG3, SG4, NET1, NET2, SUBNET1, SUBNET2, R1, R2,
VM1, VM2, FIP1 and FIP2
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability of the security group to filter packets cross tenant.
@@ -262,7 +263,7 @@ Test execution
* **Test assertion 2:** The ping operation is successful
* Test action 11: Delete SG1, SG2, NET1, SUBNET1, R1, VM1, VM2 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability of the security group to filter packets in one tenant.
@@ -318,7 +319,7 @@ Test execution
* **Test assertion 3:** Can SSH to VM1 successfully
* Test action 11: Delete SG1, SG2, NET1, SUBNET1, R1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability of multiple security groups to filter packets.
@@ -378,7 +379,7 @@ Test execution
* **Test assertion 2:** The ping operation is successful
* Test action 12: Delete SG1, SG2, NET1, SUBNET1, R1, VM1, VM2 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability of port security to disable security group.
@@ -435,7 +436,7 @@ Test execution
* **Test assertion 2:** The ping operation is successful
* Test action 13: Delete SG1, SG2, SG3, NET1, SUBNET1, R1, VM1, VM2 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to update port with a new security group.
diff --git a/docs/testing/user/testspecification/tempest_osinterop/index.rst b/docs/testing/user/testspecification/tempest_osinterop/index.rst
new file mode 100644
index 00000000..d2a54e86
--- /dev/null
+++ b/docs/testing/user/testspecification/tempest_osinterop/index.rst
@@ -0,0 +1,38 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) Huawei Technologies Co.,Ltd and others
+
+=============================================
+OpenStack Interoperability Test Specification
+=============================================
+
+The test cases documented here are the API test cases in the OpenStack
+Interop guideline 2018.11 as implemented by the RefStack client.
+
+References
+================
+
+- OpenStack Governance/Interop
+
+ - https://wiki.openstack.org/wiki/Governance/InteropWG
+
+- OpenStack Interoperability guidelines (version 2018.11)
+
+ - https://github.com/openstack/interop/blob/master/2018.11.json
+
+- Refstack client
+
+ - https://github.com/openstack/refstack-client
+
+
+All OpenStack interop test cases addressed in OVP are covered in the
+following test specification documents.
+
+.. toctree::
+ :maxdepth: 1
+
+ tempest_osinterop_compute.rst
+ tempest_osinterop_identity.rst
+ tempest_osinterop_image.rst
+ tempest_osinterop_network.rst
+ tempest_osinterop_volume.rst
diff --git a/docs/testing/user/testspecification/vimoperationscompute/index.rst b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_compute.rst
index 64f4356b..3e663d98 100644
--- a/docs/testing/user/testspecification/vimoperationscompute/index.rst
+++ b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_compute.rst
@@ -6,15 +6,12 @@
VIM compute operations test specification
=========================================
-.. toctree::
- :maxdepth: 2
-
Scope
=====
The VIM compute operations test area evaluates the ability of the system under
test to support VIM compute operations. The test cases documented here are the
-compute API test cases in the OpenStack Interop guideline 2016.8 as implemented
+compute API test cases in the OpenStack Interop guideline 2018.11 as implemented
by the RefStack client. These test cases will evaluate basic OpenStack (as a VIM)
compute operations, including:
@@ -25,22 +22,7 @@ compute operations, including:
- Basic server operations
- Volume management operations
-References
-================
-
-- OpenStack Governance/Interop
-
- - https://wiki.openstack.org/wiki/Governance/InteropWG
-
-- OpenStack Interoperability guidelines (version 2016.08)
-
- - https://github.com/openstack/interop/blob/master/2016.08.json
-
-- Refstack client
-
- - https://github.com/openstack/refstack-client
-
-Definitions and abbreviations
+Definitions and Abbreviations
=============================
The following terms and abbreviations are used in conjunction with this test area
@@ -48,14 +30,15 @@ The following terms and abbreviations are used in conjunction with this test are
- API - Application Programming Interface
- NFVi - Network Functions Virtualization infrastructure
- SUT - System Under Test
-- UUID - Universally Unique Identifier
+- UUID - Universally Unique IDentifier
- VIM - Virtual Infrastructure Manager
- VM - Virtual Machine
System Under Test (SUT)
=======================
-The system under test is assumed to be the NFVi and VIM deployed with a Pharos compliant infrastructure.
+The system under test is assumed to be the NFVi and VIM deployed with a Pharos
+compliant infrastructure.
Test Area Structure
====================
@@ -68,16 +51,17 @@ the same state as before the test.
For brevity, the test cases in this test area are summarized together based on
the operations they are testing.
-All these test cases are included in the test case dovetail.osinterop.tc001 of
+All these test cases are included in the test case functest.tempest.osinterop of
OVP test suite.
Test Descriptions
=================
+----------------------
API Used and Reference
----------------------
-Servers: https://developer.openstack.org/api-ref/compute/
+Servers: https://docs.openstack.org/api-ref/compute/
- create server
- delete server
@@ -96,7 +80,7 @@ Servers: https://developer.openstack.org/api-ref/compute/
- create keypair
- delete keypair
-Block storage: https://developer.openstack.org/api-ref/block-storage
+Block storage: https://docs.openstack.org/api-ref/block-storage/
- create volume
- delete volume
@@ -392,7 +376,6 @@ tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.t
tempest.api.compute.servers.test_list_server_filters.ListServerFiltersTestJSON.test_list_servers_filtered_by_name_wildcard
tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_changes_since_future_date
tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_changes_since_invalid_date
-tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_limits
tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_limits_greater_than_actual_count
tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_limits_pass_negative_value
tempest.api.compute.servers.test_list_servers_negative.ListServersNegativeTestJSON.test_list_servers_by_limits_pass_string
@@ -426,6 +409,7 @@ tempest.api.compute.servers.test_create_server.ServersTestJSON.test_verify_creat
tempest.api.compute.servers.test_create_server.ServersTestJSON.test_verify_server_details
tempest.api.compute.servers.test_create_server.ServersTestManualDisk.test_verify_created_server_vcpus
tempest.api.compute.servers.test_create_server.ServersTestManualDisk.test_verify_server_details
+tempest.api.compute.servers.test_delete_server.DeleteServersTestJSON.test_delete_active_server
Test preconditions
------------------
@@ -503,91 +487,91 @@ Test execution
* **Test assertion 31:** Verify the filtered server list is empty
* Test action 33: List all existent servers and filter the server list by a invalid change-since date format
* **Test assertion 32:** Verify a bad request error is returned in the response
-* Test action 34: List all existent servers and filter the server list by display limit '1'
-* **Test assertion 33:** Verify the length of filtered server list is 1
-* Test action 35: List all existent servers and filter the server list by a
+* Test action 34: List all existent servers and filter the server list by a
display limit value greater than the length of the server list
-* **Test assertion 34:** Verify the length of filtered server list equals to the length of server list
-* Test action 36: List all existent servers and filter the server list by display limit '-1'
+* **Test assertion 33:** Verify the length of filtered server list equals to the length of server list
+* Test action 35: List all existent servers and filter the server list by display limit '-1'
+* **Test assertion 34:** Verify a bad request error is returned in the response
+* Test action 36: List all existent servers and filter the server list by a string type limit value 'testing'
* **Test assertion 35:** Verify a bad request error is returned in the response
-* Test action 37: List all existent servers and filter the server list by a string type limit value 'testing'
-* **Test assertion 36:** Verify a bad request error is returned in the response
-* Test action 38: List all existent servers and filter the server list by a nonexistent flavor
+* Test action 37: List all existent servers and filter the server list by a nonexistent flavor
+* **Test assertion 36:** Verify the filtered server list is empty
+* Test action 38: List all existent servers and filter the server list by a nonexistent image
* **Test assertion 37:** Verify the filtered server list is empty
-* Test action 39: List all existent servers and filter the server list by a nonexistent image
+* Test action 39: List all existent servers and filter the server list by a nonexistent server name
* **Test assertion 38:** Verify the filtered server list is empty
-* Test action 40: List all existent servers and filter the server list by a nonexistent server name
-* **Test assertion 39:** Verify the filtered server list is empty
-* Test action 41: List all existent servers in detail and search the server list for a deleted server
-* **Test assertion 40:** Verify the deleted server is not in the server list
-* Test action 42: List all existent servers and filter the server list by a nonexistent server status
-* **Test assertion 41:** Verify the filtered server list is empty
-* Test action 43: List all existent servers in detail
-* **Test assertion 42:** Verify a provided deleted server's id is not in the server list
-* Test action 44: Lock a provided server VM10 and retrieve the server's status
-* **Test assertion 43:** Verify VM10 is in 'ACTIVE' status
-* Test action 45: Stop VM10
-* **Test assertion 44:** Verify stop VM10 failed
-* Test action 46: Unlock VM10 and stop VM10 again
-* **Test assertion 45:** Verify VM10 is stopped and in 'SHUTOFF' status
-* Test action 47: Start VM10
-* **Test assertion 46:** Verify VM10 is in 'ACTIVE' status
-* Test action 48: Delete metadata item 'key1' from a provided server
-* **Test assertion 47:** Verify the metadata item is removed
-* Test action 49: Get metadata item 'key2' from a provided server
-* **Test assertion 48:** Verify the metadata item is correct
-* Test action 50: List all metadata key/value pair for a provided server
-* **Test assertion 49:** Verify all metadata are retrieved correctly
-* Test action 51: Set metadata {'meta2': 'data2', 'meta3': 'data3'} for a provided server
-* **Test assertion 50:** Verify server's metadata are replaced correctly
-* Test action 52: Set metadata item nova's value to 'alt' for a provided server
-* **Test assertion 51:** Verify server's metadata are set correctly
-* Test action 53: Update metadata {'key1': 'alt1', 'key3': 'value3'} for a provided server
-* **Test assertion 52:** Verify server's metadata are updated correctly
-* Test action 54: Create a server with empty name parameter
-* **Test assertion 53:** Verify create server failed
-* Test action 55: Hard reboot a provided server
-* **Test assertion 54:** Verify server is rebooted successfully
-* Test action 56: Soft reboot a nonexistent server
-* **Test assertion 55:** Verify reboot failed, an error is returned in the response
-* Test action 57: Rebuild a provided server with new image, new server name and metadata
-* **Test assertion 56:** Verify server is rebuilt successfully, server image, name and metadata are correct
-* Test action 58: Create a server VM11
-* Test action 59: Delete VM11 and wait for VM11 to reach termination
-* Test action 60: Rebuild VM11 with another image
+* Test action 40: List all existent servers in detail and search the server list for a deleted server
+* **Test assertion 39:** Verify the deleted server is not in the server list
+* Test action 41: List all existent servers and filter the server list by a nonexistent server status
+* **Test assertion 40:** Verify the filtered server list is empty
+* Test action 42: List all existent servers in detail
+* **Test assertion 41:** Verify a provided deleted server's id is not in the server list
+* Test action 43: Lock a provided server VM10 and retrieve the server's status
+* **Test assertion 42:** Verify VM10 is in 'ACTIVE' status
+* Test action 44: Stop VM10
+* **Test assertion 43:** Verify stop VM10 failed
+* Test action 45: Unlock VM10 and stop VM10 again
+* **Test assertion 44:** Verify VM10 is stopped and in 'SHUTOFF' status
+* Test action 46: Start VM10
+* **Test assertion 45:** Verify VM10 is in 'ACTIVE' status
+* Test action 47: Delete metadata item 'key1' from a provided server
+* **Test assertion 46:** Verify the metadata item is removed
+* Test action 48: Get metadata item 'key2' from a provided server
+* **Test assertion 47:** Verify the metadata item is correct
+* Test action 49: List all metadata key/value pair for a provided server
+* **Test assertion 48:** Verify all metadata are retrieved correctly
+* Test action 50: Set metadata {'meta2': 'data2', 'meta3': 'data3'} for a provided server
+* **Test assertion 49:** Verify server's metadata are replaced correctly
+* Test action 51: Set metadata item nova's value to 'alt' for a provided server
+* **Test assertion 50:** Verify server's metadata are set correctly
+* Test action 52: Update metadata {'key1': 'alt1', 'key3': 'value3'} for a provided server
+* **Test assertion 51:** Verify server's metadata are updated correctly
+* Test action 53: Create a server with empty name parameter
+* **Test assertion 52:** Verify create server failed
+* Test action 54: Hard reboot a provided server
+* **Test assertion 53:** Verify server is rebooted successfully
+* Test action 55: Soft reboot a nonexistent server
+* **Test assertion 54:** Verify reboot failed, an error is returned in the response
+* Test action 56: Rebuild a provided server with new image, new server name and metadata
+* **Test assertion 55:** Verify server is rebuilt successfully, server image, name and metadata are correct
+* Test action 57: Create a server VM11
+* Test action 58: Delete VM11 and wait for VM11 to reach termination
+* Test action 59: Rebuild VM11 with another image
+* **Test assertion 56:** Verify rebuild server failed, an error is returned in the response
+* Test action 60: Rebuild a nonexistent server
* **Test assertion 57:** Verify rebuild server failed, an error is returned in the response
-* Test action 61: Rebuild a nonexistent server
-* **Test assertion 58:** Verify rebuild server failed, an error is returned in the response
-* Test action 62: Stop a provided server
-* **Test assertion 59:** Verify server reaches 'SHUTOFF' status
-* Test action 63: Start the stopped server
-* **Test assertion 60:** Verify server reaches 'ACTIVE' status
-* Test action 64: Stop a provided server
-* **Test assertion 61:** Verify stop server failed, an error is returned in the response
-* Test action 65: Create a server VM12 and wait it to reach 'ACTIVE' status
-* Test action 66: Update VM12's IPv4 and IPv6 access addresses
-* **Test assertion 62:** Verify VM12's access addresses have been updated correctly
-* Test action 67: Create a server VM13 and wait it to reach 'ACTIVE' status
-* Test action 68: Update VM13's server name with non-ASCII characters '\u00CD\u00F1st\u00E1\u00F1c\u00E9'
-* **Test assertion 63:** Verify VM13's server name has been updated correctly
-* Test action 69: Update the server name of a nonexistent server
-* **Test assertion 64:** Verify update server name failed, an 'object not found' error is returned in the response
-* Test action 70: Update a provided server's name with a 256-character long name
-* **Test assertion 65:** Verify update server name failed, a bad request is returned in the response
-* Test action 71: Update a provided server's server name with an empty string
-* **Test assertion 66:** Verify update server name failed, a bad request error is returned in the response
-* Test action 72: Get the number of vcpus of a provided server
-* Test action 73: Get the number of vcpus stated by the server's flavor
-* **Test assertion 67:** Verify that the number of vcpus reported by the server
+* Test action 61: Stop a provided server
+* **Test assertion 58:** Verify server reaches 'SHUTOFF' status
+* Test action 62: Start the stopped server
+* **Test assertion 59:** Verify server reaches 'ACTIVE' status
+* Test action 63: Stop a provided server
+* **Test assertion 60:** Verify stop server failed, an error is returned in the response
+* Test action 64: Create a server VM12 and wait it to reach 'ACTIVE' status
+* Test action 65: Update VM12's IPv4 and IPv6 access addresses
+* **Test assertion 61:** Verify VM12's access addresses have been updated correctly
+* Test action 66: Create a server VM13 and wait it to reach 'ACTIVE' status
+* Test action 67: Update VM13's server name with non-ASCII characters '\u00CD\u00F1st\u00E1\u00F1c\u00E9'
+* **Test assertion 62:** Verify VM13's server name has been updated correctly
+* Test action 68: Update the server name of a nonexistent server
+* **Test assertion 63:** Verify update server name failed, an 'object not found' error is returned in the response
+* Test action 69: Update a provided server's name with a 256-character long name
+* **Test assertion 64:** Verify update server name failed, a bad request is returned in the response
+* Test action 70: Update a provided server's server name with an empty string
+* **Test assertion 65:** Verify update server name failed, a bad request error is returned in the response
+* Test action 71: Get the number of vcpus of a provided server
+* Test action 72: Get the number of vcpus stated by the server's flavor
+* **Test assertion 66:** Verify that the number of vcpus reported by the server
matches the amount stated by the server's flavor
-* Test action 74: Create a server VM14
-* **Test assertion 68:** Verify VM14's server attributes are set correctly
-* Test action 75: Get the number of vcpus of a provided server (manual disk configuration)
-* Test action 76: Get the number of vcpus stated by the server's flavor (manual disk configuration)
-* **Test assertion 69:** Verify that the number of vcpus reported by the server
+* Test action 73: Create a server VM14
+* **Test assertion 67:** Verify VM14's server attributes are set correctly
+* Test action 74: Get the number of vcpus of a provided server (manual disk configuration)
+* Test action 75: Get the number of vcpus stated by the server's flavor (manual disk configuration)
+* **Test assertion 68:** Verify that the number of vcpus reported by the server
matches the amount stated by the server's flavor (manual disk configuration)
-* Test action 77: Create a server VM15 (manual disk configuration)
-* **Test assertion 70:** Verify VM15's server attributes are set correctly (manual disk configuration)
+* Test action 76: Create a server VM15 (manual disk configuration)
+* **Test assertion 69:** Verify VM15's server attributes are set correctly (manual disk configuration)
+* Test action 77: Create a server VM16 and then delete it when its status is 'ACTIVE'
+* **Test assertion 70:** Verify VM16 is deleted successfully
* Test action 78: Delete all VMs created
Pass / fail criteria
@@ -602,6 +586,7 @@ Specifically, the test verifies that:
* Create a server with a metadata that exceeds the length limit is not allowed
* Create a server with an invalid flavor, an invalid image or an invalid network UUID is not allowed
* Delete a server with a server ID that exceeds the length limit or a nonexistent server ID is not allowed
+* Delete a server which status is 'ACTIVE' is allowed
* A provided server's host name is the same as the server name
* Create a server with an invalid IPv6 access address is not allowed
* A created server is in the (detailed) list of servers
@@ -691,3 +676,132 @@ Post conditions
---------------
N/A
+
+--------------------------------------------------------------------------
+Test Case 8 - List Compute service availability zones with the Compute API
+--------------------------------------------------------------------------
+
+Test case specification
+-----------------------
+
+This test case evaluates the Compute API ability of listing availability zones
+with a non admin user, the reference is,
+
+tempest.api.compute.servers.test_availability_zone.AZV2TestJSON.test_get_availability_zone_list_with_non_admin_user
+
+Test preconditions
+------------------
+
+* Compute volume extension API
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+''''''''''''''
+
+* Test action 1: List availability zones with a non admin user
+* **Test assertion 1:** The list is not empty
+
+Pass / fail criteria
+''''''''''''''''''''
+
+This test evaluates the functionality of listing availability zones with a non admin user.
+Specifically, the test verifies that:
+
+* Non admin users can list availability zones.
+
+In order to pass this test, all test assertions listed in the test execution above need to pass.
+
+Post conditions
+---------------
+
+N/A
+
+-------------------------------------------------
+Test Case 9 - List Flavors within the Compute API
+-------------------------------------------------
+
+Test case specification
+-----------------------
+
+This test case evaluates the Compute API ability of listing flavors, the reference is,
+
+tempest.api.compute.flavors.test_flavors.FlavorsV2TestJSON.test_list_flavors
+tempest.api.compute.flavors.test_flavors.FlavorsV2TestJSON.test_list_flavors_with_detail
+
+Test preconditions
+------------------
+
+* Compute volume extension API
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+''''''''''''''
+
+* Test action 1: List all flavors
+* **Test assertion 1:** One given flavor is list in the all flavors' list
+* Test action 2: List all flavors with details
+* **Test assertion 2:** One given flavor is list in the all flavors' list
+
+Pass / fail criteria
+''''''''''''''''''''
+
+This test evaluates the functionality of listing flavors within the Compute API.
+Specifically, the test verifies that:
+
+* Can list flavors with/without details within the Compute API.
+
+In order to pass this test, all test assertions listed in the test execution above need to pass.
+
+Post conditions
+---------------
+
+N/A
+
+--------------------------------------------------------
+Test Case 10 - Keypair operations within the Compute API
+--------------------------------------------------------
+
+Test case specification
+-----------------------
+
+This test case evaluates the Compute API ability of creating keypair with type,
+the reference is,
+
+tempest.api.compute.keypairs.test_keypairs_v22.KeyPairsV22TestJSON.test_keypairsv22_create_list_show_with_type
+
+Test preconditions
+------------------
+
+* Compute server extension API
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+''''''''''''''
+
+* Test action 1: Create a keypair with type 'x509' and a random name
+* **Test assertion 1:** The keypair type received in the response body is equal to 'x509'
+* Test action 2: Show the details of this created keypair
+* **Test assertion 2:** The keypair type received in the response body is equal to 'x509'
+* Test action 3: List all keypairs and find the one with the same name as given in test action 1
+* **Test assertion 3:** The keypair type of this keypair is equal to 'x509'
+
+Pass / fail criteria
+''''''''''''''''''''
+
+This test evaluates the functionality of keypair operations within the Compute API.
+Specifically, the test verifies that:
+
+* Can create keypair by specifying keypair type.
+
+In order to pass this test, all test assertions listed in the test execution above need to pass.
+
+Post conditions
+---------------
+
+N/A
diff --git a/docs/testing/user/testspecification/vimoperationsidentity/index.rst b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_identity.rst
index 9790e75e..34a71168 100644
--- a/docs/testing/user/testspecification/vimoperationsidentity/index.rst
+++ b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_identity.rst
@@ -6,9 +6,6 @@
VIM identity operations test specification
==========================================
-.. toctree::
- :maxdepth: 2
-
Scope
=====
@@ -17,23 +14,7 @@ support VIM identity operations. The tests in this area will evaluate
API discovery operations within the Identity v3 API, auth operations within
the Identity API.
-References
-================
-
-- OpenStack interoperability guidelines (version 2016.08)
-
- - https://github.com/openstack/interop/blob/master/2016.08.json
-
-- Openstack interoperability
-
- - https://www.openstack.org/brand/interop/
-
-- OpenStack interoperability test cases excluding object storage
-
- - https://refstack.openstack.org/api/v1/guidelines/2016.08/tests?target=compute&type=required&alias=true&flag=false
-
-
-Definitions and abbreviations
+Definitions and Abbreviations
=============================
The following terms and abbreviations are used in conjunction with this test area
@@ -53,23 +34,23 @@ Test Area Structure
The test area is structured based on VIM identity operations. Each test case
is able to run independently, i.e. irrelevant of the state created by a previous test.
-All these test cases are included in the test case dovetail.osinterop.tc001 of
+All these test cases are included in the test case functest.tempest.osinterop of
OVP test suite.
Dependency Description
======================
The VIM identity operations test cases are a part of the OpenStack
-interoperability tempest test cases. For Danube based dovetail release, the
-OpenStack interoperability guidelines (version 2016.08) is adopted, which is
-valid for Kilo, Liberty, Mitaka and Newton releases of Openstack.
+interoperability tempest test cases. For Fraser based dovetail release, the
+OpenStack interoperability guidelines (version 2018.11) is adopted, which is
+valid for Mitaka, Newton, Ocata and Pike releases of Openstack.
Test Descriptions
=================
-----------------------------------------------------
-API discovery operations within the Identity v3 API
-----------------------------------------------------
+-----------------------------------------------------------------
+Test Case 1 - API discovery operations within the Identity v3 API
+-----------------------------------------------------------------
Use case specification
-----------------------
@@ -98,7 +79,7 @@ Test execution
'stable', 'experimental', 'supported', 'deprecated' are all of the identity api 'status'
values.
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test case passes if all test action steps execute successfully and all assertions
@@ -110,14 +91,15 @@ Post conditions
None
-------------------------------------------
-Auth operations within the Identity API
-------------------------------------------
+-----------------------------------------------------
+Test Case 2 - Auth operations within the Identity API
+-----------------------------------------------------
Use case specification
-----------------------
tempest.api.identity.v3.test_tokens.TokensV3Test.test_create_token
+tempest.api.identity.v3.test_tokens.TokensV3Test.test_validate_token
Test preconditions
-------------------
@@ -138,8 +120,16 @@ Test execution
passes if it is equal to the user_name which is used to get token.
* Test action 4: Get the method in getting token response message, the test
passes if it is equal to the password which is used to get token.
+* Test action 5: Get the token by system credentials and show the token,
+ the test passes if the response bodies of the get and show operations are the same.
+* Test action 6: Get the user_id in showing token response message, the test
+ passes if it is equal to the user_id which is used to get token.
+* Test action 7: Get the username in showing token response message, the test
+ passes if it is equal to the username which is used to get token.
+* Test action 8: Delete this token by non-admin compute client, the test passes
+ if it raises a NotFound exception.
-Pass / fail criteria
+Pass / Fail criteria
'''''''''''''''''''''
This test case passes if all test action steps execute successfully and all assertions
@@ -151,3 +141,39 @@ Post conditions
None
+--------------------------------------------------------
+Test Case 3 - Catalog operations within the Identity API
+--------------------------------------------------------
+
+Use case specification
+-----------------------
+
+tempest.api.identity.v3.test_catalog.IdentityCatalogTest.test_catalog_standardization
+
+Test preconditions
+-------------------
+
+None
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+'''''''''''''''
+
+* Test action 1: Show all catalogs by non-admin catalog client, the test passes
+ if the catalog types getting in the show response message equal to the
+ standard service values. Standard catalog types of 'keystone', 'nova', 'glance' and
+ 'swift' should be 'identity', 'compute', 'image' and 'object-store' respectively.
+
+Pass / Fail criteria
+'''''''''''''''''''''
+
+This test case passes if all test action steps execute successfully and all assertions
+are affirmed. If any test steps fails to execute successfully or any of the assertions
+is not met, the test case fails.
+
+Post conditions
+---------------
+
+None
diff --git a/docs/testing/user/testspecification/vimoperationsimage/index.rst b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_image.rst
index 2970207e..b6cdb77f 100644
--- a/docs/testing/user/testspecification/vimoperationsimage/index.rst
+++ b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_image.rst
@@ -6,30 +6,16 @@
VIM image operations test specification
=======================================
-.. toctree::
- :maxdepth: 2
-
Scope
=====
The VIM image test area evaluates the ability of the system under test to support
VIM image operations. The test cases documented here are the Image API test cases
-in the Openstack Interop guideline 2016.8 as implemented by the Refstack client.
+in the Openstack Interop guideline 2018.11 as implemented by the Refstack client.
These test cases will evaluate basic Openstack (as a VIM) image operations including
image creation, image list, image update and image deletion capabilities using Glance v2 API.
-References
-==========
-
-- OpenStack Interoperability guidelines (version 2016.08)
-
- - https://github.com/openstack/interop/blob/master/2016.08.json
-
-- Refstack client
-
- - https://github.com/openstack/refstack-client
-
-Definitions and abbreviations
+Definitions and Abbreviations
=============================
The following terms and abbreviations are used in conjunction with this test area
@@ -54,16 +40,17 @@ to run independently, i.e. irrelevant of the state created by a previous test.
For brevity, the test cases in this test area are summarized together based on
the operations they are testing.
-All these test cases are included in the test case dovetail.osinterop.tc001 of
+All these test cases are included in the test case functest.tempest.osinterop of
OVP test suite.
Test Descriptions
=================
+----------------------
API Used and Reference
----------------------
-Images: https://developer.openstack.org/api-ref/image/v2/
+Images: https://docs.openstack.org/api-ref/image/v2/
- create image
- delete image
@@ -75,9 +62,9 @@ Images: https://developer.openstack.org/api-ref/image/v2/
- add image tag
- delete image tag
----------------------------------------
-Image get tests using the Glance v2 API
----------------------------------------
+-----------------------------------------------------
+Test Case 1 - Image get tests using the Glance v2 API
+-----------------------------------------------------
Test case specification
-----------------------
@@ -117,7 +104,7 @@ Test execution
whether the 6 images' ids are not in the show list.
* **Test assertion 6:** The 6 images' ids are not found in the show list.
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
The first two test cases evaluate the ability to use Glance v2 API to show image
@@ -137,9 +124,9 @@ Post conditions
None
---------------------------------------
-CRUD image operations in Images API v2
---------------------------------------
+----------------------------------------------------
+Test Case 2 - CRUD image operations in Images API v2
+----------------------------------------------------
Test case specification
-----------------------
@@ -161,7 +148,7 @@ Test execution
* Test action 2: List all images and check whether the ids listed are in the created images list.
* **Test assertion 1:** The ids get from the list images API are in the created images list.
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the ability to use Glance v2 API to list images.
@@ -176,9 +163,9 @@ Post conditions
None
-----------------------------------------
-Image list tests using the Glance v2 API
-----------------------------------------
+------------------------------------------------------
+Test Case 3 - Image list tests using the Glance v2 API
+------------------------------------------------------
Test case specification
-----------------------
@@ -237,7 +224,7 @@ Test execution
the 6 ids are not in the show list.
* **Test assertion 8:** The stored 6 ids are not found in the show list.
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the ability to use Glance v2 API to list images with
@@ -258,9 +245,9 @@ Post conditions
None
-------------------------------------------
-Image update tests using the Glance v2 API
-------------------------------------------
+--------------------------------------------------------
+Test Case 4 - Image update tests using the Glance v2 API
+--------------------------------------------------------
Test case specification
-----------------------
@@ -304,7 +291,7 @@ Test execution
and check whether the ids are not in the show list.
* **Test assertion 6:** The two ids are not found in the show list.
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the ability to use Glance v2 API to update images with
@@ -321,9 +308,9 @@ Post conditions
None
---------------------------------------------
-Image deletion tests using the Glance v2 API
---------------------------------------------
+----------------------------------------------------------
+Test Case 5 - Image deletion tests using the Glance v2 API
+----------------------------------------------------------
Test case specification
-----------------------
@@ -362,7 +349,7 @@ Test execution
and check whether the ids are in the list.
* **Test assertion 5:** The two ids are not found in the list.
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
The first three test cases evaluate the ability to use Glance v2 API to delete images
diff --git a/docs/testing/user/testspecification/vimoperationsnetwork/index.rst b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_network.rst
index bf929828..d8e7413b 100644
--- a/docs/testing/user/testspecification/vimoperationsnetwork/index.rst
+++ b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_network.rst
@@ -6,30 +6,16 @@
VIM network operations test specification
=========================================
-.. toctree::
- :maxdepth: 2
-
Scope
=====
The VIM network test area evaluates the ability of the system under test to support
VIM network operations. The test cases documented here are the network API test cases
-in the Openstack Interop guideline 2016.8 as implemented by the Refstack client.
+in the Openstack Interop guideline 2018.11 as implemented by the Refstack client.
These test cases will evaluate basic Openstack (as a VIM) network operations including
basic CRUD operations on L2 networks, L2 network ports and security groups.
-References
-==========
-
-- OpenStack Interoperability guidelines (version 2016.08)
-
- - https://github.com/openstack/interop/blob/master/2016.08.json
-
-- Refstack client
-
- - https://github.com/openstack/refstack-client
-
-Definitions and abbreviations
+Definitions and Abbreviations
=============================
The following terms and abbreviations are used in conjunction with this test area
@@ -56,16 +42,17 @@ the same state as before the test.
For brevity, the test cases in this test area are summarized together based on
the operations they are testing.
-All these test cases are included in the test case dovetail.osinterop.tc001 of
+All these test cases are included in the test case functest.tempest.osinterop of
OVP test suite.
Test Descriptions
=================
+----------------------
API Used and Reference
----------------------
-Network: http://developer.openstack.org/api-ref/networking/v2/index.html
+Network: https://docs.openstack.org/api-ref/network/v2/index.html
- create network
- update network
@@ -97,9 +84,9 @@ Network: http://developer.openstack.org/api-ref/networking/v2/index.html
- show security group rule
- delete security group rule
----------------------------------------------------------
-Basic CRUD operations on L2 networks and L2 network ports
----------------------------------------------------------
+-----------------------------------------------------------------------
+Test Case 1 - Basic CRUD operations on L2 networks and L2 network ports
+-----------------------------------------------------------------------
Test case specification
-----------------------
@@ -113,6 +100,7 @@ tempest.api.network.test_networks.NetworksTest.test_create_delete_subnet_without
tempest.api.network.test_networks.NetworksTest.test_create_delete_subnet_all_attributes
tempest.api.network.test_networks.NetworksTest.test_create_update_delete_network_subnet
tempest.api.network.test_networks.NetworksTest.test_delete_network_with_subnet
+tempest.api.network.test_networks.NetworksTest.test_external_network_visibility
tempest.api.network.test_networks.NetworksTest.test_list_networks
tempest.api.network.test_networks.NetworksTest.test_list_networks_fields
tempest.api.network.test_networks.NetworksTest.test_list_subnets
@@ -127,10 +115,9 @@ tempest.api.network.test_ports.PortsTestJSON.test_create_port_in_allowed_allocat
tempest.api.network.test_ports.PortsTestJSON.test_create_update_delete_port
tempest.api.network.test_ports.PortsTestJSON.test_list_ports
tempest.api.network.test_ports.PortsTestJSON.test_list_ports_fields
+tempest.api.network.test_ports.PortsTestJSON.test_port_list_filter_by_router_id
tempest.api.network.test_ports.PortsTestJSON.test_show_port
tempest.api.network.test_ports.PortsTestJSON.test_show_port_fields
-tempest.api.network.test_ports.PortsTestJSON.test_update_port_with_security_group_and_extra_attributes
-tempest.api.network.test_ports.PortsTestJSON.test_update_port_with_two_security_groups_and_extra_attributes
Test preconditions
------------------
@@ -178,65 +165,64 @@ Test execution
found after deletion
* Test action 9: Create a network and create a subnet of this network, then delete this network
* **Test assertion 9:** The subnet has also been deleted after deleting the network
-* Test action 10: Create a network and list all networks
-* **Test assertion 10:** The network created is found in the list
-* Test action 11: Create a network and list networks with the id and name of the created network
-* **Test assertion 11:** The id and name of the list network equal to the created network's id and name
-* Test action 12: Create a network and create a subnet of this network, then list all subnets
-* **Test assertion 12:** The subnet created is found in the list
-* Test action 13: Create a network and create a subnet of this network, then list subnets with
+* Test action 10: List all external networks, find the one with the same public_network_id
+ as defined in tempest.conf and list its subnets
+* **Test assertion 10:** The external network can be found, no internal network got by the
+ list operation, if this external network is shared, the subnets list is not empty,
+ otherwise, it should be empty
+* Test action 11: Create a network and list all networks
+* **Test assertion 11:** The network created is found in the list
+* Test action 12: Create a network and list networks with the id and name of the created network
+* **Test assertion 12:** The id and name of the list network equal to the created network's id and name
+* Test action 13: Create a network and create a subnet of this network, then list all subnets
+* **Test assertion 13:** The subnet created is found in the list
+* Test action 14: Create a network and create a subnet of this network, then list subnets with
the id and network_id of the created subnet
-* **Test assertion 13:** The id and network_id of the list subnet equal to the created subnet
-* Test action 14: Create a network and show network's details with the id of the created network
-* **Test assertion 14:** The id and name returned in the response equal to the created network's id and name
-* Test action 15: Create a network and just show network's id and name info with the id of the created network
-* **Test assertion 15:** The keys returned in the response are only id and name, and the values
+* **Test assertion 14:** The id and network_id of the list subnet equal to the created subnet
+* Test action 15: Create a network and show network's details with the id of the created network
+* **Test assertion 15:** The id and name returned in the response equal to the created network's id and name
+* Test action 16: Create a network and just show network's id and name info with the id of the created network
+* **Test assertion 16:** The keys returned in the response are only id and name, and the values
of all the keys equal to network's id and name
-* Test action 16: Create a network and create a subnet of this network, then show subnet's details
+* Test action 17: Create a network and create a subnet of this network, then show subnet's details
with the id of the created subnet
-* **Test assertion 16:** The id and cidr info returned in the response equal to the created
+* **Test assertion 17:** The id and cidr info returned in the response equal to the created
subnet's id and cidr
-* Test action 17: Create a network and create a subnet of this network, then show subnet's id and
+* Test action 18: Create a network and create a subnet of this network, then show subnet's id and
network_id info with the id of the created subnet
-* **Test assertion 17:** The keys returned in the response are just id and network_id, and the values
+* **Test assertion 18:** The keys returned in the response are just id and network_id, and the values
of all the keys equal to subnet's id and network_id
-* Test action 18: Create a network and create a subnet of this network, then update subnet's
+* Test action 19: Create a network and create a subnet of this network, then update subnet's
name, host_routes, dns_nameservers and gateway_ip
-* **Test assertion 18:** The name, host_routes, dns_nameservers and gateway_ip returned in the
+* **Test assertion 19:** The name, host_routes, dns_nameservers and gateway_ip returned in the
response equal to the values used to update the subnet
-* Test action 19: Create 2 networks and bulk create 2 ports with the ids of the created networks
-* **Test assertion 19:** The network_id of each port equals to the one used to create the port and
+* Test action 20: Create 2 networks and bulk create 2 ports with the ids of the created networks
+* **Test assertion 20:** The network_id of each port equals to the one used to create the port and
the admin_state_up of each port is True
-* Test action 20: Create a network and create a subnet of this network by setting allocation_pools,
+* Test action 21: Create a network and create a subnet of this network by setting allocation_pools,
then create a port with the created network's id
-* **Test assertion 20:** The ip_address of the created port is in the range of the allocation_pools
-* Test action 21: Create a network and create a port with its id, then update the port's name and
+* **Test assertion 21:** The ip_address of the created port is in the range of the allocation_pools
+* Test action 22: Create a network and create a port with its id, then update the port's name and
set its admin_state_up to be False
-* **Test assertion 21:** The name returned in the response equals to the name used to update
+* **Test assertion 22:** The name returned in the response equals to the name used to update
the port and the port's admin_state_up is False
-* Test action 22: Create a network and create a port with its id, then list all ports
-* **Test assertion 22:** The created port is found in the list
-* Test action 23: Create a network and create a port with its id, then list ports with the id
- and mac_address of the created port
+* Test action 23: Create a network and create a port with its id, then list all ports
* **Test assertion 23:** The created port is found in the list
-* Test action 24: Create a network and create a port with its id, then show the port's details
-* **Test assertion 24:** The key 'id' is in the details
-* Test action 25: Create a network and create a port with its id, then show the port's id
+* Test action 24: Create a network and create a port with its id, then list ports with the id
+ and mac_address of the created port
+* **Test assertion 24:** The created port is found in the list
+* Test action 25: Create a network and create a subnet, port with its id, create a router
+ and add this port as this router's interface, then list ports with this router id
+* **Test assertion 25:** The number of the ports list is 1, the port id and device_id
+ getting with the list operation are the same as the ones got when creating them
+* Test action 26: Create a network and create a port with its id, then show the port's details
+* **Test assertion 26:** The key 'id' is in the details
+* Test action 27: Create a network and create a port with its id, then show the port's id
and mac_address info with the port's id
-* **Test assertion 25:** The keys returned in the response are just id and mac_address,
+* **Test assertion 27:** The keys returned in the response are just id and mac_address,
and the values of all the keys equal to port's id and mac_address
-* Test action 26: Create a network, 2 subnets (SUBNET1 and SUBNET2) and 2 security groups
- (SG1 and SG2), create a port with SG1 and SUBNET1, then update the port's security group to SG2
- and its subnet_id to SUBNET2
-* **Test assertion 26:** The port's subnet_id equals to SUBNET2's id and its security_group_ids
- equals to SG2's id
-* Test action 27: Create a network, 2 subnets (SUBNET1 and SUBNET2) and 3 security groups
- (SG1, SG2 and SG3), create a port with SG1 and SUBNET1, then update the port's security group to
- SG2 and SG3 and its subnet_id to SUBNET2
-* **Test assertion 27:** The port's subnet_id equal to SUBNET2's id and its security_group_ids
- equals to the ids of SG2 and SG3
-
-Pass / fail criteria
+
+Pass / Fail criteria
''''''''''''''''''''
These test cases evaluate the ability of basic CRUD operations on L2 networks and L2 network ports.
@@ -257,9 +243,9 @@ Post conditions
N/A
-----------------------------------------
-Basic CRUD operations on security groups
-----------------------------------------
+------------------------------------------------------
+Test Case 2 - Basic CRUD operations on security groups
+------------------------------------------------------
Test case specification
-----------------------
@@ -346,7 +332,7 @@ Test execution
* Test action 19: Generate a random uuid and use this id to show security group rule
* **Test assertion 19:** Failed to show security group rule because of nonexistent id of security group rule
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
These test cases evaluate the ability of Basic CRUD operations on security groups and security group rules.
@@ -366,3 +352,175 @@ Post conditions
---------------
N/A
+
+---------------------------------------------
+Test Case 3 - CRUD operations on subnet pools
+---------------------------------------------
+
+Test case specification
+-----------------------
+
+tempest.api.network.test_subnetpools_extensions.SubnetPoolsTestJSON.test_create_list_show_update_delete_subnetpools
+
+Test preconditions
+------------------
+
+Neutron is available.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+''''''''''''''
+
+* Test action 1: Create a subnetpool SNP1 with a specific name and get the name from the response body
+* **Test assertion 1:** The name got from the body is the same as the name used to create SNP1
+* Test action 2: Show SNP1 and get the name from the response body
+* **Test assertion 2:** The name got from the body is the same as the name used to create SNP1
+* Test action 3: Update the name of SNP1 and get the new name from the response body
+* **Test assertion 3:** The name got from the body is the same as the name used to update SNP1
+* Test action 4: Delete SNP1
+
+
+Pass / Fail criteria
+''''''''''''''''''''
+
+These test cases evaluate the ability of Basic CRUD operations on subnetpools.
+Specifically it verifies that:
+
+* Subnetpools can be created, updated, shown and deleted.
+
+In order to pass this test, all test assertions listed in the test execution above need to pass.
+
+Post conditions
+---------------
+
+N/A
+
+----------------------------------------
+Test Case 4 - CRUD operations on routers
+----------------------------------------
+
+Test case specification
+-----------------------
+
+tempest.api.network.test_routers.RoutersTest.test_add_multiple_router_interfaces
+tempest.api.network.test_routers.RoutersTest.test_add_remove_router_interface_with_port_id
+tempest.api.network.test_routers.RoutersTest.test_add_remove_router_interface_with_subnet_id
+tempest.api.network.test_routers.RoutersTest.test_create_show_list_update_delete_router
+tempest.api.network.test_routers.RoutersTest.test_update_delete_extra_route
+tempest.api.network.test_routers.RoutersTest.test_update_router_admin_state
+tempest.api.network.test_routers_negative.RoutersNegativeTest.test_add_router_interfaces_on_overlapping_subnets_returns_400
+tempest.api.network.test_routers_negative.RoutersNegativeTest.test_delete_non_existent_router_returns_404
+tempest.api.network.test_routers_negative.RoutersNegativeTest.test_router_add_gateway_invalid_network_returns_404
+tempest.api.network.test_routers_negative.RoutersNegativeTest.test_router_add_gateway_net_not_external_returns_400
+tempest.api.network.test_routers_negative.RoutersNegativeTest.test_router_remove_interface_in_use_returns_409
+tempest.api.network.test_routers_negative.RoutersNegativeTest.test_show_non_existent_router_returns_404
+tempest.api.network.test_routers_negative.RoutersNegativeTest.test_update_non_existent_router_returns_404
+
+Test preconditions
+------------------
+
+Neutron is available.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+''''''''''''''
+
+* Test action 1: Create 2 networks NET1 and NET2, create SubNet1 of NET1 and SubNet2 with cidr of NET2,
+ create a router R1, add SubNet1 and SubNet2 to be R1's interfaces and get port details
+* **Test assertion 1:** The device_id and subnet_id of the port equals to the router id
+ and subnet id getting from the creating response body
+* Test action 2: Create a network NET1, create SubNet1 of NET1, create P1 of NET1, create a router R1,
+ add P1 to be its interface, store the response body B1, show P1 and store the response body B2 and
+ remove the interface with port id
+* **Test assertion 2:** B1 includes keys 'subnet_id' and 'port_id', 'device_id' of B2 equals to router id
+* Test action 3: Create a network NET1, create SubNet1 of NET1, create a router R1, add SubNet1 to be its
+ interface, store the response body B1, show port details with 'port_id' in B1
+* **Test assertion 3:** B1 includes keys 'subnet_id' and 'port_id', 'device_id' equals to router id
+* Test action 4: Create a router R1 with name, admin_state_up False and external_network_id Ext-Net,
+ store the request body B1, show R1 with 'id' in B1, list all routers, update R1's name
+* **Test assertion 4:** 'name', 'admin_state_up' and 'network_id' in B1 equal to the name, False and Ext-Net,
+ 'name' in show details equals to 'name' in B1, 'network_id' in show details equals to Ext-Net,
+ R1 is in the router list, the 'name' has been updated
+* Test action 5: Create a router R1 with admin_state_up=True, create 4 networks and 4 subnets with different
+ cidr, add these 4 subnets to R1 as its interfaces, update R1 by given routes and show R1, delete extra
+ route of R1 and then show R1
+* **Test assertion 5:** The number of routes queals to 4 and routes is empty after deletion
+* Test action 6: Create a router R1, update 'admin_state_up' of R1 to be True
+* **Test assertion 6:** 'admin_state_up' of R1 is False, 'admin_state_up' of R1 is True after updating
+* Test action 7: Create 2 networks NET1 and NET2, create SubNet1 of NET1 and SubNet2 of NET2,
+ create a router R1, add SubNet1 to be R1's interface and then trying to add SubNet2 to be R1's interface
+* **Test assertion 7:** It raises a BadRequest exception when trying to add SubNet2
+* Test action 8: Try to delete router with a random name
+* **Test assertion 8:** It raises a NotFound exception when trying delete operation
+* Test action 9: Create a router R1 and try to update it with wrong external_gateway_info
+* **Test assertion 9:** It raises a NotFound exception when trying to update R1
+* Test action 10: Create an internal network NET1, create SubNet1 with cidr of NET1,
+ create a router R1 and try to update it with NET1 as external_gateway_info
+* **Test assertion 10:** It raises a BadRequest exception when trying to update R1
+* Test action 11: Create a network NET1, create SubNet1 of NET1, create a router R1, add SubNet1 to
+ the interface of R1 and try to delete R1
+* **Test assertion 11:** It raises a Conflict exception when trying to delete R1
+* Test action 12: Try to show router with a random name
+* **Test assertion 12:** It raises a NotFound exception when trying to show router
+* Test action 13: Try to update router with a random name
+* **Test assertion 13:** It raises a NotFound exception when trying to update router
+
+
+Pass / Fail criteria
+''''''''''''''''''''
+
+These test cases evaluate the ability of Basic CRUD operations on routers.
+Specifically it verifies that:
+
+* Routers can be created, updated, shown and deleted.
+* Can not show, update and delete non existent router.
+* Can not remove interface in use.
+
+In order to pass this test, all test assertions listed in the test execution above need to pass.
+
+Post conditions
+---------------
+
+N/A
+
+----------------------------------------------
+Test Case 5 - List versions within Network API
+----------------------------------------------
+
+Test case specification
+-----------------------
+
+tempest.api.network.test_versions.NetworksApiDiscovery.test_api_version_resources
+
+Test preconditions
+------------------
+
+Neutron is available.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+''''''''''''''
+
+* Test action 1: List network API versions
+* **Test assertion 1:** The network API version is 'v2.0'
+
+Pass / Fail criteria
+''''''''''''''''''''
+
+This test case evaluates the ability of listing network API versions.
+Specifically it verifies that:
+
+* The network API version is 'v2.0'.
+
+In order to pass this test, all test assertions listed in the test execution above need to pass.
+
+Post conditions
+---------------
+
+N/A
diff --git a/docs/testing/user/testspecification/vimoperationsvolume/index.rst b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_volume.rst
index ce51b954..263e87cc 100644
--- a/docs/testing/user/testspecification/vimoperationsvolume/index.rst
+++ b/docs/testing/user/testspecification/tempest_osinterop/tempest_osinterop_volume.rst
@@ -6,15 +6,12 @@
VIM volume operations test specification
=========================================
-.. toctree::
- :maxdepth: 2
-
Scope
=====
The VIM volume operations test area evaluates the ability of the system under
test to support VIM volume operations. The test cases documented here are the
-volume API test cases in the OpenStack Interop guideline 2016.8 as implemented
+volume API test cases in the OpenStack Interop guideline 2018.11 as implemented
by the RefStack client. These test cases will evaluate basic OpenStack (as a VIM)
volume operations, including:
@@ -27,22 +24,7 @@ volume operations, including:
- Volume metadata operations
- Volume snapshot operations
-References
-================
-
-- OpenStack Governance/Interop
-
- - https://wiki.openstack.org/wiki/Governance/InteropWG
-
-- OpenStack Interoperability guidelines (version 2016.08)
-
- - https://github.com/openstack/interop/blob/master/2016.08.json
-
-- Refstack client
-
- - https://github.com/openstack/refstack-client
-
-Definitions and abbreviations
+Definitions and Abbreviations
=============================
The following terms and abbreviations are used in conjunction with this test area
@@ -69,16 +51,17 @@ the same state as before the test.
For brevity, the test cases in this test area are summarized together based on
the operations they are testing.
-All these test cases are included in the test case dovetail.osinterop.tc001 of
+All these test cases are included in the test case functest.tempest.osinterop of
OVP test suite.
Test Descriptions
=================
+----------------------
API Used and Reference
----------------------
-Block storage: https://developer.openstack.org/api-ref/block-storage
+Block storage: https://docs.openstack.org/api-ref/block-storage/
- create volume
- delete volume
@@ -94,68 +77,14 @@ Block storage: https://developer.openstack.org/api-ref/block-storage
- update snapshot
- delete snapshot
-------------------------------------------------------------------------
-Test Case 1 - Volume attach and detach operations with the Cinder v2 API
-------------------------------------------------------------------------
-
-Test case specification
------------------------
-
-tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_attach_detach_volume_to_instance
-tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_get_volume_attachment
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_attach_volumes_with_nonexistent_volume_id
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_detach_volumes_with_invalid_volume_id
-
-Test preconditions
-------------------
-
-* Volume extension API
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-''''''''''''''
-* Test action 1: Create a server VM1
-* Test action 2: Attach a provided VOL1 to VM1
-* **Test assertion 1:** Verify VOL1 is in 'in-use' status
-* Test action 3: Detach VOL1 from VM1
-* **Test assertion 2:** Verify VOL1 is in 'available' status
-* Test action 4: Create a server VM2
-* Test action 5: Attach a provided VOL2 to VM2 and wait for VOL2 to reach 'in-use' status
-* Test action 6: Retrieve VOL2's attachment information ATTCH_INFO
-* **Test assertion 3:** Verify ATTCH_INFO is correct
-* Test action 7: Create a server VM3 and wait for VM3 to reach 'ACTIVE' status
-* Test action 8: Attach a non-existent volume to VM3
-* **Test assertion 4:** Verify attach volume failed, a 'NOT FOUND' error is returned in the response
-* Test action 9: Detach a volume from a server by using an invalid volume ID
-* **Test assertion 5:** Verify detach volume failed, a 'NOT FOUND' error is returned in the response
-
-Pass / fail criteria
-''''''''''''''''''''
-
-This test evaluates the volume API ability of attaching a volume to a server
-and detaching a volume from a server. Specifically, the test verifies that:
-
-* Volumes can be attached and detached from servers.
-* Volume attachment information can be retrieved.
-* Attach and detach a volume using an invalid volume ID is not allowed.
-
-In order to pass this test, all test assertions listed in the test execution above need to pass.
-
-Post conditions
----------------
-
-N/A
-
---------------------------------------------------------------------------------
-Test Case 2 - Volume service availability zone operations with the Cinder v2 API
---------------------------------------------------------------------------------
+--------------------------------------------------------------------------------------
+Test Case 1 - Volume service availability zone operations with the Cinder v2 or v3 API
+--------------------------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_availability_zone.AvailabilityZoneV2TestJSON.test_get_availability_zone_list
+tempest.api.volume.test_availability_zone.AvailabilityZoneTestJSON.test_get_availability_zone_list
Test preconditions
------------------
@@ -170,7 +99,7 @@ Test execution
* Test action 1: List all existent availability zones
* **Test assertion 1:** Verify the availability zone list length is greater than 0
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of listing availability zones.
@@ -185,14 +114,14 @@ Post conditions
N/A
---------------------------------------------------------------
-Test Case 3 - Volume cloning operations with the Cinder v2 API
---------------------------------------------------------------
+--------------------------------------------------------------------
+Test Case 2 - Volume cloning operations with the Cinder v2 or v3 API
+--------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_volumes_get.VolumesV2GetTest.test_volume_create_get_update_delete_as_clone
+tempest.api.volume.test_volumes_get.VolumesGetTest.test_volume_create_get_update_delete_as_clone
Test preconditions
------------------
@@ -220,7 +149,7 @@ Test execution
* Test action 7: Update the name of VOL3 and description with the original value
* **Test assertion 6:** Verify VOL3's bootable flag is 'False'
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of creating a cloned volume from a source volume,
@@ -239,15 +168,15 @@ Post conditions
N/A
---------------------------------------------------------------------
-Test Case 4 - Image copy-to-volume operations with the Cinder v2 API
---------------------------------------------------------------------
+--------------------------------------------------------------------------
+Test Case 3 - Image copy-to-volume operations with the Cinder v2 or v3 API
+--------------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_volume_bootable
-tempest.api.volume.test_volumes_get.VolumesV2GetTest.test_volume_create_get_update_delete_from_image
+tempest.api.volume.test_volumes_actions.VolumesActionsTest.test_volume_bootable
+tempest.api.volume.test_volumes_get.VolumesGetTest.test_volume_create_get_update_delete_from_image
Test preconditions
------------------
@@ -279,7 +208,7 @@ Test execution
* Test action 11: Update the name of VOL3 and description with the original value
* **Test assertion 8:** Verify VOL3's bootable flag is 'True'
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of updating volume's bootable flag and creating
@@ -299,20 +228,20 @@ Post conditions
N/A
-----------------------------------------------------------------------------
-Test Case 5 - Volume creation and deletion operations with the Cinder v2 API
-----------------------------------------------------------------------------
+----------------------------------------------------------------------------------
+Test Case 4 - Volume creation and deletion operations with the Cinder v2 or v3 API
+----------------------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_volumes_get.VolumesV2GetTest.test_volume_create_get_update_delete
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_invalid_size
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_nonexistent_source_volid
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_nonexistent_volume_type
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_without_passing_size
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_size_negative
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_size_zero
+tempest.api.volume.test_volumes_get.VolumesGetTest.test_volume_create_get_update_delete
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_create_volume_with_invalid_size
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_create_volume_with_nonexistent_source_volid
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_create_volume_with_nonexistent_volume_type
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_create_volume_without_passing_size
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_create_volume_with_size_negative
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_create_volume_with_size_zero
Test preconditions
------------------
@@ -350,7 +279,7 @@ Test execution
* Test action 13: Create a volume with volume size '0'
* **Test assertion 12:** Verify create volume failed, a bad request error is returned in the response
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of creating a volume, getting volume
@@ -372,14 +301,14 @@ Post conditions
N/A
---------------------------------------------------------------------------------
-Test Case 6 - Volume service extension listing operations with the Cinder v2 API
---------------------------------------------------------------------------------
+--------------------------------------------------------------------------------------
+Test Case 5 - Volume service extension listing operations with the Cinder v2 or v3 API
+--------------------------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_extensions.ExtensionsV2TestJSON.test_list_extensions
+tempest.api.volume.test_extensions.ExtensionsTestJSON.test_list_extensions
Test preconditions
------------------
@@ -396,7 +325,7 @@ Test execution
* Test action 1: List all cinder service extensions
* **Test assertion 1:** Verify all extensions are list in the extension list
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of listing all existent volume service extensions.
@@ -410,16 +339,16 @@ Post conditions
N/A
-----------------------------------------------------------
-Test Case 7 - Volume GET operations with the Cinder v2 API
-----------------------------------------------------------
+----------------------------------------------------------------
+Test Case 6 - Volume GET operations with the Cinder v2 or v3 API
+----------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_get_invalid_volume_id
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_get_volume_without_passing_volume_id
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_volume_get_nonexistent_volume_id
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_get_invalid_volume_id
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_get_volume_without_passing_volume_id
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_volume_get_nonexistent_volume_id
Test preconditions
------------------
@@ -439,7 +368,7 @@ Test execution
* Test action 3: Retrieve a volume with a nonexistent volume ID
* **Test assertion 3:** Verify retrieve volume failed, a 'Not Found' error is returned in the response
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of getting volumes.
@@ -454,32 +383,32 @@ Post conditions
N/A
---------------------------------------------------------------
-Test Case 8 - Volume listing operations with the Cinder v2 API
---------------------------------------------------------------
+--------------------------------------------------------------------
+Test Case 7 - Volume listing operations with the Cinder v2 or v3 API
+--------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_by_name
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_details_by_name
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_param_display_name_and_status
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_with_detail_param_display_name_and_status
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_with_detail_param_metadata
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_with_details
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_with_param_metadata
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volumes_list_by_availability_zone
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volumes_list_by_status
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volumes_list_details_by_availability_zone
-tempest.api.volume.test_volumes_list.VolumesV2ListTestJSON.test_volumes_list_details_by_status
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_list_volumes_detail_with_invalid_status
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_list_volumes_detail_with_nonexistent_name
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_list_volumes_with_invalid_status
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_list_volumes_with_nonexistent_name
-tempest.api.volume.v2.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_details_pagination
-tempest.api.volume.v2.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_details_with_multiple_params
-tempest.api.volume.v2.test_volumes_list.VolumesV2ListTestJSON.test_volume_list_pagination
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_by_name
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_details_by_name
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_param_display_name_and_status
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_with_detail_param_display_name_and_status
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_with_detail_param_metadata
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_with_details
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_with_param_metadata
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volumes_list_by_availability_zone
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volumes_list_by_status
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volumes_list_details_by_availability_zone
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volumes_list_details_by_status
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_list_volumes_detail_with_invalid_status
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_list_volumes_detail_with_nonexistent_name
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_list_volumes_with_invalid_status
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_list_volumes_with_nonexistent_name
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_details_pagination
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_details_with_multiple_params
+tempest.api.volume.test_volumes_list.VolumesListTestJSON.test_volume_list_pagination
Test preconditions
------------------
@@ -540,7 +469,7 @@ Test execution
* Test action 22: List all existent volumes and paginate the volume list by desired volume IDs
* **Test assertion 24:** Verify only the desired volumes are listed in the filtered volume list
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of getting a list of volumes and filtering the volume list.
@@ -558,15 +487,15 @@ Post conditions
N/A
----------------------------------------------------------------
-Test Case 9 - Volume metadata operations with the Cinder v2 API
----------------------------------------------------------------
+---------------------------------------------------------------------
+Test Case 8 - Volume metadata operations with the Cinder v2 or v3 API
+---------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_volume_metadata.VolumesV2MetadataTest.test_crud_volume_metadata
-tempest.api.volume.test_volume_metadata.VolumesV2MetadataTest.test_update_volume_metadata_item
+tempest.api.volume.test_volume_metadata.VolumesMetadataTest.test_crud_volume_metadata
+tempest.api.volume.test_volume_metadata.VolumesMetadataTest.test_update_show_volume_metadata_item
Test preconditions
------------------
@@ -591,7 +520,7 @@ Test execution
* Test action 6: Update one metadata item 'key3' of VOL2
* **Test assertion 5:** Verify the metadata of VOL2 is updated
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of creating metadata for a volume, getting the
@@ -610,14 +539,14 @@ Post conditions
N/A
----------------------------------------------------------------------------------
-Test Case 10 - Verification of read-only status on volumes with the Cinder v2 API
----------------------------------------------------------------------------------
+--------------------------------------------------------------------------------------
+Test Case 9 - Verification of read-only status on volumes with the Cinder v2 or v3 API
+--------------------------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_volume_readonly_update
+tempest.api.volume.test_volumes_actions.VolumesActionsTest.test_volume_readonly_update
Test preconditions
------------------
@@ -635,7 +564,7 @@ Test execution
* Test action 2: Update a provided volume VOL1's read-only access mode to 'False'
* **Test assertion 2:** Verify VOL1 is not in read-only access mode
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of setting and updating volume read-only access mode.
@@ -650,78 +579,25 @@ Post conditions
N/A
--------------------------------------------------------------------
-Test Case 11 - Volume reservation operations with the Cinder v2 API
--------------------------------------------------------------------
+----------------------------------------------------------------------------------------
+Test Case 10 - Volume snapshot creation/deletion operations with the Cinder v2 or v3 API
+----------------------------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_volumes_actions.VolumesV2ActionsTest.test_reserve_unreserve_volume
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_reserve_volume_with_negative_volume_status
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_reserve_volume_with_nonexistent_volume_id
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_unreserve_volume_with_nonexistent_volume_id
-
-Test preconditions
-------------------
-
-* Volume extension API
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Test execution
-''''''''''''''
-
-* Test action 1: Update a provided volume VOL1 as reserved
-* **Test assertion 1:** Verify VOL1 is in 'attaching' status
-* Test action 2: Update VOL1 as un-reserved
-* **Test assertion 2:** Verify VOL1 is in 'available' status
-* Test action 3: Update a provided volume VOL2 as reserved
-* Test action 4: Update VOL2 as reserved again
-* **Test assertion 3:** Verify update VOL2 status failed, a bad request error is returned in the response
-* Test action 5: Update VOL2 as un-reserved
-* Test action 6: Update a non-existent volume as reserved by using an invalid volume ID
-* **Test assertion 4:** Verify update non-existent volume as reserved failed, a 'Not Found' error is returned in the response
-* Test action 7: Update a non-existent volume as un-reserved by using an invalid volume ID
-* **Test assertion 5:** Verify update non-existent volume as un-reserved failed, a 'Not Found' error is returned in the response
-
-Pass / fail criteria
-''''''''''''''''''''
-
-This test case evaluates the volume API ability of reserving and un-reserving volumes.
-Specifically, the test verifies that:
-
-* Volume can be reserved and un-reserved.
-* Update a non-existent volume as reserved is not allowed.
-* Update a non-existent volume as un-reserved is not allowed.
-
-In order to pass this test, all test assertions listed in the test execution above need to pass.
-
-Post conditions
----------------
-
-N/A
-
-----------------------------------------------------------------------------------
-Test Case 12 - Volume snapshot creation/deletion operations with the Cinder v2 API
-----------------------------------------------------------------------------------
-
-Test case specification
------------------------
-
-tempest.api.volume.test_snapshot_metadata.SnapshotV2MetadataTestJSON.test_crud_snapshot_metadata
-tempest.api.volume.test_snapshot_metadata.SnapshotV2MetadataTestJSON.test_update_snapshot_metadata_item
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_create_volume_with_nonexistent_snapshot_id
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_delete_invalid_volume_id
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_delete_volume_without_passing_volume_id
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_volume_delete_nonexistent_volume_id
-tempest.api.volume.test_volumes_snapshots.VolumesV2SnapshotTestJSON.test_snapshot_create_get_list_update_delete
-tempest.api.volume.test_volumes_snapshots.VolumesV2SnapshotTestJSON.test_volume_from_snapshot
-tempest.api.volume.test_volumes_snapshots_list.VolumesV2SnapshotListTestJSON.test_snapshots_list_details_with_params
-tempest.api.volume.test_volumes_snapshots_list.VolumesV2SnapshotListTestJSON.test_snapshots_list_with_params
-tempest.api.volume.test_volumes_snapshots_negative.VolumesV2SnapshotNegativeTestJSON.test_create_snapshot_with_nonexistent_volume_id
-tempest.api.volume.test_volumes_snapshots_negative.VolumesV2SnapshotNegativeTestJSON.test_create_snapshot_without_passing_volume_id
+tempest.api.volume.test_snapshot_metadata.SnapshotMetadataTestJSON.test_crud_snapshot_metadata
+tempest.api.volume.test_snapshot_metadata.SnapshotMetadataTestJSON.test_update_show_snapshot_metadata_item
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_create_volume_with_nonexistent_snapshot_id
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_delete_invalid_volume_id
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_delete_volume_without_passing_volume_id
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_volume_delete_nonexistent_volume_id
+tempest.api.volume.test_volumes_snapshots.VolumesSnapshotTestJSON.test_snapshot_create_get_list_update_delete
+tempest.api.volume.test_volumes_snapshots.VolumesSnapshotTestJSON.test_volume_from_snapshot
+tempest.api.volume.test_volumes_snapshots_list.VolumesSnapshotListTestJSON.test_snapshots_list_details_with_params
+tempest.api.volume.test_volumes_snapshots_list.VolumesSnapshotListTestJSON.test_snapshots_list_with_params
+tempest.api.volume.test_volumes_snapshots_negative.VolumesSnapshotNegativeTestJSON.test_create_snapshot_with_nonexistent_volume_id
+tempest.api.volume.test_volumes_snapshots_negative.VolumesSnapshotNegativeTestJSON.test_create_snapshot_without_passing_volume_id
Test preconditions
------------------
@@ -783,7 +659,7 @@ Test execution
* Test action 27: Create a snapshot from a volume by using an empty volume ID
* **Test assertion 21:** Verify create snapshot failed, a 'Not Found' error is returned in the response
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of managing snapshot and snapshot metadata.
@@ -813,16 +689,16 @@ Post conditions
N/A
---------------------------------------------------------------
-Test Case 13 - Volume update operations with the Cinder v2 API
---------------------------------------------------------------
+--------------------------------------------------------------------
+Test Case 11 - Volume update operations with the Cinder v2 or v3 API
+--------------------------------------------------------------------
Test case specification
-----------------------
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_update_volume_with_empty_volume_id
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_update_volume_with_invalid_volume_id
-tempest.api.volume.test_volumes_negative.VolumesV2NegativeTest.test_update_volume_with_nonexistent_volume_id
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_update_volume_with_empty_volume_id
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_update_volume_with_invalid_volume_id
+tempest.api.volume.test_volumes_negative.VolumesNegativeTest.test_update_volume_with_nonexistent_volume_id
Test preconditions
------------------
@@ -842,7 +718,7 @@ Test execution
* Test action 3: Update a non-existent volume by using a random generated volume ID
* **Test assertion 3:** Verify update volume failed, a 'Not Found' error is returned in the response
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test case evaluates the volume API ability of updating volume attributes.
@@ -858,3 +734,41 @@ Post conditions
---------------
N/A
+
+-------------------------------------------------------------------
+Test Case 12 - Volume list version operation with the Cinder v3 API
+-------------------------------------------------------------------
+
+Test case specification
+-----------------------
+
+tempest.api.volume.test_versions.VersionsTest.test_list_versions
+
+Test preconditions
+------------------
+
+* Volume API
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+''''''''''''''
+
+* Test action 1: List volume API versions
+* **Test assertion 1:** The list versions operation is successful executed
+
+Pass / Fail criteria
+''''''''''''''''''''
+
+This test case evaluates the volume API ability of listing volume API versions.
+Specifically, the test verifies that:
+
+* Successfully listing volume API versions.
+
+In order to pass this test, all test assertions listed in the test execution above need to pass.
+
+Post conditions
+---------------
+
+N/A
diff --git a/docs/testing/user/testspecification/tempest_trunk_ports/index.rst b/docs/testing/user/testspecification/tempest_trunk_ports/index.rst
new file mode 100644
index 00000000..f82fc468
--- /dev/null
+++ b/docs/testing/user/testspecification/tempest_trunk_ports/index.rst
@@ -0,0 +1,130 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV
+
+================================
+Neutron Trunk Port Tempest Tests
+================================
+
+Scope
+=====
+
+This test area evaluates the ability of a system under test to support Neutron
+trunk ports. The test area specifically validates port and sub-port API CRUD
+operations, by means of both positive and negative tests.
+
+
+References
+================
+
+- `OpenStack API reference <https://docs.openstack.org/api-ref/network/v2/#trunk-networking>`_
+
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVI and VIM deployed on a Pharos
+compliant infrastructure.
+
+
+Test Area Structure
+====================
+
+The test area is structured in individual tests as listed below. Each test case
+is able to run independently, i.e. irrelevant of the state created by a previous
+test. For detailed information on the individual steps and assertions performed
+by the tests, review the Python source code accessible via the following links:
+
+- `Neutron Trunk API tests <https://github.com/openstack/neutron-tempest-plugin/blob/0.3.0/neutron_tempest_plugin/api/test_trunk.py>`_
+- `Neutron Trunk API trunk details <https://github.com/openstack/neutron-tempest-plugin/blob/0.3.0/neutron_tempest_plugin/api/test_trunk_details.py>`_
+- `Neutron Trunk API negative tests <https://github.com/openstack/neutron-tempest-plugin/blob/0.3.0/neutron_tempest_plugin/api/test_trunk_negative.py>`_
+
+
+**Trunk port and sub-port CRUD operations:**
+
+These tests cover the CRUD (Create, Read, Update, Delete) life-cycle operations
+of trunk ports and subports.
+
+Implementation:
+`TrunkTestJSON <https://github.com/openstack/neutron-tempest-plugin/blob/0.3.0/neutron_tempest_plugin/api/test_trunk.py#L71>`_
+
+
+- neutron_tempest_plugin.api.test_trunk.TrunkTestJSON.test_add_subports
+- neutron_tempest_plugin.api.test_trunk.TrunkTestJSON.test_create_show_delete_trunk
+- neutron_tempest_plugin.api.test_trunk.TrunkTestJSON.test_create_trunk_empty_subports_list
+- neutron_tempest_plugin.api.test_trunk.TrunkTestJSON.test_create_trunk_subports_not_specified
+- neutron_tempest_plugin.api.test_trunk.TrunkTestJSON.test_delete_trunk_with_subport_is_allowed
+- neutron_tempest_plugin.api.test_trunk.TrunkTestJSON.test_get_subports
+- neutron_tempest_plugin.api.test_trunk.TrunkTestJSON.test_list_trunks
+- neutron_tempest_plugin.api.test_trunk.TrunkTestJSON.test_remove_subport
+
+
+**API for listing query results:**
+
+These tests verify that listing operations of trunk port objects work. This
+functionality is required for CLI and UI operations.
+
+Implementation:
+`TrunksSearchCriteriaTest <https://github.com/openstack/neutron-tempest-plugin/blob/0.3.0/neutron_tempest_plugin/api/test_trunk.py#L306>`_
+
+- neutron_tempest_plugin.api.test_trunk.TrunksSearchCriteriaTest.test_list_no_pagination_limit_0
+- neutron_tempest_plugin.api.test_trunk.TrunksSearchCriteriaTest.test_list_pagination
+- neutron_tempest_plugin.api.test_trunk.TrunksSearchCriteriaTest.test_list_pagination_page_reverse_asc
+- neutron_tempest_plugin.api.test_trunk.TrunksSearchCriteriaTest.test_list_pagination_page_reverse_desc
+- neutron_tempest_plugin.api.test_trunk.TrunksSearchCriteriaTest.test_list_pagination_with_marker
+- neutron_tempest_plugin.api.test_trunk.TrunksSearchCriteriaTest.test_list_sorts_asc
+- neutron_tempest_plugin.api.test_trunk.TrunksSearchCriteriaTest.test_list_sorts_desc
+
+
+**Query trunk port details:**
+
+These tests validate that all attributes of trunk port objects can be queried.
+
+Implementation:
+`TestTrunkDetailsJSON <https://github.com/openstack/neutron-tempest-plugin/blob/0.3.0/neutron_tempest_plugin/api/test_trunk_details.py#L20>`_
+
+- neutron_tempest_plugin.api.test_trunk_details.TestTrunkDetailsJSON.test_port_resource_empty_trunk_details
+- neutron_tempest_plugin.api.test_trunk_details.TestTrunkDetailsJSON.test_port_resource_trunk_details_no_subports
+- neutron_tempest_plugin.api.test_trunk_details.TestTrunkDetailsJSON.test_port_resource_trunk_details_with_subport
+
+
+**Negative tests:**
+
+These group of tests comprise negative tests which verify that invalid operations
+are handled correctly by the system under test.
+
+Implementation:
+`TrunkTestNegative <https://github.com/openstack/neutron-tempest-plugin/blob/0.3.0/neutron_tempest_plugin/api/test_trunk_negative.py#L27>`_
+
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_add_subport_duplicate_segmentation_details
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_add_subport_passing_dict
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_add_subport_port_id_disabled_trunk
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_add_subport_port_id_uses_parent_port_id
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_create_subport_missing_segmentation_id
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_create_subport_nonexistent_port_id
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_create_subport_nonexistent_trunk
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_create_trunk_duplicate_subport_segmentation_ids
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_create_trunk_nonexistent_port_id
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_create_trunk_nonexistent_subport_port_id
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_create_trunk_with_subport_missing_port_id
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_create_trunk_with_subport_missing_segmentation_id
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_create_trunk_with_subport_missing_segmentation_type
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_delete_port_in_use_by_subport
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_delete_port_in_use_by_trunk
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_delete_trunk_disabled_trunk
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_remove_subport_not_found
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_remove_subport_passing_dict
+- neutron_tempest_plugin.api.test_trunk_negative.TrunkTestJSON.test_remove_subport_port_id_disabled_trunk
+
+
+**Scenario tests (tests covering more than one functionality):**
+
+In contrast to the API tests above, these tests validate more than one specific
+API capability. Instead they verify that a simple scenario (example workflow)
+functions as intended. To this end, they boot up two VMs with trunk ports and
+sub ports and verify connectivity between those VMs.
+
+Implementation:
+`TrunkTest <https://github.com/openstack/neutron-tempest-plugin/blob/0.3.0/neutron_tempest_plugin/scenario/test_trunk.py#L41>`_
+
+- neutron_tempest_plugin.scenario.test_trunk.TrunkTest.test_trunk_subport_lifecycle
diff --git a/docs/testing/user/testspecification/machinelifecycle/index.rst b/docs/testing/user/testspecification/tempest_vm_lifecycle/index.rst
index b0cc0d79..1621d8f5 100644
--- a/docs/testing/user/testspecification/machinelifecycle/index.rst
+++ b/docs/testing/user/testspecification/tempest_vm_lifecycle/index.rst
@@ -59,44 +59,45 @@ Each test case is able to run independently, i.e. irrelevant of the state
created by a previous test. Specifically, every test performs clean-up
operations which return the system to the same state as before the test.
-All these test cases are included in the test case dovetail.tempest.tc004 of
+All these test cases are included in the test case functest.tempest.vm_lifecycle of
OVP test suite.
Test Descriptions
=================
+----------------------
API Used and Reference
----------------------
-Block storage: https://developer.openstack.org/api-ref/block-storage
+Block storage: https://docs.openstack.org/api-ref/block-storage/
- create volume
- delete volume
- attach volume to server
- detach volume from server
-Security Groups: https://developer.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
+Security Groups: https://docs.openstack.org/api-ref/network/v2/index.html#security-groups-security-groups
- create security group
- delete security group
-Networks: https://developer.openstack.org/api-ref/networking/v2/index.html#networks
+Networks: https://docs.openstack.org/api-ref/network/v2/index.html#networks
- create network
- delete network
-Routers and interface: https://developer.openstack.org/api-ref/networking/v2/index.html#routers-routers
+Routers and interface: https://docs.openstack.org/api-ref/network/v2/index.html#routers-routers
- create router
- delete router
- add interface to router
-Subnets: https://developer.openstack.org/api-ref/networking/v2/index.html#subnets
+Subnets: https://docs.openstack.org/api-ref/network/v2/index.html#subnets
- create subnet
- delete subnet
-Servers: https://developer.openstack.org/api-ref/compute/
+Servers: https://docs.openstack.org/api-ref/compute/
- create keypair
- create server
@@ -119,17 +120,17 @@ Servers: https://developer.openstack.org/api-ref/compute/
- migrate server
- live-migrate server
-Ports: https://developer.openstack.org/api-ref/networking/v2/index.html#ports
+Ports: https://docs.openstack.org/api-ref/network/v2/index.html#ports
- create port
- delete port
-Floating IPs: https://developer.openstack.org/api-ref/networking/v2/index.html#floating-ips-floatingips
+Floating IPs: https://docs.openstack.org/api-ref/network/v2/index.html#floating-ips-floatingips
- create floating IP
- delete floating IP
-Availability zone: https://developer.openstack.org/api-ref/compute/
+Availability zone: https://docs.openstack.org/api-ref/compute/
- get availability zone
@@ -173,7 +174,7 @@ Test execution
* **Test assertion 8:** Verify VM1's addresses have been refreshed after disassociating FIP1
* Test action 10: Delete SG1, IMG1, KEYP1, VOL1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates a minimum basic scenario. Specifically, the test verifies that:
@@ -226,7 +227,7 @@ Test execution
* **Test assertion 3:** Verify SRC_HOST does not equal to DST_HOST
* Test action 11: Delete KEYP1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to cold migrate VMs. Specifically, the test verifies that:
@@ -273,7 +274,7 @@ Test execution
* **Test assertion 3:** Verify can ping FIP1 successfully and can SSH to VM1 via FIP1
* Test action 8: Delete KEYP1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to pause and unpause VMs. Specifically, the test verifies that:
@@ -318,7 +319,7 @@ Test execution
* **Test assertion 1:** Verify can ping FIP1 successfully and can SSH to VM1 via FIP1
* Test action 6: Delete KEYP1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to reboot servers. Specifically, the test verifies that:
@@ -361,7 +362,7 @@ Test execution
* **Test assertion 1:** Verify can ping FIP1 successfully and can SSH to VM1 via FIP1
* Test action 6: Delete KEYP1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to rebuild servers. Specifically, the test verifies that:
@@ -406,7 +407,7 @@ Test execution
* **Test assertion 1:** Verify can ping FIP1 successfully and can SSH to VM1 via FIP1
* Test action 8: Delete KEYP1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to resize servers. Specifically, the test verifies that:
@@ -452,7 +453,7 @@ Test execution
* **Test assertion 2:** Verify can ping FIP1 successfully and can SSH to VM1 via FIP1
* Test action 8: Delete KEYP1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to stop and start servers. Specifically, the test verifies that:
@@ -500,7 +501,7 @@ Test execution
* **Test assertion 2:** Verify can ping FIP1 successfully and can SSH to VM1 via FIP1
* Test action 8: Delete KEYP1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to suspend and resume servers. Specifically, the test verifies that:
@@ -552,7 +553,7 @@ Test execution
* **Test assertion 4:** Verify VM1 status is 'ACTIVE'
* Test action 10: Delete KEYP1, VM1 and FIP1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to suspend and resume servers in sequence.
@@ -574,7 +575,7 @@ Test Case 10 - Resize volume backed server
Test case specification
-----------------------
-tempest.scenario.test_server_advanced_ops.TestServerAdvancedOps.test_resize_volume_backed_server_confirm
+tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_resize_volume_backed_server_confirm
Test preconditions
------------------
@@ -596,7 +597,7 @@ Test execution
* **Test assertion 1:** VM1's status is 'ACTIVE'
* Test action 6: Delete VM1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to resize volume backed servers.
@@ -645,7 +646,7 @@ Test execution
* **Test assertion 1:** Verify the values written and read are the same
* Test action 9: Delete SG1, KEYP1 and VM1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to shelve and unshelve servers.
@@ -693,7 +694,7 @@ Test execution
* **Test assertion 1:** Verify T_STAMP1 equals to T_STAMP2
* Test action 9: Delete SG1, KEYP1 and VM1
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates the ability to shelve and unshelve volume backed servers.
diff --git a/docs/testing/user/testspecification/tempest_volume/index.rst b/docs/testing/user/testspecification/tempest_volume/index.rst
new file mode 100644
index 00000000..64888c99
--- /dev/null
+++ b/docs/testing/user/testspecification/tempest_volume/index.rst
@@ -0,0 +1,128 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) Ericsson AB
+
+=================================
+Tempest Volume test specification
+=================================
+
+
+Scope
+=====
+
+This test area evaluates the ability of a system under test to manage volumes.
+
+The test area specifically validates the creation, the deletion and
+the attachment/detach volume operations.
+tests.
+
+
+References
+==========
+
+N/A
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the NFVi and VIM in operation on a
+Pharos compliant infrastructure.
+
+Test Area Structure
+===================
+
+The test area is structured in individual tests as listed below.
+For detailed information on the individual steps and assertions performed
+by the tests, review the Python source code accessible via the following links:
+
+All these test cases are included in the test case functest.tempest.volume of
+OVP test suite.
+
+
+----------------------------------------------
+Test Case 1 - Attach Detach Volume to Instance
+----------------------------------------------
+
+Test case specification
+-----------------------
+
+Implementation: `Attach Detach Volume to Instance <https://github.com/openstack/tempest/blob/17.1.0/tempest/api/volume/test_volumes_actions.py>`_
+
+* tempest.api.volume.test_volumes_actions.VolumesActionsTest.test_attach_detach_volume_to_instance
+
+Test preconditions
+------------------
+
+* Volume extension API
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+''''''''''''''
+* Test action 1: Create a server VM1
+* Test action 2: Attach a provided VOL1 to VM1
+* **Test assertion 1:** Verify VOL1 is in 'in-use' status
+* Test action 3: Detach VOL1 from VM1
+* **Test assertion 2:** Verify detach volume VOL1 successfully and VOL1 is in 'available' status
+
+Pass / fail criteria
+''''''''''''''''''''
+
+This test evaluates the volume API ability of attaching a volume to a server
+and detaching a volume from a server. Specifically, the test verifies that:
+
+* Volumes can be attached and detached from servers.
+
+In order to pass this test, all test assertions listed in the test execution above need to pass.
+
+Post conditions
+---------------
+
+N/A
+
+--------------------------------------
+Test Case 2 - Volume Boot Pattern test
+--------------------------------------
+
+Test case specification
+-----------------------
+
+Implementation: `Volume Boot Pattern test <https://github.com/openstack/tempest/blob/17.1.0/tempest/scenario/test_volume_boot_pattern.py>`_
+
+* tempest.scenario.test_volume_boot_pattern.TestVolumeBootPattern.test_volume_boot_pattern
+
+Test preconditions
+------------------
+
+* Volume extension API
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Test execution
+''''''''''''''
+* Test action 1:Create in Cinder some bootable volume VOL1 importing a Glance image
+* Test action 2:Boot an instance VM1 from the bootable volume VOL1
+* Test action 3:Write content to the VOL1
+* Test action 4:Delete VM1 and Boot a new instance VM2 from the volume VOL1
+* Test action 5:Check written content in the instance
+* **Test assertion 1:** Verify the content of written file in action 3
+* Test action 6:Create a volume snapshot VOL2 while the instance VM2 is running
+* Test action 7:Boot an additional instance VM3 from the new snapshot based volume VOL2
+* Test action 8:Check written content in the instance booted from snapshot
+* **Test assertion 2:** Verify the content of written file in action 3
+
+Pass / fail criteria
+''''''''''''''''''''
+
+This test evaluates the volume storage consistency. Specifically, the test verifies that:
+
+* The content of written file in the volume.
+
+In order to pass this test, all test assertions listed in the test execution above need to pass.
+
+Post conditions
+---------------
+
+N/A
diff --git a/docs/testing/user/testspecification/vnf/index.rst b/docs/testing/user/testspecification/vnf/index.rst
new file mode 100644
index 00000000..d13634a1
--- /dev/null
+++ b/docs/testing/user/testspecification/vnf/index.rst
@@ -0,0 +1,222 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) Ericsson AB
+
+======================
+VNF test specification
+======================
+
+.. toctree::
+ :maxdepth: 2
+
+Scope
+=====
+
+The VNF test area evaluates basic NFV capabilities of the system under test.
+These capabilities include creating a small number of virtual machines,
+establishing the SUT VNF, VNFs which are going to support the test activities
+and an Orchestrator as well as verifying the proper behavior of the basic VNF.
+
+
+References
+==========
+
+This test area references the following specifications and guides:
+
+- Functest repo for detailed description of the vEPC testcase
+
+ - https://github.com/opnfv/functest/blob/master/docs/testing/user/userguide/test_details.rst#juju_epc
+
+- Functest repo for detailed description of the vIMS testcase
+
+ - https://github.com/opnfv/functest/blob/master/docs/testing/user/userguide/test_details.rst#cloudify_ims
+
+- 3GPP LTE
+
+ - https://www.3gpp.org/technologies/keywords-acronyms/98-lte
+
+- ETSI - TS 24.301
+
+ - https://www.etsi.org/deliver/etsi_ts/124300_124399/124301/10.03.00_60/ts_124301v100300p.pdf
+
+- Cloudify clearwater: opnfv-cloudify-clearwater [1]
+
+ - https://github.com/Orange-OpenSource/opnfv-cloudify-clearwater
+
+
+Definitions and abbreviations
+=============================
+
+The following terms and abbreviations are used in conjunction with this test
+area
+
+- 3GPP - 3rd Generation Partnership Project
+- EPC - Evolved Packet Core
+- ETSI - European Telecommunications Standards Institute
+- IMS - IP Multimedia Core Network Subsystem
+- LTE - Long Term Evolution
+- NFV - Network functions virtualization
+- OAI - Open Air Interface
+- TS - Technical Specifications
+- VM - Virtual machine
+- VNF - Virtual Network Function
+
+
+
+System Under Test (SUT)
+=======================
+
+The system under test is assumed to be the VNF and VIM in operation on a
+Pharos compliant infrastructure.
+
+
+Test Area Structure
+===================
+
+The test area is structured in two separate tests which are executed
+sequentially. The order of the tests is arbitrary as there are no dependencies
+across the tests. Specifically, every test performs clean-up operations which
+return the system to the same state as before the test.
+
+
+Test Descriptions
+=================
+
+------------------
+Test Case 1 - vEPC
+------------------
+
+Short name
+----------
+
+functest.vnf.vepc
+
+
+Use case specification
+----------------------
+
+The Evolved Packet Core (EPC) is the main component of the System Architecture
+Evolution (SAE) which forms the core of the 3GPP LTE specification.
+
+vEPC has been integrated in Functest to demonstrate the capability to deploy
+a complex mobility-specific NFV scenario on the OPNFV platform. The OAI EPC
+supports most of the essential functions defined by the 3GPP Technical Specs;
+hence the successful execution of functional tests on the OAI EPC provides a
+good endorsement of the underlying NFV platform.
+
+
+Test preconditions
+------------------
+
+At least one compute node is available. No further pre-configuration needed.
+
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+Methodology for verifying connectivity
+''''''''''''''''''''''''''''''''''''''
+
+This integration also includes ABot, a Test Orchestration system that enables
+test scenarios to be defined in high-level DSL. ABot is also deployed as a VM
+on the OPNFV platform; and this provides an example of the automation driver
+and the Test VNF being both deployed as separate VNFs on the underlying OPNFV
+platform.
+
+
+Test execution
+''''''''''''''
+* Test action 1: Deploy Juju controller (VNF Manager) using Bootstrap command.
+* Test action 2: Deploy ABot (Orchestrator) and OAI EPC as Juju charms.
+ Configuration of ABot and OAI EPC components is handled through built-in Juju
+ relations.
+* Test action 3: Execution of ABot feature files triggered by Juju actions.
+ This executes a suite of LTE signalling tests on the OAI EPC.
+* Test action 4: ABot test results are parsed accordingly.
+* Test action 5: The deployed VMs are deleted.
+
+
+Pass / Fail criteria
+''''''''''''''''''''
+
+The VNF Manager (juju) should be deployed successfully
+
+Test executor (ABot), test Orchestration system is deployed and enables test
+scenarios to be defined in high-level DSL
+
+VMs which are act as VNFs (including the VNF that is the SUT for test case) are
+following the 3GPP technical specifications accordingly.
+
+
+Post conditions
+---------------
+
+The clean-up operations are run.
+
+------------------
+Test Case 2 - vIMS
+------------------
+
+Short name
+----------
+
+functest.vnf.vims
+
+Use case specification
+----------------------
+
+The IP Multimedia Subsystem or IP Multimedia Core Network Subsystem (IMS) is
+an architectural framework for delivering IP multimedia services.
+
+vIMS test case is integrated to demonstrate the capability to deploy a
+relatively complex NFV scenario on top of the OPNFV infrastructure.
+
+Example of a real VNF deployment to show the NFV capabilities of the platform.
+The IP Multimedia Subsytem is a typical Telco test case, referenced by ETSI.
+It provides a fully functional VoIP System.
+
+Test preconditions
+------------------
+
+Certain ubuntu server and cloudify images version refer to dovetail testing user guide.
+
+At least 30G RAMs and 50 vcpu cores required.
+
+Basic test flow execution description and pass/fail criteria
+------------------------------------------------------------
+
+vIMS has been integrated in Functest to demonstrate the capability to deploy
+a relatively complex NFV scenario on the OPNFV platform. The deployment of a
+complete functional VNF allows the test of most of the essential functions
+needed for a NFV platform.
+
+
+Test execution
+''''''''''''''
+* Test action 1: Deploy a VNF orchestrator (Cloudify).
+* Test action 2: Deploy a Clearwater vIMS (IP Multimedia Subsystem) VNF from
+ this orchestrator based on a TOSCA blueprint defined in repository of
+ opnfv-cloudify-clearwater [1].
+* Test action 3: Run suite of signaling tests on top of this VNF
+* Test action 4: Collect test results.
+* Test action 5: The deployed VMs are deleted.
+
+
+Pass / Fail criteria
+''''''''''''''''''''
+
+The VNF orchestrator (Cloudify) should be deployed successfully.
+
+The Clearwater vIMS (IP Multimedia Subsystem) VNF from this orchestrator
+should be deployed successfully.
+
+The suite of signaling tests on top of vIMS should be run successfully.
+
+The test scenarios on the NFV platform should be executed successfully following
+the ETSI standards accordingly.
+
+
+Post conditions
+---------------
+
+All resources created during the test run have been cleaned-up
diff --git a/docs/testing/user/testspecification/vping/index.rst b/docs/testing/user/testspecification/vping/index.rst
index 486ee842..355018c6 100644
--- a/docs/testing/user/testspecification/vping/index.rst
+++ b/docs/testing/user/testspecification/vping/index.rst
@@ -3,7 +3,7 @@
.. (c) Ericsson AB
========================
-Vping test specification
+Vping Test Specification
========================
.. toctree::
@@ -21,13 +21,9 @@ means of ICMP packets.
References
==========
-- Neutron Client
+- OpenStack Shade
- - https://docs.openstack.org/developer/python-neutronclient/usage/library.html
-
-- Nova Client
-
- - https://docs.openstack.org/developer/python-novaclient/ref/v2/servers.html
+ - https://docs.openstack.org/shade/latest/user/index.html
- SSHClient
@@ -35,10 +31,10 @@ References
- SCPClient
- - https://pypi.python.org/pypi/scp
+ - https://pypi.org/project/scp/
-Definitions and abbreviations
+Definitions and Abbreviations
=============================
The following terms and abbreviations are used in conjunction with this test
@@ -46,10 +42,10 @@ area
- ICMP - Internet Control Message Protocol
- L3 - Layer 3
-- NFVi - Network functions virtualization infrastructure
-- SCP - Secure Copy
-- SSH - Secure Shell
-- VM - Virtual machine
+- NFVi - Network Functions Virtualization infrastructure
+- SCP - Secure CoPy
+- SSH - Secure SHell
+- VM - Virtual Machine
System Under Test (SUT)
@@ -77,7 +73,7 @@ Test Case 1 - vPing using userdata provided by nova metadata service
Short name
----------
-dovetail.vping.tc001.userdata
+functest.vping.userdata
Use case specification
@@ -110,39 +106,46 @@ checking the console output of the source VMs.
Test execution
''''''''''''''
-* Test action 1:
- * Create a private tenant network by using neutron client
- * Create one subnet and one router in the network by neutron client
- * Add one interface between the subnet and router
- * Add one gateway route to the router by neutron client
- * Store the network id in the response
+* **Test action 1:**
+ * Create a private tenant network by using neutron client
+ * Create one subnet and one router in the network by neutron client
+ * Add one interface between the subnet and router
+ * Add one gateway route to the router by neutron client
+ * Store the network id in the response
* **Test assertion 1:** The network id, subnet id and router id can be found in the response
-* Test action 2:
- * Create an security group by using neutron client
- * Store the security group id parameter in the response
+* **Test action 2:**
+ * Create an security group by using neutron client
+ * Store the security group id parameter in the response
* **Test assertion 2:** The security group id can be found in the response
-* Test action 3: boot VM1 by using nova client with configured name, image, flavor, private tenant
- network created in test action 1, security group created in test action 2
+* **Test action 3:**
+ * Boot VM1 by using nova client with configured name, image, flavor, private tenant
+ network created in test action 1 and security group created in test action 2
* **Test assertion 3:** The VM1 object can be found in the response
-* Test action 4: Generate ping script with the IP of VM1 to be passed as userdata provided by
- the **nova metadata service**.
-* Test action 5: Boot VM2 by using nova client with configured name, image, flavor, private tenant
- network created in test action 1, security group created in test action 2, userdata created
- in test action 4
+* **Test action 4:**
+ * Generate ping script with the IP of VM1 to be passed as userdata provided by
+ the **nova metadata service**
+* **Test action 5:**
+ * Boot VM2 by using nova client with configured name, image, flavor, private tenant
+ network created in test action 1, security group created in test action 2, and
+ userdata created in test action 4
* **Test assertion 4:** The VM2 object can be found in the response
-* Test action 6: Inside VM2, the ping script is executed automatically when booted and it contains a
- loop doing the ping until the return code is 0 or timeout reached. For each ping, when the return
- code is 0, "vPing OK" is printed in the VM2 console-log, otherwise, "vPing KO" is printed.
- Monitoring the console-log of VM2 to see the response generated by the script.
+* **Test action 6:**
+ * Inside VM2, the ping script is executed automatically when booted and it contains a
+ loop doing the ping until the return code is 0 or timeout reached
+ * For each ping, when the return code is 0, "vPing OK" is printed in the VM2 console-log,
+ otherwise, "vPing KO" is printed
+ * Monitoring the console-log of VM2 to see the response generated by the script
* **Test assertion 5:** "vPing OK" is detected, when monitoring the console-log in VM2
-* Test action 7: delete VM1, VM2
+* **Test action 7:**
+ * Delete VM1, VM2
* **Test assertion 6:** VM1 and VM2 are not present in the VM list
-* Test action 8: delete security group, gateway, interface, router, subnet and network
+* **Test action 8:**
+ * Delete security group, gateway, interface, router, subnet and network
* **Test assertion 7:** The security group, gateway, interface, router, subnet and network are
no longer present in the lists after deleting
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates basic NFVi capabilities of the system under test.
@@ -177,7 +180,7 @@ Test Case 2 - vPing using SSH to a floating IP
Short name
----------
-dovetail.vping.tc002.ssh
+functest.vping.ssh
Use case specification
@@ -211,45 +214,55 @@ Test execution
''''''''''''''
-* Test action 1:
- * Create a private tenant network by neutron client
- * Create one subnet and one router are created in the network by using neutron client
- * Create one interface between the subnet and router
- * Add one gateway route to the router by neutron client
- * Store the network id in the response
+* **Test action 1:**
+ * Create a private tenant network by neutron client
+ * Create one subnet and one router are created in the network by using neutron client
+ * Create one interface between the subnet and router
+ * Add one gateway route to the router by neutron client
+ * Store the network id in the response
* **Test assertion 1:** The network id, subnet id and router id can be found in the response
-* Test action 2:
- * Create an security group by using neutron client
- * Store the security group id parameter in the response
+* **Test action 2:**
+ * Create an security group by using neutron client
+ * Store the security group id parameter in the response
* **Test assertion 2:** The security group id can be found in the response
-* Test action 3: Boot VM1 by using nova client with configured name, image, flavor, private tenant
- network created in test action 1, security group created in test action 2
+* **Test action 3:**
+ * Boot VM1 by using nova client with configured name, image, flavor, private tenant
+ network created in test action 1 and security group created in test action 2
* **Test assertion 3:** The VM1 object can be found in the response
-* Test action 4: Boot VM2 by using nova client with configured name, image, flavor, private tenant
- network created in test action 1, security group created in test action 2
+* **Test action 4:**
+ * Boot VM2 by using nova client with configured name, image, flavor, private tenant
+ network created in test action 1 and security group created in test action 2
* **Test assertion 4:** The VM2 object can be found in the response
-* Test action 5: create one floating IP by using neutron client, storing the floating IP address
- returned in the response
+* **Test action 5:**
+ * Create one floating IP by using neutron client, storing the floating IP address
+ returned in the response
* **Test assertion 5:** Floating IP address can be found in the response
-* Test action 6: Assign the floating IP address created in test action 5 to VM2 by using nova client
+* **Test action 6:**
+ * Assign the floating IP address created in test action 5 to VM2 by using nova client
* **Test assertion 6:** The assigned floating IP can be found in the VM2 console log file
-* Test action 7: Establish SSH connection between the test host and VM2 through the floating IP
+* **Test action 7:**
+ * Establish SSH connection between the test host and VM2 through the floating IP
* **Test assertion 7:** SSH connection between the test host and VM2 is established within
300 seconds
-* Test action 8: Copy the Ping script from the test host to VM2 by using SCPClient
+* **Test action 8:**
+ * Copy the Ping script from the test host to VM2 by using SCPClient
* **Test assertion 8:** The Ping script can be found inside VM2
-* Test action 9: Inside VM2, to execute the Ping script to ping VM1, the Ping script contains a
- loop doing the ping until the return code is 0 or timeout reached, for each ping, when the return
- code is 0, "vPing OK" is printed in the VM2 console-log, otherwise, "vPing KO" is printed.
- Monitoring the console-log of VM2 to see the response generated by the script.
+* **Test action 9:**
+ * Inside VM2, to execute the Ping script to ping VM1, the Ping script contains a
+ loop doing the ping until the return code is 0 or timeout reached
+ * For each ping, when the return code is 0, "vPing OK" is printed in the VM2 console-log,
+ otherwise, "vPing KO" is printed
+ * Monitoring the console-log of VM2 to see the response generated by the script
* **Test assertion 9:** "vPing OK" is detected, when monitoring the console-log in VM2
-* Test action 10: delete VM1, VM2
+* **Test action 10:**
+ * Delete VM1, VM2
* **Test assertion 10:** VM1 and VM2 are not present in the VM list
-* Test action 11: delete floating IP, security group, gateway, interface, router, subnet and network
+* **Test action 11:**
+ * Delete floating IP, security group, gateway, interface, router, subnet and network
* **Test assertion 11:** The security group, gateway, interface, router, subnet and network are
no longer present in the lists after deleting
-Pass / fail criteria
+Pass / Fail criteria
''''''''''''''''''''
This test evaluates basic NFVi capabilities of the system under test.
diff --git a/docs/testing/user/testspecification/vpn/index.rst b/docs/testing/user/testspecification/vpn/index.rst
index 13c0cfa6..9d0a4ad3 100644
--- a/docs/testing/user/testspecification/vpn/index.rst
+++ b/docs/testing/user/testspecification/vpn/index.rst
@@ -13,9 +13,8 @@ Scope
=====
The VPN test area evaluates the ability of the system under test to support VPN
-networking for virtual workloads. The tests in this test area will evaluate
-establishing VPN networks, publishing and communication between endpoints using
-BGP and tear down of the networks.
+networking for virtual workloads. This test area tests CRUD (Create, Read, Update,
+Delete) operations of BGPVPN API.
References
==========
@@ -44,14 +43,9 @@ The following terms and abbreviations are used in conjunction with this test
area
- BGP - Border gateway protocol
-- eRT - Export route target
-- IETF - Internet Engineering Task Force
-- iRT - Import route target
- NFVi - Network functions virtualization infrastructure
-- Tenant - An isolated set of virtualized infrastructures
- VM - Virtual machine
- VPN - Virtual private network
-- VLAN - Virtual local area network
System Under Test (SUT)
@@ -64,466 +58,202 @@ Pharos compliant infrastructure.
Test Area Structure
===================
-The test area is structured in four separate tests which are executed
+The test area is structured in several tempest tests which are executed
sequentially. The order of the tests is arbitrary as there are no dependencies
across the tests. Specifially, every test performs clean-up operations which
return the system to the same state as before the test.
-The test area evaluates the ability of the SUT to establish connectivity
-between Virtual Machines using an appropriate route target configuration,
-reconfigure the route targets to remove connectivity between the VMs, then
-reestablish connectivity by re-association.
-
Test Descriptions
=================
-----------------------------------------------------------------
-Test Case 1 - VPN provides connectivity between Neutron subnets
-----------------------------------------------------------------
-
-Short name
-----------
-
-dovetail.sdnvpn.tc001.subnet_connectivity
-
-
-Use case specification
-----------------------
-
-This test evaluates the use case where an NFVi tenant uses a BGPVPN to provide
-connectivity between VMs on different Neutron networks and subnets that reside
-on different hosts.
-
-
-Test preconditions
-------------------
-
-2 compute nodes are available, denoted Node1 and Node2 in the following.
-
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Methodology for verifying connectivity
-''''''''''''''''''''''''''''''''''''''
-
-Connectivity between VMs is tested by sending ICMP ping packets between
-selected VMs. The target IPs are passed to the VMs sending pings by means of a
-custom user data script. Whether or not a ping was successful is determined by
-checking the console output of the source VMs.
-
-
-Test execution
-''''''''''''''
-
-* Create Neutron network N1 and subnet SN1 with IP range 10.10.10.0/24
-* Create Neutron network N2 and subnet SN2 with IP range 10.10.11.0/24
-
-* Create VM1 on Node1 with a port in network N1
-* Create VM2 on Node1 with a port in network N1
-* Create VM3 on Node2 with a port in network N1
-* Create VM4 on Node1 with a port in network N2
-* Create VM5 on Node2 with a port in network N2
-
-* Create VPN1 with eRT<>iRT
-* Create network association between network N1 and VPN1
-
-* VM1 sends ICMP packets to VM2 using ``ping``
-
-* **Test assertion 1:** Ping from VM1 to VM2 succeeds: ``ping`` exits with return code 0
-
-* VM1 sends ICMP packets to VM3 using ``ping``
-
-* **Test assertion 2:** Ping from VM1 to VM3 succeeds: ``ping`` exits with return code 0
-
-* VM1 sends ICMP packets to VM4 using ``ping``
-
-* **Test assertion 3:** Ping from VM1 to VM4 fails: ``ping`` exits with a non-zero return code
-
-* Create network association between network N2 and VPN1
-
-* VM4 sends ICMP packets to VM5 using ``ping``
-
-* **Test assertion 4:** Ping from VM4 to VM5 succeeds: ``ping`` exits with return code 0
-
-* Configure iRT=eRT in VPN1
-
-* VM1 sends ICMP packets to VM4 using ``ping``
-
-* **Test assertion 5:** Ping from VM1 to VM4 succeeds: ``ping`` exits with return code 0
-
-* VM1 sends ICMP packets to VM5 using ``ping``
-
-* **Test assertion 6:** Ping from VM1 to VM5 succeeds: ``ping`` exits with return code 0
-
-* Delete all instances: VM1, VM2, VM3, VM4 and VM5
-
-* Delete all networks and subnets: networks N1 and N2 including subnets SN1 and SN2
-
-* Delete all network associations and VPN1
-
-
-Pass / fail criteria
-''''''''''''''''''''
-
-This test evaluates the capability of the NFVi and VIM to provide routed IP
-connectivity between VMs by means of BGP/MPLS VPNs. Specifically, the test
-verifies that:
-
-* VMs in the same Neutron subnet have IP connectivity regardless of BGP/MPLS
- VPNs (test assertion 1, 2, 4)
-
-* VMs in different Neutron subnets do not have IP connectivity by default - in
- this case without associating VPNs with the same import and export route
- targets to the Neutron networks (test assertion 3)
-
-* VMs in different Neutron subnets have routed IP connectivity after
- associating both networks with BGP/MPLS VPNs which have been configured with
- the same import and export route targets (test assertion 5, 6). Hence,
- adjusting the ingress and egress route targets enables as well as prohibits
- routing.
-
-In order to pass this test, all test assertions listed in the test execution
-above need to pass.
-
-
-Post conditions
----------------
-
-N/A
-
-------------------------------------------------------------
-Test Case 2 - VPNs ensure traffic separation between tenants
-------------------------------------------------------------
+------------------------------------
+Test Case 1 - Tempest API CRUD Tests
+------------------------------------
Short Name
----------
-dovetail.sdnvpn.tc002.tenant_separation
+functest.tempest.bgpvpn
Use case specification
----------------------
-This test evaluates if VPNs provide separation of traffic such that overlapping
-IP ranges can be used.
+This test case combines multiple CRUD (Create, Read, Update, Delete) tests for
+the objects defined by the BGPVPN API extension of Neutron.
+
+These tests are implemented in the upstream `networking-bgpvpn project repository
+<https://github.com/openstack/networking-bgpvpn>`_ as a Tempest plugin.
Test preconditions
------------------
-2 compute nodes are available, denoted Node1 and Node2 in the following.
+The VIM is operational and the networking-bgpvpn service plugin for Neutron is
+correctly configured and loaded. At least one compute node is available.
Basic test flow execution description and pass/fail criteria
------------------------------------------------------------
-Methodology for verifying connectivity
-''''''''''''''''''''''''''''''''''''''
-
-Connectivity between VMs is tested by establishing an SSH connection. Moreover,
-the command "hostname" is executed at the remote VM in order to retrieve the
-hostname of the remote VM. The retrieved hostname is furthermore compared
-against an expected value. This is used to verify tenant traffic separation,
-i.e., despite overlapping IPs, a connection is made to the correct VM as
-determined by means of the hostname of the target VM.
-
-
-
-Test execution
-''''''''''''''
-
-* Create Neutron network N1
-* Create subnet SN1a of network N1 with IP range 10.10.10.0/24
-* Create subnet SN1b of network N1 with IP range 10.10.11.0/24
-
-* Create Neutron network N2
-* Create subnet SN2a of network N2 with IP range 10.10.10.0/24
-* Create subnet SN2b of network N2 with IP range 10.10.11.0/24
-
-* Create VM1 on Node1 with a port in network N1 and IP 10.10.10.11.
-* Create VM2 on Node1 with a port in network N1 and IP 10.10.10.12.
-* Create VM3 on Node2 with a port in network N1 and IP 10.10.11.13.
-* Create VM4 on Node1 with a port in network N2 and IP 10.10.10.12.
-* Create VM5 on Node2 with a port in network N2 and IP 10.10.11.13.
-
-* Create VPN1 with iRT=eRT=RT1
-* Create network association between network N1 and VPN1
-
-* VM1 attempts to execute the command ``hostname`` on the VM with IP 10.10.10.12 via SSH.
-
-* **Test assertion 1:** VM1 can successfully connect to the VM with IP
- 10.10.10.12. via SSH and execute the remote command ``hostname``. The
- retrieved hostname equals the hostname of VM2.
-
-* VM1 attempts to execute the command ``hostname`` on the VM with IP 10.10.11.13 via SSH.
-
-* **Test assertion 2:** VM1 can successfully connect to the VM with IP
- 10.10.11.13 via SSH and execute the remote command ``hostname``. The
- retrieved hostname equals the hostname of VM3.
-
-* Create VPN2 with iRT=eRT=RT2
-* Create network association between network N2 and VPN2
-
-* VM4 attempts to execute the command ``hostname`` on the VM with IP 10.10.11.13 via SSH.
-
-* **Test assertion 3:** VM4 can successfully connect to the VM with IP
- 10.10.11.13 via SSH and execute the remote command ``hostname``. The
- retrieved hostname equals the hostname of VM5.
-
-* VM4 attempts to execute the command ``hostname`` on the VM with IP 10.10.11.11 via SSH.
-
-* **Test assertion 4:** VM4 cannot connect to the VM with IP 10.10.11.11 via SSH.
+List of test cases
-* Delete all instances: VM1, VM2, VM3, VM4 and VM5
+* networking_bgpvpn_tempest.tests.api.test_create_bgpvpn
+* networking_bgpvpn_tempest.tests.api.test_create_bgpvpn_as_non_admin_fail
+* networking_bgpvpn_tempest.tests.api.test_delete_bgpvpn_as_non_admin_fail
+* networking_bgpvpn_tempest.tests.api.test_show_bgpvpn_as_non_owner_fail
+* networking_bgpvpn_tempest.tests.api.test_list_bgpvpn_as_non_owner_fail
+* networking_bgpvpn_tempest.tests.api.test_show_netassoc_as_non_owner_fail
+* networking_bgpvpn_tempest.tests.api.test_list_netassoc_as_non_owner_fail
+* networking_bgpvpn_tempest.tests.api.test_associate_disassociate_network
+* networking_bgpvpn_tempest.tests.api.test_update_route_target_non_admin_fail
+* networking_bgpvpn_tempest.tests.api.test_create_bgpvpn_with_invalid_routetargets
+* networking_bgpvpn_tempest.tests.api.test_update_bgpvpn_invalid_routetargets
+* networking_bgpvpn_tempest.tests.api.test_associate_invalid_network
+* networking_bgpvpn_tempest.tests.api.test_disassociate_invalid_network
+* networking_bgpvpn_tempest.tests.api.test_associate_disassociate_router
+* networking_bgpvpn_tempest.tests.api.test_attach_associated_subnet_to_associated_router
-* Delete all networks and subnets: networks N1 and N2 including subnets SN1a, SN1b, SN2a and SN2b
-
-* Delete all network associations, VPN1 and VPN2
-
-
-Pass / fail criteria
-''''''''''''''''''''
-
-This test evaluates the capability of the NFVi and VIM to provide routed IP
-connectivity between VMs by means of BGP/MPLS VPNs. Specifically, the test
-verifies that:
-
-* VMs in the same Neutron subnet (still) have IP connectivity between each
- other when a BGP/MPLS VPN is associated with the network (test assertion 1).
-
-* VMs in different Neutron subnets have routed IP connectivity between each
- other when BGP/MPLS VPNs with the same import and expert route targets are
- associated with both networks (assertion 2).
-
-* VMs in different Neutron networks and BGP/MPLS VPNs with different import and
- export route targets can have overlapping IP ranges. The BGP/MPLS VPNs
- provide traffic separation (assertion 3 and 4).
-
-In order to pass this test, all test assertions listed in the test execution
-above need to pass.
-
-
-Post conditions
----------------
-
-N/A
-
---------------------------------------------------------------------------------
-Test Case 3 - VPN provides connectivity between subnets using router association
---------------------------------------------------------------------------------
-
-Short Name
-----------
-
-dovetail.sdnvpn.tc004.router_association
-
-
-Use case specification
-----------------------
-
-This test evaluates if a VPN provides connectivity between two subnets by
-utilizing two different VPN association mechanisms: a router association and a
-network association.
-
-Specifically, the test network topology comprises two networks N1 and N2 with
-corresponding subnets. Additionally, network N1 is connected to a router R1.
-This test verifies that a VPN V1 provides connectivity between both networks
-when applying a router association to router R1 and a network association to
-network N2.
-
-
-Test preconditions
-------------------
-
-2 compute nodes are available, denoted Node1 and Node2 in the following.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-
-Methodology for verifying connectivity
-''''''''''''''''''''''''''''''''''''''
-
-Connectivity between VMs is tested by sending ICMP ping packets between
-selected VMs. The target IPs are passed to the VMs sending pings by means of a
-custom user data script. Whether or not a ping was successful is determined by
-checking the console output of the source VMs.
+The tests include both positive tests and negative tests. The latter are
+identified with the suffix "_fail" in their name.
Test execution
''''''''''''''
-* Create a network N1, a subnet SN1 with IP range 10.10.10.0/24 and a connected router R1
-* Create a network N2, a subnet SN2 with IP range 10.10.11.0/24
-
-* Create VM1 on Node1 with a port in network N1
-* Create VM2 on Node1 with a port in network N1
-* Create VM3 on Node2 with a port in network N1
-* Create VM4 on Node1 with a port in network N2
-* Create VM5 on Node2 with a port in network N2
-
-* Create VPN1 with eRT<>iRT so that connected subnets should not reach each other
-
-* Create route association between router R1 and VPN1
-
-* VM1 sends ICMP packets to VM2 using ``ping``
+The tests are executed sequentially and a separate pass/fail result is recorded
+per test.
-* **Test assertion 1:** Ping from VM1 to VM2 succeeds: ``ping`` exits with return code 0
+In general, every test case performs the API operations indicated in its name
+and asserts that the action succeeds (positive test) or a specific exception
+is triggered (negative test). The following describes the test execution
+per test in further detail.
-* VM1 sends ICMP packets to VM3 using ``ping``
-* **Test assertion 2:** Ping from VM1 to VM3 succeeds: ``ping`` exits with return code 0
+networking_bgpvpn_tempest.tests.api.test_create_bgpvpn
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create a BGPVPN as an admin.
+* **Test assertion**: The API call succeeds.
-* VM1 sends ICMP packets to VM4 using ``ping``
-* **Test assertion 3:** Ping from VM1 to VM4 fails: ``ping`` exits with a non-zero return code
+networking_bgpvpn_tempest.tests.api.test_create_bgpvpn_as_non_admin_fail
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Attempt to create a BGPVPN as non-admin.
+* **Test assertion**: Creating a BGPVPN as non-admin fails.
-* Create network association between network N2 and VPN1
-* VM4 sends ICMP packets to VM5 using ``ping``
+networking_bgpvpn_tempest.tests.api.test_delete_bgpvpn_as_non_admin_fail
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create BGPVPN vpn1 as admin.
+* Attempt to delete vpn1 as non-admin.
+* **Test assertion**: The deletion of vpn1 as non-admin fails.
-* **Test assertion 4:** Ping from VM4 to VM5 succeeds: ``ping`` exits with return code 0
-* Change VPN1 so that iRT=eRT
+networking_bgpvpn_tempest.tests.api.test_show_bgpvpn_as_non_owner_fail
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create a BGPVPN vpn1 as admin in project1.
+* **Test assertion**: Attempting to retrieve detailed properties of vpn1
+ in project2 fails.
-* VM1 sends ICMP packets to VM4 using ``ping``
-* **Test assertion 5:** Ping from VM1 to VM4 succeeds: ``ping`` exits with return code 0
+networking_bgpvpn_tempest.tests.api.test_list_bgpvpn_as_non_owner_fail
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create a BGPVPN vpn1 as admin in project1.
+* Retrieve a list of all BGPVPNs in project2.
+* **Test assertion**: The list of BGPVPNs retrieved in project2 does not
+ include vpn1.
-* VM1 sends ICMP packets to VM5 using ``ping``
-* **Test assertion 6:** Ping from VM1 to VM5 succeeds: ``ping`` exits with return code 0
+networking_bgpvpn_tempest.tests.api.test_show_netassoc_as_non_owner_fail
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create BGPVPN vpn1 as admin in project1.
+* Associate vpn1 with a Neutron network in project1
+* **Test assertion**: Retrieving detailed properties of the network association
+ fails in project2.
-* Delete all instances: VM1, VM2, VM3, VM4 and VM5
-* Delete all networks, subnets and routers: networks N1 and N2 including subnets SN1 and SN2, router R1
+networking_bgpvpn_tempest.tests.api.test_list_netassoc_as_non_owner_fail
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create BGPVPN vpn1 as admin in project1.
+* Create network association net-assoc1 with vpn1 and Neutron network net1
+ in project1.
+* Retrieve a list of all network associations in project2.
+* **Test assertion**: The retrieved list of network associations does not
+ include network association net-assoc1.
-* Delete all network and router associations and VPN1
+networking_bgpvpn_tempest.tests.api.test_associate_disassociate_network
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create a BGPVPN vpn1 as admin.
+* Associate vpn1 with a Neutron network net1.
+* **Test assertion**: The metadata of vpn1 includes the UUID of net1.
+* Diassociate vpn1 from the Neutron network.
+* **Test assertion**: The metadata of vpn1 does not include the UUID of net1.
-Pass / fail criteria
-''''''''''''''''''''
-This test evaluates the capability of the NFVi and VIM to provide routed IP
-connectivity between VMs by means of BGP/MPLS VPNs. Specifically, the test
-verifies that:
+networking_bgpvpn_tempest.tests.api.test_update_route_target_non_admin_fail
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create a BGPVPN vpn1 as admin with specific route targets.
+* Attempt to update vpn1 with different route targets as non-admin.
+* **Test assertion**: The update fails.
-* VMs in the same Neutron subnet have IP connectivity regardless of the import
- and export route target configuration of BGP/MPLS VPNs (test assertion 1, 2, 4)
-* VMs in different Neutron subnets do not have IP connectivity by default - in
- this case without associating VPNs with the same import and export route
- targets to the Neutron networks or connected Neutron routers (test assertion 3).
+networking_bgpvpn_tempest.tests.api.test_create_bgpvpn_with_invalid_routetargets
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Attempt to create a BGPVPN as admin with invalid route targets.
+* **Test assertion**: The creation of the BGPVPN fails.
-* VMs in two different Neutron subnets have routed IP connectivity after
- associating the first network and a router connected to the second network
- with BGP/MPLS VPNs which have been configured with the same import and export
- route targets (test assertion 5, 6). Hence, adjusting the ingress and egress
- route targets enables as well as prohibits routing.
-* Network and router associations are equivalent methods for binding Neutron networks
- to VPN.
+networking_bgpvpn_tempest.tests.api.test_update_bgpvpn_invalid_routetargets
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create a BGPVPN vpn1 as admin with empty route targets.
+* Attempt to update vpn1 with invalid route targets.
+* **Test assertion**: The update of the route targets fails.
-In order to pass this test, all test assertions listed in the test execution
-above need to pass.
+networking_bgpvpn_tempest.tests.api.test_associate_invalid_network
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create BGPVPN vpn1 as admin.
+* Attempt to associate vpn1 with a non-existing Neutron network.
+* **Test assertion**: Creating the network association fails.
-Post conditions
----------------
-N/A
+networking_bgpvpn_tempest.tests.api.test_disassociate_invalid_network
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create BGPVPN vpn1 as admin.
+* Create network association net-assoc1 with vpn1 and Neutron network net1.
+* Attempt to delete net-assoc1 with an invalid network UUID.
+* **Test assertion**: The deletion of the net-assoc fails.
----------------------------------------------------------------------------------------------------
-Test Case 4 - Verify interworking of router and network associations with floating IP functionality
----------------------------------------------------------------------------------------------------
-
-Short Name
-----------
-
-dovetail.sdnvpn.tc008.router_association_floating_ip
+networking_bgpvpn_tempest.tests.api.test_associate_disassociate_router
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create a BGPVPN vpn1 as admin.
+* Associate vpn1 with a Neutron router router1.
+* **Test assertion**: The metadata of vpn1 includes the UUID of router1.
+* Disassociate router1 from vpn1.
+* **Test assertion**: The metadata of vpn1 does not include the UUID of router1.
-Use case specification
-----------------------
-
-This test evaluates if both the router association and network association
-mechanisms interwork with floating IP functionality.
-
-Specifically, the test network topology comprises two networks N1 and N2 with
-corresponding subnets. Additionally, network N1 is connected to a router R1.
-This test verifies that i) a VPN V1 provides connectivity between both networks
-when applying a router association to router R1 and a network association to
-network N2 and ii) a VM in network N1 is reachable externally by means of a
-floating IP.
-
-
-Test preconditions
-------------------
-
-At least one compute node is available.
-
-Basic test flow execution description and pass/fail criteria
-------------------------------------------------------------
-Methodology for verifying connectivity
-''''''''''''''''''''''''''''''''''''''
+networking_bgpvpn_tempest.tests.api.test_attach_associated_subnet_to_associated_router
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+* Create BGPVPN vpn1 as admin.
+* Associate vpn1 with Neutron network net1.
+* Create BGPVPN vpn2
+* Associate vpn2 with Neutron router router1.
+* Attempt to add the subnet of net1 to router1
+* **Test assertion**: The association fails.
-Connectivity between VMs is tested by sending ICMP ping packets between
-selected VMs. The target IPs are passed to the VMs sending pings by means of a
-custom user data script. Whether or not a ping was successful is determined by
-checking the console output of the source VMs.
-
-
-Test execution
-''''''''''''''
-
-* Create a network N1, a subnet SN1 with IP range 10.10.10.0/24 and a connected router R1
-* Create a network N2 with IP range 10.10.20.0/24
-
-* Create VM1 with a port in network N1
-* Create VM2 with a port in network N2
-
-* Create VPN1
-* Create a router association between router R1 and VPN1
-* Create a network association between network N2 and VPN1
-
-
-* VM1 sends ICMP packets to VM2 using ``ping``
-
-* **Test assertion 1:** Ping from VM1 to VM2 succeeds: ``ping`` exits with return code 0
-
-* Assign a floating IP to VM1
-
-* The host running the test framework sends ICMP packets to VM1 using ``ping``
-
-* **Test assertion 2:** Ping from the host running the test framework to the
- floating IP of VM1 succeeds: ``ping`` exits with return code 0
-
-* Delete floating IP assigned to VM1
-
-* Delete all instances: VM1, VM2
-
-* Delete all networks, subnets and routers: networks N1 and N2 including subnets SN1 and SN2, router R1
-
-* Delete all network and router associations as well as VPN1
Pass / fail criteria
''''''''''''''''''''
-This test evaluates the capability of the NFVi and VIM to provide routed IP
-connectivity between VMs by means of BGP/MPLS VPNs. Specifically, the test
-verifies that:
-
-* VMs in the same Neutron subnet have IP connectivity regardless of the import
- and export route target configuration of BGP/MPLS VPNs (test assertion 1)
-
-* VMs connected to a network which has been associated with a BGP/MPLS VPN are
- reachable through floating IPs.
-
-In order to pass this test, all test assertions listed in the test execution
-above need to pass.
+This test validates that all supported CRUD operations (create, read, update,
+delete) can be applied to the objects of the Neutron BGPVPN extension. In
+order to pass this test, all test assertions listed in the test execution above
+need to pass.
Post conditions
diff --git a/docs/testing/user/userguide/api_testing_guide.rst b/docs/testing/user/userguide/api_testing_guide.rst
new file mode 100644
index 00000000..119beff7
--- /dev/null
+++ b/docs/testing/user/userguide/api_testing_guide.rst
@@ -0,0 +1,373 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV, Huawei Technologies Co.,Ltd and others.
+
+===============================
+Running Dovetail by RESTful API
+===============================
+
+Overview
+--------
+
+Dovetail framework provides RESTful APIs for end users to run all OVP test cases.
+Also it provides a Swagger UI for users to find out all APIs and try them out.
+
+
+Definitions and abbreviations
+-----------------------------
+
+- REST - Representational State Transfer
+- API - Application Programming Interface
+- OVP - OPNFV Verification Program
+- UI - User Interface
+
+
+Environment Preparation
+-----------------------
+
+
+Install Docker
+^^^^^^^^^^^^^^
+
+The main prerequisite software for Dovetail is Docker. Please refer to official
+Docker installation guide that is relevant to your Test Host's operating system.
+
+
+Configuring the Test Host Environment
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+For convenience and as a convention, we will create a home directory for storing
+all Dovetail related config items and results files:
+
+.. code-block:: bash
+
+ $ mkdir -p ${HOME}/dovetail
+ $ export DOVETAIL_HOME=${HOME}/dovetail
+
+
+Installing Dovetail API
+-----------------------
+
+The Dovetail project maintains a Docker image that has both Dovetail API and
+Dovetail CLI preinstalled. This Docker image is tagged with versions.
+Before pulling the Dovetail image, check the OPNFV's OVP web page first to
+determine the right tag for OVP testing.
+
+
+Downloading Dovetail Docker Image
+^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+The first version of Dovetail API is ovp-3.0.0.
+
+.. code-block:: bash
+
+ $ sudo docker pull opnfv/dovetail:ovp-3.0.0
+ ovp-3.0.0: Pulling from opnfv/dovetail
+ 6abc03819f3e: Pull complete
+ 05731e63f211: Pull complete
+ 0bd67c50d6be: Pull complete
+ 3f737f5d00b2: Pull complete
+ c93fd0792ebd: Pull complete
+ 77d9a9603ec6: Pull complete
+ 9463cdd9c628: Pull complete
+ Digest: sha256:45e2ffdbe217a4e6723536afb5b6a3785d318deff535da275f34cf8393af458d
+ Status: Downloaded newer image for opnfv/dovetail:ovp-3.0.0
+
+
+Deploying Dovetail API
+^^^^^^^^^^^^^^^^^^^^^^
+
+The Dovetail API can be deployed by running a Dovetail container with the Docker
+image downloaded before.
+
+.. code-block:: bash
+
+ $ docker run -itd -p <swagger_port>:80 -p <api_port>:5000 --privileged=true \
+ -e SWAGGER_HOST=<host_ip>:<api_port> -e DOVETAIL_HOME=/home/ovp \
+ -v /home/ovp:/home/ovp -v /var/run/docker.sock:/var/run/docker.sock \
+ opnfv/dovetail:<version>
+
+
+In the container, it uses 2 ports for Swagger UI (port 80) and API (port 5000)
+respectively. So in order to access to these 2 services outside the container,
+it needs to map them to the host ports. It can be any available ports in the host.
+
+The env SWAGGER_HOST is optional. If you will access the Swagger UI webpage with
+the same host deploying this container, there is no need to set SWAGGER_HOST.
+Otherwise, if you will access the Swagger UI webpage from other machines, then
+it needs to set SWAGGER_HOST.
+
+
+Using Dovetail API
+------------------
+
+Here give the guide of where to find out all APIs and how to use them.
+
+
+Swagger UI Webpage
+^^^^^^^^^^^^^^^^^^
+
+After deploying Dovetail container, the Swagger UI webpage can be accessed with
+any browser. The url is ``http://localhost:<swagger_port>/dovetail-api/index.html``
+if accessing from the same host as deploying this container. Otherwise, the url
+is ``http://<host_ip>:<swagger_port>/dovetail-api/index.html``.
+
+
+Calling APIs
+^^^^^^^^^^^^
+
+There are totally 5 APIs provided by Dovetail.
+
+ * Get all test suites
+
+ * Get all test cases
+
+ * Run test cases
+
+ * Run test cases with execution ID
+
+ * Get status of test cases
+
+Here give some easy guide of how to call these APIs. For more detailed infomation,
+please refer to the Swagger UI page.
+
+
+Getting All Test Suites
+=======================
+
+ * This is a **GET** function with no parameter to get all test suites defined
+ in Dovetail container.
+
+ * The request URL is ``http://<host_ip>:<api_port>/api/v1/scenario/nfvi/testsuites``.
+
+ * The response body is structured as:
+
+ .. code-block:: bash
+
+ {
+ "testsuites": {
+ "debug": {
+ "name": "debug",
+ "testcases_list": {
+ "optional": [
+ "functest.vping.userdata"
+ ]
+ }
+ },
+ "healthcheck": {
+ "name": "healthcheck",
+ "testcases_list": {
+ "optional": [
+ "functest.healthcheck.connection_check"
+ ]
+ }
+ }
+ }
+ }
+
+
+Getting All Test Cases
+======================
+
+ * This is a **GET** function without no parameter to get all test cases integrated
+ in Dovetail container.
+
+ * The request URL is ``http://<host_ip>:<api_port>/api/v1/scenario/nfvi/testcases``.
+
+ * The response body is structured as:
+
+ .. code-block:: bash
+
+ {
+ "testcases": [
+ {
+ "description": "This test case will verify the high availability of the user service provided by OpenStack (keystone) on control node.",
+ "scenario": "nfvi",
+ "subTestCase": null,
+ "testCaseName": "yardstick.ha.keystone"
+ },
+ {
+ "description": "testing for vping using userdata",
+ "scenario": "nfvi",
+ "subTestCase": null,
+ "testCaseName": "functest.vping.userdata"
+ },
+ {
+ "description": "tempest smoke test cases about volume",
+ "scenario": "nfvi",
+ "subTestCase": [
+ "tempest.api.volume.test_volumes_actions.VolumesActionsTest.test_attach_detach_volume_to_instance[compute,id-fff42874-7db5-4487-a8e1-ddda5fb5288d,smoke]",
+ "tempest.scenario.test_volume_boot_pattern.TestVolumeBootPattern.test_volume_boot_pattern[compute,id-557cd2c2-4eb8-4dce-98be-f86765ff311b,image,slow,volume]"
+ ],
+ "testCaseName": "functest.tempest.volume"
+ }
+ ]
+ }
+
+
+Running Test Cases
+==================
+
+ * This is a **POST** function with some parameters to run a subset of the whole test cases.
+
+ * The request URL is ``http://<host_ip>:<api_port>/api/v1/scenario/nfvi/execution``.
+
+ * The request body is structured as following. The ``conf`` section is used to
+ give all configuration items those are required to run test cases. They are
+ the same as all configuration files provided under ``$DOVETAIL_HOME/pre_config/``.
+ If you already have these files under this directory, the whole ``conf`` section
+ can be ignored. If you provide these configuration items with the request body,
+ then the corresponding files under ``$DOVETAIL_HOME/pre_config/`` will be ignored
+ by Dovetail. The ``testcase``, ``testsuite``, ``testarea`` and ``deploy_scenario``
+ correspond to ``--testcase``, ``--testsuite``, ``--testarea`` and ``--deploy-scenario``
+ defined with Dovetail CLI. The ``options`` section support to set all options
+ which have already been implemented by Dovetail CLI including ``--optional``,
+ ``--mandatory``, ``--no-clean``, ``--no-api-validation``, ``--offline``,
+ ``--report``, ``--stop`` and ``--debug``. For options list in ``options`` section,
+ they are set to be ``True``, otherwise, they are set to be ``False``.
+
+ .. code-block:: bash
+
+ {
+ "conf": {
+ "vm_images": "/home/ovp/images",
+ "pods": {
+ "nodes": [
+ {
+ "name": "node1",
+ "role": "Controller",
+ "ip": "192.168.117.222",
+ "user": "root",
+ "password": "root",
+ }
+ ],
+ "process_info": [
+ {
+ "testcase_name": "yardstick.ha.rabbitmq",
+ "attack_host": "node1",
+ "attack_process": "rabbitmq"
+ }
+ ]
+ },
+ "tempest_conf": {
+ "compute": {
+ "min_compute_nodes": "2",
+ "volume_device_name": "vdb",
+ "max_microversion": "2.65"
+ }
+ },
+ "hosts": {
+ "192.168.141.101": [
+ "volume.os.com",
+ "compute.os.com"
+ ]
+ },
+ "envs": {
+ "OS_USERNAME": "admin",
+ "OS_PASSWORD": "admin",
+ "OS_AUTH_URL": "https://192.168.117.222:5000/v3",
+ "EXTERNAL_NETWORK": "ext-net"
+ }
+ },
+ "testcase": [
+ "functest.vping.ssh",
+ "yardstick.ha.rabbitmq"
+ ],
+ "testsuite": "ovp.2019.12",
+ "testarea": [
+ "vping",
+ "ha"
+ ],
+ "deploy_scenario": "os-nosdn-ovs-ha",
+ "options": [
+ "debug",
+ "report"
+ ]
+ }
+
+
+ * The response body is structured as:
+
+ .. code-block:: bash
+
+ {
+ "result": [
+ {
+ "endTime": null,
+ "executionId": "a65e24c0-1803-11ea-84f4-0242ac110004",
+ "results": null,
+ "scenario": "nfvi",
+ "status": "IN_PROGRESS",
+ "testCaseName": "functest.vping.ssh",
+ "testSuiteName": "ovp.2019.12",
+ "timestart": null
+ }
+ ]
+ }
+
+
+Running Test Cases with Execution ID
+====================================
+
+ * This is a **POST** function with some parameters to run a subset of
+ whole test cases and set the execution ID instead of using the random one.
+
+ * The request URL is ``http://<host_ip>:<api_port>/api/v1/scenario/nfvi/execution/{exec_id}``.
+
+ * It's almost the same as the above running test cases API except the execution ID.
+
+
+Getting Status of Test Cases
+============================
+
+ * This is a **POST** function to get the status of some test cases by using
+ the execution ID received in the response body of `Running Test Cases`_ or
+ `Running Test Cases with Execution ID`_ APIs.
+
+ * The request URL is ``http://<host_ip>:<api_port>/api/v1/scenario/nfvi/execution/status/{exec_id}``.
+
+ * The request body is structured as:
+
+ .. code-block:: bash
+
+ {
+ "testcase": [
+ "functest.vping.ssh"
+ ]
+ }
+
+ * The response body is structured as:
+
+ .. code-block:: bash
+
+ {
+ "result": [
+ {
+ "endTime": "2019-12-06 08:39:23",
+ "executionId": "a65e24c0-1803-11ea-84f4-0242ac110004",
+ "results": {
+ "criteria": "PASS",
+ "sub_testcase": [],
+ "timestart": "2019-12-06 08:38:40",
+ "timestop":"2019-12-06 08:39:23"
+ },
+ "scenario": "nfvi",
+ "status": "COMPLETED",
+ "testCaseName": "functest.vping.ssh",
+ "testSuiteName": "ovp.2019.12",
+ "timestart":"2019-12-06 08:38:40"
+ }
+ ]
+ }
+
+
+
+
+Getting Test Results
+^^^^^^^^^^^^^^^^^^^^
+
+Each time you call the running test case API, Dovetail creates a directory with the
+execution ID as the name under ``$DOVETAIL_HOME`` to store results on the host.
+You can find all result files under ``$DOVETAIL_HOME/<executionId>/results``.
+If you run test cases with ``report`` option, then there will be a tarball file
+under ``$DOVETAIL_HOME/<executionId>`` which can be upload to OVP portal.
diff --git a/docs/testing/user/userguide/cli_reference.rst b/docs/testing/user/userguide/cli_reference.rst
index 39e72c93..7dd5c8e4 100644
--- a/docs/testing/user/userguide/cli_reference.rst
+++ b/docs/testing/user/userguide/cli_reference.rst
@@ -51,35 +51,49 @@ Commands List
| dovetail run --help | -h | Show usage of command "dovetail run" |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
-| dovetail run | Run Dovetail with all test areas within default test suite "compliance_set" |
+| dovetail run | Run Dovetail with all test cases within default test suite |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
-| dovetail run --testsuite <test_suite_name> | Run Dovetail with all test areas within test suite <test_suite_name> |
+| dovetail run --testsuite <test_suite_name> | Run Dovetail with all test cases within test suite <test_suite_name> |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
| dovetail run --testsuite <test_suite_name> --testarea <test_area_name> | Run Dovetail with test area <test_area_name> within test suite <test_suite_name>. |
-| | Test area can be chosen from (mandatory, optional, osinterop, ha, vping, ipv6, tempest, sdnvpn). |
+| | Test area can be chosen from (vping, tempest, security, ha, stress, bgpvpn, vnf, snaps). |
| | Repeat option to set multiple test areas. |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
-| dovetail run --debug | -d | Run Dovetail with a debug mode and show all debug logs |
+| dovetail run --testcase <test_case_name> | Run Dovetail with one or more specified test cases. |
+| | Repeat option to set multiple test cases. |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
-| dovetail run --offline | Run Dovetail offline, use local docker images and will not update them |
+| dovetail run --mandatory --testsuite <test_suite_name> | Run Dovetail with all mandatory test cases within test suite <test_suite_name> |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
-| dovetail run --report | -r <db_url> | Push results to local or official DB |
+| dovetail run --optional --testsuite <test_suite_name> | Run Dovetail with all optional test cases within test suite <test_suite_name> |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
-| dovetail run --yardstick_tag | -y <yardstick_docker_image_tag> | Specify yardstick's docker image tag |
+| dovetail run --debug | -d | Run Dovetail with debug mode and show all debug logs |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
-| dovetail run --functest_tag | -f <functest_docker_image_tag> | Specify functest's docker image tag |
+| dovetail run --offline | Run Dovetail offline, use local docker images instead of download online |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
-| dovetail run --bottlenecks_tag | -b <bottlenecks_docker_image_tag> | Specify bottlenecks' docker image tag |
+| dovetail run --report | -r <db_url> | Package the results directory which can be used to upload to OVP web portal |
| | |
+------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
+| dovetail run --deploy-scenario <deploy_scenario_name> | Specify the deploy scenario having as project name 'ovs' |
+| | |
++------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
+| dovetail run --no-api-validation | Disable strict API response validation |
+| | |
++------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
+| dovetail run --no-clean | -n | Keep all Containers created for debuging |
+| | |
++------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
+| dovetail run --stop | -s | Stop immediately when one test case failed |
+| | |
++------------------------------------------------------------------------+---------------------------------------------------------------------------------------------------+
+
Commands Examples
=================
@@ -104,7 +118,7 @@ Dovetail Commands
.. code-block:: bash
root@1f230e719e44:~/dovetail/dovetail# dovetail --version
- dovetail, version 0.7.0
+ dovetail, version 2018.9.0
Dovetail List Commands
----------------------
@@ -121,14 +135,42 @@ Dovetail List Commands
.. code-block:: bash
- root@1f230e719e44:~/dovetail/dovetail# dovetail list debug
- - example
- dovetail.example.tc002
- - osinterop
- dovetail.osinterop.tc001
- - vping
- dovetail.vping.tc001
- dovetail.vping.tc002
+ root@1f230e719e44:~/dovetail/dovetail# dovetail list ovp.2019.12
+ - mandatory
+ functest.vping.userdata
+ functest.vping.ssh
+ functest.tempest.osinterop
+ functest.tempest.compute
+ functest.tempest.identity_v3
+ functest.tempest.image
+ functest.tempest.network_api
+ functest.tempest.volume
+ functest.tempest.neutron_trunk_ports
+ functest.tempest.ipv6_api
+ functest.security.patrole
+ yardstick.ha.nova_api
+ yardstick.ha.neutron_server
+ yardstick.ha.keystone
+ yardstick.ha.glance_api
+ yardstick.ha.cinder_api
+ yardstick.ha.cpu_load
+ yardstick.ha.disk_load
+ yardstick.ha.haproxy
+ yardstick.ha.rabbitmq
+ yardstick.ha.database
+ bottlenecks.stress.ping
+ - optional
+ functest.tempest.ipv6_scenario
+ functest.tempest.multi_node_scheduling
+ functest.tempest.network_security
+ functest.tempest.vm_lifecycle
+ functest.tempest.network_scenario
+ functest.tempest.bgpvpn
+ functest.security.patrole_vxlan_dependent
+ yardstick.ha.neutron_l3_agent
+ yardstick.ha.controller_restart
+ functest.vnf.vims
+ functest.vnf.vepc
Dovetail Show Commands
----------------------
@@ -145,37 +187,54 @@ Dovetail Show Commands
.. code-block:: bash
- root@1f230e719e44:~/dovetail/dovetail# dovetail show dovetail.vping.tc001
+ root@1f230e719e44:~/dovetail/dovetail# dovetail show functest.vping.ssh
---
- dovetail.vping.tc001:
- name: dovetail.vping.tc001
- objective: testing for vping using userdata
+ functest.vping.ssh:
+ name: functest.vping.ssh
+ objective: testing for vping using ssh
validate:
type: functest
- testcase: vping_userdata
+ testcase: vping_ssh
+ image_name: opnfv/functest-healthcheck
report:
+ source_archive_files:
+ - functest.log
+ dest_archive_files:
+ - vping_logs/functest.vping.ssh.log
+ check_results_file:
+ - 'functest_results.txt'
+ portal_key_file: vping_logs/functest.vping.ssh.log
sub_testcase_list:
.. code-block:: bash
- root@1f230e719e44:~/dovetail/dovetail# dovetail show ipv6.tc001
+ root@1f230e719e44:~/dovetail/dovetail# dovetail show functest.tempest.image
---
- dovetail.ipv6.tc001:
- name: dovetail.ipv6.tc001
- objective: Bulk creation and deletion of IPv6 networks, ports and subnets
+ functest.tempest.image:
+ name: functest.tempest.image
+ objective: tempest smoke test cases about image
validate:
type: functest
testcase: tempest_custom
pre_condition:
- - 'cp /home/opnfv/userconfig/pre_config/tempest_conf.yaml /usr/local/lib/python2.7/dist-packages/functest/opnfv_tests/openstack/tempest/custom_tests/tempest_conf.yaml'
- pre_copy:
- src_file: tempest_custom.txt
- dest_path: /usr/local/lib/python2.7/dist-packages/functest/opnfv_tests/openstack/tempest/custom_tests/test_list.txt
+ - 'cp /home/opnfv/userconfig/pre_config/tempest_conf.yaml /usr/lib/python2.7/site-packages/functest/opnfv_tests/openstack/tempest/custom_tests/tempest_conf.yaml'
+ - 'cp /home/opnfv/userconfig/tempest_custom_testcases.yaml /usr/lib/python2.7/site-packages/xtesting/ci/testcases.yaml'
+ - 'cp /home/opnfv/functest/results/tempest_custom.txt /usr/lib/python2.7/site-packages/functest/opnfv_tests/openstack/tempest/custom_tests/test_list.txt'
report:
+ source_archive_files:
+ - functest.log
+ - tempest_custom/rally.log
+ - tempest_custom/tempest-report.html
+ dest_archive_files:
+ - tempest_logs/functest.tempest.image.functest.log
+ - tempest_logs/functest.tempest.image.log
+ - tempest_logs/functest.tempest.image.html
+ check_results_file:
+ - 'functest_results.txt'
+ portal_key_file: tempest_logs/functest.tempest.image.html
sub_testcase_list:
- - tempest.api.network.test_networks.BulkNetworkOpsIpV6Test.test_bulk_create_delete_network[id-d4f9024d-1e28-4fc1-a6b1-25dbc6fa11e2,smoke]
- - tempest.api.network.test_networks.BulkNetworkOpsIpV6Test.test_bulk_create_delete_port[id-48037ff2-e889-4c3b-b86a-8e3f34d2d060,smoke]
- - tempest.api.network.test_networks.BulkNetworkOpsIpV6Test.test_bulk_create_delete_subnet[id-8936533b-c0aa-4f29-8e53-6cc873aec489,smoke]
+ - tempest.api.image.v2.test_images.BasicOperationsImagesTest.test_register_upload_get_image_file[id-139b765e-7f3d-4b3d-8b37-3ca3876ee318,smoke]
+ - tempest.api.image.v2.test_versions.VersionsTest.test_list_versions[id-659ea30a-a17c-4317-832c-0f68ed23c31d,smoke]
Dovetail Run Commands
----------------------
@@ -188,25 +247,43 @@ Dovetail Run Commands
Dovetail compliance test entry!
Options:
- -b, --bott_tag TEXT Overwrite tag for bottlenecks docker container (e.g. cvp.0.4.0)
- -f, --func_tag TEXT Overwrite tag for functest docker container (e.g. cvp.0.5.0)
- -y, --yard_tag TEXT Overwrite tag for yardstick docker container (e.g. danube.3.2)
- --testarea TEXT compliance testarea within testsuite
- --offline run in offline method, which means not to update the docker upstream images, functest, yardstick, etc.
- -r, --report TEXT push results to DB (e.g. --report http://192.168.135.2:8000/api/v1/results)
- --testsuite TEXT compliance testsuite.
- -d, --debug Flag for showing debug log on screen.
- -h, --help Show this message and exit.
+ --opnfv-ci Only enabled when running with OPNFV CI jobs and pushing results to TestAPI DB
+ --optional Run all optional test cases.
+ --mandatory Run all mandatory test cases.
+ --deploy-scenario TEXT Specify the DEPLOY_SCENARIO which will be used as input by each testcase respectively
+ -n, --no-clean Keep all Containers created for debuging.
+ --no-api-validation disable strict API response validation
+ --offline run in offline method, which means not to update the docker upstream images, functest, yardstick, etc.
+ -r, --report Create a tarball file to upload to OVP web portal
+ -s, --stop Flag for stopping on test case failure.
+ -d, --debug Flag for showing debug log on screen.
+ --testcase TEXT Compliance testcase. Specify option multiple times to include multiple test cases.
+ --testarea TEXT Compliance testarea within testsuite. Specify option multiple times to include multiple test areas.
+ --testsuite TEXT compliance testsuite.
+ -h, --help Show this message and exit.
.. code-block:: bash
- root@1f230e719e44:~/dovetail/dovetail# dovetail run --testsuite proposed_tests --testarea vping --offline -r http://192.168.135.2:8000/api/v1/results
- 2017-10-12 14:57:51,278 - run - INFO - ================================================
- 2017-10-12 14:57:51,278 - run - INFO - Dovetail compliance: proposed_tests!
- 2017-10-12 14:57:51,278 - run - INFO - ================================================
- 2017-10-12 14:57:51,278 - run - INFO - Build tag: daily-master-b80bca76-af5d-11e7-879a-0242ac110002
- 2017-10-12 14:57:51,336 - run - WARNING - There is no hosts file /home/jenkins/opnfv/slave_root/workspace/dovetail-compass-huawei-pod7-proposed_tests-danube/cvp/pre_config/hosts.yaml, may be some issues with domain name resolution.
- 2017-10-12 14:57:51,517 - run - INFO - >>[testcase]: dovetail.vping.tc001
- 2017-10-12 14:58:21,325 - run - INFO - Results have been pushed to database and stored with local file /home/dovetail/results/results.json.
- 2017-10-12 14:58:21,337 - run - INFO - >>[testcase]: dovetail.vping.tc002
- 2017-10-12 14:58:48,862 - run - INFO - Results have been pushed to database and stored with local file /home/dovetail/results/results.json.
+ root@1f230e719e44:~/dovetail/dovetail# dovetail run --testcase functest.vping.ssh --offline -r --deploy-scenario os-nosdn-ovs-ha
+ 2019-12-06 02:51:52,634 - run - INFO - ================================================
+ 2019-12-06 02:51:52,634 - run - INFO - Dovetail compliance: ovp.2019.12!
+ 2019-12-06 02:51:52,634 - run - INFO - ================================================
+ 2019-12-06 02:51:52,634 - run - INFO - Build tag: daily-master-5b58584a-17d3-11ea-878a-0242ac110002
+ 2019-12-06 02:51:52,634 - run - INFO - DEPLOY_SCENARIO : os-nosdn-ovs-ha
+ 2019-12-06 02:51:53,077 - run - INFO - >>[testcase]: functest.vping.ssh
+ 2019-12-06 02:51:53,078 - dovetail.test_runner.DockerRunner - WARNING - There is no hosts file /home/ovp/pre_config/hosts.yaml. This may cause some issues with domain name resolution.
+ 2019-12-06 02:51:54,048 - dovetail.test_runner.DockerRunner - INFO - Get hardware info of all nodes list in file /home/ovp/pre_config/pod.yaml ...
+ 2019-12-06 02:51:54,049 - dovetail.test_runner.DockerRunner - INFO - Hardware info of all nodes are stored in file /home/dovetail/results/all_hosts_info.json.
+ 2019-12-06 02:51:54,073 - dovetail.container.Container - WARNING - There is no hosts file /home/ovp/pre_config/hosts.yaml. This may cause some issues with domain name resolution.
+ 2019-12-06 02:52:57,982 - dovetail.report.Report - INFO - Results have been stored with files: ['/home/ovp/results/functest_results.txt'].
+ 2019-12-06 02:52:57,986 - dovetail.report.Report - INFO -
+
+ Dovetail Report
+ Version: 2019.12
+ Build Tag: daily-master-5b58584a-17d3-11ea-878a-0242ac110002
+ Test Date: 2019-12-06 02:52:57 UTC
+ Duration: 64.91 s
+
+ Pass Rate: 100.00% (1/1)
+ vping: pass rate 100.00%
+ -functest.vping.ssh PASS
diff --git a/docs/testing/user/userguide/images/tocsa_vnf_test_environment.png b/docs/testing/user/userguide/images/tocsa_vnf_test_environment.png
new file mode 100644
index 00000000..78b3f74a
--- /dev/null
+++ b/docs/testing/user/userguide/images/tocsa_vnf_test_environment.png
Binary files differ
diff --git a/docs/testing/user/userguide/images/tosca_vnf_test_flow.png b/docs/testing/user/userguide/images/tosca_vnf_test_flow.png
new file mode 100644
index 00000000..87dc8ec4
--- /dev/null
+++ b/docs/testing/user/userguide/images/tosca_vnf_test_flow.png
Binary files differ
diff --git a/docs/testing/user/userguide/index.rst b/docs/testing/user/userguide/index.rst
index 70f78b5c..98ca56e0 100644
--- a/docs/testing/user/userguide/index.rst
+++ b/docs/testing/user/userguide/index.rst
@@ -4,12 +4,14 @@
.. _dovetail-testing_user_guide:
-********************************************************
-OPNFV Verified Program Testing User Guide
-********************************************************
+======================
+OVP Testing User Guide
+======================
.. toctree::
:maxdepth: 2
testing_guide.rst
cli_reference.rst
+ api_testing_guide.rst
+ vnf_test_guide.rst
diff --git a/docs/testing/user/userguide/testing_guide.rst b/docs/testing/user/userguide/testing_guide.rst
index 54fa294b..d1c31683 100644
--- a/docs/testing/user/userguide/testing_guide.rst
+++ b/docs/testing/user/userguide/testing_guide.rst
@@ -2,16 +2,19 @@
.. http://creativecommons.org/licenses/by/4.0
.. (c) OPNFV, Huawei Technologies Co.,Ltd and others.
-==========================================
-Conducting OVP Testing with Dovetail
-==========================================
+=========================================
+Conducting OVP NFVI Testing with Dovetail
+=========================================
Overview
-------------------------------
+--------
-The Dovetail testing framework for OVP consists of two major parts: the testing client that
+This guide provides the instructions for the OVP Infrastructure testing. For the
+OVP VNF testing, please refer to the next section.
+
+The Dovetail testing framework for OVP consists of two major parts: the testing client which
executes all test cases in a lab (vendor self-testing or a third party lab),
-and the server system that is hosted by the OVP administrator to store and
+and the server system which is hosted by the OVP administrator to store and
view test results based on a web API. The following diagram illustrates
this overall framework.
@@ -25,10 +28,10 @@ the System Under Test (SUT) itself.
The above diagram assumes that the tester's Test Host is situated in a DMZ, which
has internal network access to the SUT and external access via the public Internet.
The public Internet connection allows for easy installation of the Dovetail containers.
-A singular compressed file that includes all the underlying results can be pulled from
+A single compressed file that includes all the underlying results can be pulled from
the Test Host and uploaded to the OPNFV OVP server.
This arrangement may not be supported in some labs. Dovetail also supports an offline mode of
-installation that is illustrated in the next diagram.
+installation which is illustrated in the next diagram.
.. image:: ../../../images/dovetail_offline_mode.png
:align: center
@@ -44,8 +47,7 @@ The rest of this guide will describe how to install the Dovetail tool as a
Docker container image, go over the steps of running the OVP test suite, and
then discuss how to view test results and make sense of them.
-Readers interested
-in using Dovetail for its functionalities beyond OVP testing, e.g. for in-house
+Readers interested in using Dovetail for its functionalities beyond OVP testing, e.g. for in-house
or extended testing, should consult the Dovetail developer's guide for additional
information.
@@ -107,23 +109,7 @@ Or, if the lab environment does not allow ping, try validating it using HTTPS in
Installing Prerequisite Packages on the Test Host
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
-The main prerequisite software for Dovetail are Python and Docker.
-
-In the OVP test suite for the Danube release, Dovetail requires Python 2.7. Various minor
-versions of Python 2.7.x are known to work Dovetail, but there are no assurances. Python 3.x
-is not supported at this time.
-
-Use the following steps to check if the right version of python is already installed,
-and if not, install it.
-
-.. code-block:: bash
-
- $ python --version
- Python 2.7.6
-
-If your Test Host does not have Python installed, or the version is not 2.7, you
-should consult Python installation guides corresponding to the operating system
-in your Test Host on how to install Python 2.7.
+The main prerequisite software for Dovetail is Docker.
Dovetail does not work with Docker versions prior to 1.12.3. We have validated
Dovetail with Docker 17.03 CE. Other versions of Docker later than 1.12.3 may
@@ -152,20 +138,13 @@ also work, but community support may be more available on Docker 17.03 CE or gre
If your Test Host does not have Docker installed, or Docker is older than 1.12.3,
or you have Docker version other than 17.03 CE and wish to change,
you will need to install, upgrade, or re-install in order to run Dovetail.
-The Docker installation process
-can be more complex, you should refer to the official
+If you need further assistance with Docker installation process, you should refer to the official
Docker installation guide that is relevant to your Test Host's operating system.
-The above installation steps assume that the Test Host is in the online mode. For offline
-testing, use the following offline installation steps instead.
-
-In order to install or upgrade Python offline, you may download packaged Python 2.7
-for your Test Host's operating system on a connected host, copy the packge to
-the Test Host, then install from that local copy.
-
-In order to install Docker offline, download Docker static binaries and copy the
-tar file to the Test Host, such as for Ubuntu14.04, you may follow the following link
-to install,
+The above installation steps assume that the Test Host is in the online mode.
+For offline testing, use the following offline installation steps instead.
+For instance, download Docker static binaries and copy the tar file to the
+Test Host, such as for Ubuntu14.04, you may follow the following link:
.. code-block:: bash
@@ -175,31 +154,33 @@ Configuring the Test Host Environment
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
The Test Host needs a few environment variables set correctly in order to access the
-Openstack API required to drive the Dovetail tests. For convenience and as a convention,
+OpenStack API which is required to drive the Dovetail tests. For convenience and as a convention,
we will also create a home directory for storing all Dovetail related config files and
results files:
.. code-block:: bash
- $ mkdir -p /home/dovetail
- $ export DOVETAIL_HOME=/home/dovetail
+ $ mkdir -p ${HOME}/dovetail
+ $ export DOVETAIL_HOME=${HOME}/dovetail
-Here we set dovetail home directory to be ``/home/dovetail`` for an example.
-Then create a directory named ``pre_config`` in this directory to store all
-Dovetail related config files:
+For example, here we set dovetail home directory to be ``${HOME}/dovetail``.
+Then create two directories named ``pre_config`` and ``images`` under this directory
+to store all Dovetail related config files and all test images respectively:
.. code-block:: bash
$ mkdir -p ${DOVETAIL_HOME}/pre_config
+ $ mkdir -p ${DOVETAIL_HOME}/images
+
Setting up Primary Configuration File
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
-At this point, you will need to consult your SUT (Openstack) administrator to correctly set
+At this point, you will need to consult your SUT (OpenStack) administrator to correctly set
the configurations in a file named ``env_config.sh``.
-The Openstack settings need to be configured such that the Dovetail client has all the necessary
+The OpenStack settings need to be configured such that the Dovetail client has all the necessary
credentials and privileges to execute all test operations. If the SUT uses terms
-somewhat differently from the standard Openstack naming, you will need to adjust
+somewhat differently from the standard OpenStack naming, you will need to adjust
this file accordingly.
Create and edit the file ``${DOVETAIL_HOME}/pre_config/env_config.sh`` so that
@@ -210,20 +191,17 @@ this file should contain.
$ cat ${DOVETAIL_HOME}/pre_config/env_config.sh
- # Project-level authentication scope (name or ID), recommend admin project.
+ # Project-level authentication scope (name or ID), admin project is recommended.
export OS_PROJECT_NAME=admin
- # For identity v2, it uses OS_TENANT_NAME rather than OS_PROJECT_NAME.
- export OS_TENANT_NAME=admin
-
- # Authentication username, belongs to the project above, recommend admin user.
+ # Authentication username, belongs to the project above, admin user is recommended.
export OS_USERNAME=admin
# Authentication password. Use your own password
export OS_PASSWORD=xxxxxxxx
# Authentication URL, one of the endpoints of keystone service. If this is v3 version,
- # there need some extra variables as follows.
+ # there needs some extra variables as follows.
export OS_AUTH_URL='http://xxx.xxx.xxx.xxx:5000/v3'
# Default is 2.0. If use keystone v3 API, this should be set as 3.
@@ -245,34 +223,45 @@ this file should contain.
# If using https + no cacert, should add OS_INSECURE environment parameter.
export OS_INSECURE=True
+ # The name of a network with external connectivity for allocating floating
+ # IPs. It is required that at least one Neutron network with the attribute
+ # 'router:external=True' is pre-configured on the system under test.
+ # This network is used by test cases to SSH into tenant VMs and perform
+ # operations there.
+ export EXTERNAL_NETWORK=xxx
-The OS_AUTH_URL variable is key to configure correctly, as the other admin services
-are gleaned from the identity service. HTTPS should be configured in the SUT so the
-final two variables should be uncommented. However, if SSL is disabled in the SUT, comment
-out the OS_CACERT and OS_INSECURE variables. Ensure the '/path/to/pre_config' directory in
-the above file matches the directory location of the cacert file for the OS_CACERT variable.
+ # Set an existing role used to create project and user for vping test cases.
+ # Otherwise, it will create a role 'Member' to do that.
+ export NEW_USER_ROLE=xxx
-Export all these variables into environment by,
+ # For XCI installer the following environment parameters should be added in
+ # this file. Otherwise, those parameters could be ignored.
+ export INSTALLER_TYPE=osa
+ export DEPLOY_SCENARIO=os-nosdn-nofeature
+ export XCI_FLAVOR=noha
-.. code-block:: bash
- $ source ${DOVETAIL_HOME}/pre_config/env_config.sh
-The above line may be added to your .bashrc file for convenience when repeatedly using
-Dovetail.
+The OS_AUTH_URL variable is key to configure correctly, as the other admin services
+are collected from the identity service. HTTPS should be configured in the SUT so
+either OS_CACERT or OS_INSECURE should be uncommented.
+However, if SSL is disabled in the SUT, comment out both OS_CACERT and OS_INSECURE variables.
+Ensure the '/path/to/pre_config' directory in
+the above file matches the directory location of the cacert file for the OS_CACERT variable.
The next three sections outline additional configuration files used by Dovetail. The
-tempest (tempest_conf.yaml) configuration file is required for executing the mandatory
-osinterop test cases and the optional ipv6/tempest test cases. The HA (pod.yaml) configuration
-file is required for the mandatory HA test cases and is also employed to collect SUT hardware
+tempest (tempest_conf.yaml) configuration file is required for executing all tempest
+test cases (e.g. functest.tempest.compute, functest.tempest.ipv6 ...) and
+functest.security.patrole. The HA (pod.yaml) configuration
+file is required for HA test cases and is also employed to collect SUT hardware
info. The hosts.yaml is optional for hostname/IP resolution.
Configuration for Running Tempest Test Cases (Mandatory)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
-The test cases in the test areas `osinterop` (OpenStack Interoperability tests),
-`ipv6` and `tempest` are based on Tempest. A SUT-specific configuration of
-Tempest is required in order to run those test cases successfully. The
+The test cases in the test areas `tempest` and `security`
+are based on Tempest. A SUT-specific configuration of
+Tempest is required in order to run those test cases successfully. The
corresponding SUT-specific configuration options must be supplied in the file
``$DOVETAIL_HOME/pre_config/tempest_conf.yaml``.
@@ -289,15 +278,34 @@ Here is an example of what this file should contain.
# Expected device name when a volume is attached to an instance.
volume_device_name: vdb
-Use the listing above at a minimum to execute the mandatory test areas.
+ # One sub test case of functest.tempest.osinterop will be skipped if not provide this version.
+ # The default range of microversion for tempest is [None - None].
+ # Test case functest.tempest.osinterop required the range to be [2.2 - latest].
+ max_microversion: 2.65
+
+Use the listing above as a minimum to execute the mandatory test cases.
+
+If the optional BGPVPN Tempest API tests shall be run, Tempest needs to be told
+that the BGPVPN service is available. To do that, add the following to the
+``$DOVETAIL_HOME/pre_config/tempest_conf.yaml`` configuration file:
+
+.. code-block:: bash
+
+ service_available:
+ bgpvpn: True
+
Configuration for Running HA Test Cases (Mandatory)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
-The mandatory HA test cases require OpenStack controller node info. It must include the node's
+The HA test cases require OpenStack controller node info. It must include the node's
name, role, ip, as well as the user and key_filename or password to login to the node. Users
-must create the file ``${DOVETAIL_HOME}/pre_config/pod.yaml`` to store the info. This file is
-also used as basis to collect SUT hardware information that is stored alongside results and
+must create the file ``${DOVETAIL_HOME}/pre_config/pod.yaml`` to store the info.
+For some HA test cases, they will log in the controller node 'node1' and kill the specific processes.
+The names of the specific processes may be different with the actual ones of the SUTs.
+The processes' names can also be changed with file ``${DOVETAIL_HOME}/pre_config/pod.yaml``.
+
+This file is also used as a basis to collect SUT hardware information which is stored alongside results and
uploaded to the OVP web portal. The SUT hardware information can be viewed within the
'My Results' view in the OVP web portal by clicking the SUT column 'info' link. In order to
collect SUT hardware information holistically, ensure this file has an entry for each of
@@ -309,21 +317,62 @@ Below is a sample with the required syntax when password is employed by the cont
nodes:
-
- # This can not be changed and must be node1.
+ # This info of node0 is used only for one optional test case 'yardstick.ha.controller_restart'.
+ # If you don't plan to test it, this Jumpserver node can be ignored.
+ # This can not be changed and **must** be node0.
+ name: node0
+
+ # This **must** be Jumpserver.
+ role: Jumpserver
+
+ # This is the instance IP of a node which has ipmitool installed.
+ ip: xx.xx.xx.xx
+
+ # User name of the user of this node. This user **must** have sudo privileges.
+ user: root
+
+ # Password of the user.
+ password: root
+
+ -
+ # Almost all HA test cases are trying to login to a controller node named 'node1'
+ # and then kill some processes running on it.
+ # If you don't want to reset the attack node name for each test case, this
+ # name can not be changed and **must** be node1.
name: node1
- # This must be controller.
+ # This **must** be controller.
role: Controller
- # This is the install IP of a controller node, which is the haproxy primary node
+ # This is the instance IP of a controller node, which is the haproxy primary node
ip: xx.xx.xx.xx
- # User name of this node. This user must have sudo privileges.
+ # User name of the user of this node. This user **must** have sudo privileges.
user: root
# Password of the user.
password: root
+ process_info:
+ -
+ # For all HA test cases, there are 2 parameters, 'attack_process' and 'attack_host',
+ # which support to be set by users instead of using the default values.
+ # The 'attack_process' is the process name of one HA test case which it try to kill.
+ # The 'attack_host' is the host name which the test case try to login and then kill
+ # the process running on it.
+ # Fllowing is 2 samples.
+
+ # The default attack process of yardstick.ha.rabbitmq is 'rabbitmq-server'.
+ # Here can be reset to 'rabbitmq'.
+ testcase_name: yardstick.ha.rabbitmq
+ attack_process: rabbitmq
+
+ -
+ # The default attack host for all HA test cases is 'node1'.
+ # Here can be reset to any other node given in the section 'nodes'.
+ testcase_name: yardstick.ha.glance_api
+ attack_host: node2
+
Besides the 'password', a 'key_filename' entry can be provided to login to the controller node.
Users need to create file ``$DOVETAIL_HOME/pre_config/id_rsa`` to store the private key.
A sample is provided below to show the required syntax when using a key file.
@@ -337,15 +386,43 @@ A sample is provided below to show the required syntax when using a key file.
ip: 10.1.0.50
user: root
- # Private key of this node. It must be /root/.ssh/id_rsa
- # Dovetail will move the key file from $DOVETAIL_HOME/pre_config/id_rsa
- # to /root/.ssh/id_rsa of Yardstick container
- key_filename: /root/.ssh/id_rsa
+ # Private ssh key for accessing the controller nodes. If a keyfile is
+ # being used instead of password, it **must** be put under
+ # $DOVETAIL_HOME/pre_config/ and named 'id_rsa'.
+ key_filename: /home/dovetail/pre_config/id_rsa
Under nodes, repeat entries for name, role, ip, user and password or key file for each of the
controller/compute nodes that comprise the SUT. Use a '-' to separate each of the entries.
Specify the value for the role key to be either 'Controller' or 'Compute' for each node.
+Under process_info, repeat entries for testcase_name, attack_host and attack_process
+for each HA test case. Use a '-' to separate each of the entries.
+The default attack host of all HA test cases is **node1**.
+The default attack processes of all HA test cases are list here:
+
+ +-------------------------------+-------------------------+
+ | Test Case Name | Attack Process Name |
+ +===============================+=========================+
+ | yardstick.ha.cinder_api | cinder-api |
+ +-------------------------------+-------------------------+
+ | yardstick.ha.database | mysql |
+ +-------------------------------+-------------------------+
+ | yardstick.ha.glance_api | glance-api |
+ +-------------------------------+-------------------------+
+ | yardstick.ha.haproxy | haproxy |
+ +-------------------------------+-------------------------+
+ | yardstick.ha.keystone | keystone |
+ +-------------------------------+-------------------------+
+ | yardstick.ha.neutron_l3_agent | neutron-l3-agent |
+ +-------------------------------+-------------------------+
+ | yardstick.ha.neutron_server | neutron-server |
+ +-------------------------------+-------------------------+
+ | yardstick.ha.nova_api | nova-api |
+ +-------------------------------+-------------------------+
+ | yardstick.ha.rabbitmq | rabbitmq-server |
+ +-------------------------------+-------------------------+
+
+
Configuration of Hosts File (Optional)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
@@ -353,7 +430,7 @@ If your SUT uses a hosts file to translate hostnames into the IP of OS_AUTH_URL,
to provide the hosts info in a file ``$DOVETAIL_HOME/pre_config/hosts.yaml``.
Create and edit file ``$DOVETAIL_HOME/pre_config/hosts.yaml``. Below is an example of what
-this file should contain. Note, that multiple hostnames can be specified for each IP address,
+this file should contain. Note that multiple hostnames can be specified for each IP address,
as shown in the generic syntax below the example.
.. code-block:: bash
@@ -363,7 +440,8 @@ as shown in the generic syntax below the example.
---
hosts_info:
192.168.141.101:
- - ha-vip
+ - identity.endpoint.url
+ - compute.endpoint.url
<ip>:
- <hostname1>
@@ -380,64 +458,79 @@ OPNFV's OVP web page first to determine the right tag for OVP testing.
Online Test Host
""""""""""""""""
-If the Test Host is online, you can directly pull Dovetail Docker image and download Ubuntu
-and Cirros images. All other dependent docker images will automatically be downloaded. The
-Ubuntu and Cirros images are used by Dovetail for image creation and VM instantiation within
-the SUT.
+If the Test Host is online, you can directly pull Dovetail Docker image, then all
+other dependent docker images will automatically be downloaded. Also you can download
+other related VM images such as Ubuntu and Cirros images which are used by Dovetail
+for image creation and VM instantiation within the SUT.
+
+Following given the download url for each VM images. Cirros-0.4.0 and Ubuntu-16.04
+are used by mandatory test cases, so they are the only 2 images **must** be downloaded
+before doing the test. There are also 2 other optional VM images, Ubuntu-14.04 and
+Cloudify-manager, which are used by optional test cases functest.vnf.vepc and functest.vnf.vims.
+If you don't plan to test these 2 test cases, you can skip downloading these 2 images.
.. code-block:: bash
- $ wget -nc http://artifacts.opnfv.org/sdnvpn/ubuntu-16.04-server-cloudimg-amd64-disk1.img -P ${DOVETAIL_HOME}/pre_config
- $ wget -nc http://download.cirros-cloud.net/0.3.5/cirros-0.3.5-x86_64-disk.img -P ${DOVETAIL_HOME}/pre_config
-
- $ sudo docker pull opnfv/dovetail:ovp.1.0.0
- ovp.1.0.0: Pulling from opnfv/dovetail
- 30d541b48fc0: Pull complete
- 8ecd7f80d390: Pull complete
- 46ec9927bb81: Pull complete
- 2e67a4d67b44: Pull complete
- 7d9dd9155488: Pull complete
- cc79be29f08e: Pull complete
- e102eed9bf6a: Pull complete
- 952b8a9d2150: Pull complete
- bfbb639d1f38: Pull complete
- bf7c644692de: Pull complete
- cdc345e3f363: Pull complete
- Digest: sha256:d571b1073b2fdada79562e8cc67f63018e8d89268ff7faabee3380202c05edee
- Status: Downloaded newer image for opnfv/dovetail:ovp.1.0.0
-
-An example of the <tag> is *ovp.1.0.0*.
+ $ wget -nc http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img -P ${DOVETAIL_HOME}/images
+ $ wget -nc https://cloud-images.ubuntu.com/releases/16.04/release/ubuntu-16.04-server-cloudimg-amd64-disk1.img -P ${DOVETAIL_HOME}/images
+ $ wget -nc https://cloud-images.ubuntu.com/releases/14.04/release/ubuntu-14.04-server-cloudimg-amd64-disk1.img -P ${DOVETAIL_HOME}/images
+ $ wget -nc http://repository.cloudifysource.org/cloudify/19.01.24/community-release/cloudify-docker-manager-community-19.01.24.tar -P ${DOVETAIL_HOME}/images
+
+ $ sudo docker pull opnfv/dovetail:ovp-3.0.0
+ ovp-3.0.0: Pulling from opnfv/dovetail
+ 324d088ce065: Pull complete
+ 2ab951b6c615: Pull complete
+ 9b01635313e2: Pull complete
+ 04510b914a6c: Pull complete
+ 83ab617df7b4: Pull complete
+ 40ebbe7294ae: Pull complete
+ d5db7e3e81ae: Pull complete
+ 0701bf048879: Pull complete
+ 0ad9f4168266: Pull complete
+ d949894f87f6: Pull complete
+ Digest: sha256:7449601108ebc5c40f76a5cd9065ca5e18053be643a0eeac778f537719336c29
+ Status: Downloaded newer image for opnfv/dovetail:ovp-3.0.0
Offline Test Host
"""""""""""""""""
-If the Test Host is offline, you will need to first pull the Dovetail Docker image, and all the
+If the Test Host is offline, you will need to first pull the Dovetail Docker image and all the
dependent images that Dovetail uses, to a host that is online. The reason that you need
to pull all dependent images is because Dovetail normally does dependency checking at run-time
and automatically pulls images as needed, if the Test Host is online. If the Test Host is
offline, then all these dependencies will need to be manually copied.
+The Docker images, Ubuntu and Cirros image below are necessary for all mandatory test cases.
+
+.. code-block:: bash
+
+ $ sudo docker pull opnfv/dovetail:ovp-3.0.0
+ $ sudo docker pull opnfv/functest-smoke:hunter
+ $ sudo docker pull opnfv/functest-healthcheck:hunter
+ $ sudo docker pull opnfv/yardstick:opnfv-8.0.0
+ $ sudo docker pull opnfv/bottlenecks:8.0.1-latest
+ $ wget -nc http://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img -P {ANY_DIR}
+ $ wget -nc https://cloud-images.ubuntu.com/releases/16.04/release/ubuntu-16.04-server-cloudimg-amd64-disk1.img -P ${DOVETAIL_HOME}/images
+
+The other Docker images and test images below are only used by optional test cases.
+
.. code-block:: bash
- $ sudo docker pull opnfv/dovetail:ovp.1.0.0
- $ sudo docker pull opnfv/functest:ovp.1.0.0
- $ sudo docker pull opnfv/yardstick:danube.3.2
- $ sudo docker pull opnfv/testapi:ovp.1.0.0
- $ sudo docker pull mongo:3.2.1
- $ sudo wget -nc http://artifacts.opnfv.org/sdnvpn/ubuntu-16.04-server-cloudimg-amd64-disk1.img -P {ANY_DIR}
- $ sudo wget -nc http://download.cirros-cloud.net/0.3.5/cirros-0.3.5-x86_64-disk.img -P {ANY_DIR}
+ $ sudo docker pull opnfv/functest-vnf:hunter
+ $ wget -nc https://cloud-images.ubuntu.com/releases/14.04/release/ubuntu-14.04-server-cloudimg-amd64-disk1.img -P {ANY_DIR}
+ $ wget -nc http://repository.cloudifysource.org/cloudify/19.01.24/community-release/cloudify-docker-manager-community-19.01.24.tar -P ${DOVETAIL_HOME}/images
-Once all these images are pulled, save the images, copy to the Test Host, and then load
-the Dovetail image and all dependent images at the Test Host. The final two lines above are
-to obtain the test images for transfer to the Test Host.
+Once all these images are pulled, save the images, copy them to the Test Host, and then load
+the Dovetail image and all dependent images at the Test Host.
At the online host, save the images with the command below.
.. code-block:: bash
- $ sudo docker save -o dovetail.tar opnfv/dovetail:ovp.1.0.0 \
- opnfv/functest:ovp.1.0.0 opnfv/yardstick:danube.3.2 \
- opnfv/testapi:ovp.1.0.0 mongo:3.2.1
+ $ sudo docker save -o dovetail.tar opnfv/dovetail:ovp-3.0.0 \
+ opnfv/functest-smoke:hunter opnfv/functest-healthcheck:hunter \
+ opnfv/functest-vnf:hunter \
+ opnfv/yardstick:opnfv-8.0.0 opnfv/bottlenecks:8.0.1-latest
The command above creates a dovetail.tar file with all the images, which can then be copied
to the Test Host. To load the Dovetail images on the Test Host execute the command below.
@@ -451,22 +544,26 @@ Now check to see that all Docker images have been pulled or loaded properly.
.. code-block:: bash
$ sudo docker images
- REPOSITORY TAG IMAGE ID CREATED SIZE
- opnfv/functest ovp.1.0.0 e2b286547478 6 weeks ago 1.26 GB
- opnfv/dovetail ovp.1.0.0 5d25b289451c 8 days ago 516MB
- opnfv/yardstick danube.3.2 df830d5c2cb2 6 weeks ago 1.21 GB
- opnfv/testapi ovp.1.0.0 05c6d5ebce6c 2 months ago 448 MB
- mongo 3.2.1 7e350b877a9a 19 months ago 317 MB
+ REPOSITORY TAG IMAGE ID CREATED SIZE
+ opnfv/dovetail ovp-3.0.0 4b68659da24d 22 hours ago 825MB
+ opnfv/functest-smoke hunter c0253f6de153 3 weeks ago 556MB
+ opnfv/functest-healthcheck hunter fb6d766e38e0 3 weeks ago 379MB
+ opnfv/functest-vnf hunter 31466d52d155 21 hours ago 1.1GB
+ opnfv/yardstick opnfv-8.0.0 189d7d9fbcb2 7 months ago 2.54GB
+ opnfv/bottlenecks 8.0.1-latest 44c1b9fb25aa 5 hours ago 837MB
After copying and loading the Dovetail images at the Test Host, also copy the test images
-(Ubuntu, Cirros) to the Test Host. Copy image ubuntu-16.04-server-cloudimg-amd64-disk1.img
-to ``${DOVETAIL_HOME}/pre_config/``. Copy image cirros-0.3.5-x86_64-disk.img to
-``${DOVETAIL_HOME}/pre_config/``.
+(Ubuntu, Cirros and cloudify-manager) to the Test Host.
+
+- Copy image ``cirros-0.4.0-x86_64-disk.img`` to ``${DOVETAIL_HOME}/images/``.
+- Copy image ``ubuntu-14.04-server-cloudimg-amd64-disk1.img`` to ``${DOVETAIL_HOME}/images/``.
+- Copy image ``ubuntu-16.04-server-cloudimg-amd64-disk1.img`` to ``${DOVETAIL_HOME}/images/``.
+- Copy image ``cloudify-docker-manager-community-19.01.24.tar`` to ``${DOVETAIL_HOME}/images/``.
Starting Dovetail Docker
------------------------
-Regardless of whether you pulled down the Dovetail image directly online, or loaded from
+Regardless of whether you pulled down the Dovetail image directly online, or loaded it from
a static image tar file, you are now ready to run Dovetail. Use the command below to
create a Dovetail container and get access to its shell.
@@ -478,42 +575,13 @@ create a Dovetail container and get access to its shell.
-v /var/run/docker.sock:/var/run/docker.sock \
opnfv/dovetail:<tag> /bin/bash
-The ``-e`` option sets the DOVETAIL_HOME environment variable in the container and the
-``-v`` options map files in the Test Host to files in the container. The latter option
-allows the Dovetail container to read the configuration files and write result files into
-DOVETAIL_HOME on the Test Host. The user should be within the Dovetail container shell,
-once the command above is executed.
-
-Build Local DB and Testapi Services
-^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
-
-The steps in this section only need to be executed if the user plans on storing consolidated
-results on the Test Host that can be uploaded to the OVP portal.
-
-Dovetail needs to build the local DB and testapi service for storing and reporting results
-to the OVP web portal. There is a script in the Dovetail container for building the local DB.
-The ports 27017 and 8000 are used by the DB and testapi respectively. If the Test Host is
-using these ports for existing services, to avoid conflicts, remap the ports to values that
-are unused. Execute the commands below in the Dovetail container to remap ports, as required.
-This step can be skipped if there are no port conflicts with the Test Host.
-
-.. code-block:: bash
-
- $ export mongodb_port=<new_DB_port>
- $ export testapi_port=<new_testapi_port>
-
-Within the Dovetail container, navigate to the directory and execute the shell script using
-the commands below to build the local DB and testapi services.
-
-.. code-block:: bash
-
- $ cd /home/opnfv/dovetail/dovetail/utils/local_db/
- $ ./launch_db.sh
+The ``-e`` option sets the DOVETAIL_HOME environment variable in the container
+and the ``-v`` options mounts files from the test host to the destination path
+inside the container. The latter option allows the Dovetail container to read
+the configuration files and write result files into DOVETAIL_HOME on the Test
+Host. The user should be within the Dovetail container shell, once the command
+above is executed.
-To validate the DB and testapi services are running successfully, navigate to the URL
-``http://<test_host_ip>:<testapi_port>/api/v1/results``, substituting within the URL
-the IP address of the Test Host and the testapi port number. If you can access this URL
-successfully, the services are up and running.
Running the OVP Test Suite
----------------------------
@@ -527,10 +595,10 @@ for the details of the CLI.
$ dovetail run --testsuite <test-suite-name>
-The '--testsuite' option is used to control the set of tests intended for execution
+The ``--testsuite`` option is used to control the set of tests intended for execution
at a high level. For the purposes of running the OVP test suite, the test suite name follows
-the following format, ``ovp.<major>.<minor>.<patch>``. The latest and default test suite is
- ovp.1.0.0.
+the following format, ``ovp.<release-version>``. The latest and default test suite is
+ovp.2019.12.
.. code-block:: bash
@@ -540,132 +608,178 @@ This command is equal to
.. code-block:: bash
- $ dovetail run --testsuite ovp.1.0.0
+ $ dovetail run --testsuite ovp.2019.12
Without any additional options, the above command will attempt to execute all mandatory and
-optional test cases. To restrict the breadth of the test scope, test areas can also be
-specified using the '--testarea' option. The test area can be specified broadly using arguments
-'mandatory' and 'optional'. The mandatory tests can be narrowed further using test area arguments
-'osinterop', 'vping' and 'ha'. The optional tests can be narrowed further using test area
-arguments 'ipv6', 'sdnvpn' and 'tempest'.
+optional test cases with test suite ovp.2019.12.
+To restrict the breadth of the test scope, it can also be specified using options
+``--mandatory`` or ``--optional``.
+
+.. code-block:: bash
+
+ $ dovetail run --mandatory
+
+Also there is a ``--testcase`` option provided to run a specified test case.
.. code-block:: bash
- $ dovetail run --testarea mandatory
+ $ dovetail run --testcase functest.tempest.osinterop
+
+Dovetail allows the user to disable strict API response validation implemented
+by Nova Tempest tests by means of the ``--no-api-validation`` option. Usage of
+this option is only advisable if the SUT returns Nova API responses that
+contain additional attributes. For more information on this command line option
+and its intended usage, refer to
+:ref:`dovetail-exemption_process_api_response_validation`.
+
+.. code-block:: bash
+
+ $ dovetail run --testcase functest.tempest.osinterop --no-api-validation
+
+By default, during test case execution, the respective feature is responsible to
+decide what flavor is going to use for the execution of each test scenario which is under
+of its umbrella.
+In parallel, there is also implemented a mechanism in order for the extra specs in flavors of
+executing test scenarios to be hugepages instead of the default option.
+This is happening if the name of the scenario contains the substring "ovs".
+In this case, the flavor which is going to be used for the running test case has
+'hugepage' characteristics.
+
+Taking the above into consideration and having in mind that the DEPLOY_SCENARIO
+environment parameter is not used by dovetail framework (the initial value is 'unknown'),
+we set as input, for the features that they are responsible for the test case execution,
+the DEPLOY_SCENARIO environment parameter having as substring the feature name "ovs"
+(e.g. os-nosdn-ovs-ha).
+
+Note for the users:
+ - if their system uses DPDK, they should run with ``--deploy-scenario <xx-yy-ovs-zz>``
+ (e.g. os-nosdn-ovs-ha)
+ - this is an experimental feature
+
+.. code-block:: bash
+
+ $ dovetail run --testcase functest.tempest.osinterop --deploy-scenario os-nosdn-ovs-ha
By default, results are stored in local files on the Test Host at ``$DOVETAIL_HOME/results``.
Each time the 'dovetail run' command is executed, the results in the aforementioned directory
are overwritten. To create a singular compressed result file for upload to the OVP portal or
-for archival purposes, the results need to pushed to the local DB. This can be achieved by
-using the '--report' option with an argument syntax as shown below. Note, that the Test Host
-IP address and testapi port number must be substituted with appropriate values.
+for archival purposes, the tool provides an option ``--report``.
.. code-block:: bash
- $ dovetail run --report http://<test_host_ip>:<testapi_port>/api/v1/results
+ $ dovetail run --report
If the Test Host is offline, ``--offline`` should be added to support running with
-local resources.
+local resources. Otherwise, it will try to download resources online during the run time.
.. code-block:: bash
- $ dovetail run --offline --report http://<test_host_ip>:<testapi_port>/api/v1/results
+ $ dovetail run --offline
-Below is an example of running the entire mandatory test area and the creation of the compressed
+Below is an example of running one test case and the creation of the compressed
result file on the Test Host.
.. code-block:: bash
- $ dovetail run --offline --testarea mandatory --report http://192.168.135.2:8000/api/v1/results
- 2017-09-29 07:00:55,718 - run - INFO - ================================================
- 2017-09-29 07:00:55,718 - run - INFO - Dovetail compliance: ovp.1.0.0!
- 2017-09-29 07:00:55,718 - run - INFO - ================================================
- 2017-09-29 07:00:55,719 - run - INFO - Build tag: daily-master-f0795af6-a4e3-11e7-acc5-0242ac110004
- 2017-09-29 07:00:55,956 - run - INFO - >>[testcase]: dovetail.osinterop.tc001
- 2017-09-29 07:15:19,514 - run - INFO - Results have been pushed to database and stored with local file /home/dovetail/results/results.json.
- 2017-09-29 07:15:19,514 - run - INFO - >>[testcase]: dovetail.vping.tc001
- 2017-09-29 07:17:24,095 - run - INFO - Results have been pushed to database and stored with local file /home/dovetail/results/results.json.
- 2017-09-29 07:17:24,095 - run - INFO - >>[testcase]: dovetail.vping.tc002
- 2017-09-29 07:20:42,434 - run - INFO - Results have been pushed to database and stored with local file /home/dovetail/results/results.json.
- 2017-09-29 07:20:42,434 - run - INFO - >>[testcase]: dovetail.ha.tc001
- ...
+ $ dovetail run --offline --testcase functest.vping.userdata --report
+ 2019-12-04 07:31:13,156 - run - INFO - ================================================
+ 2019-12-04 07:31:13,157 - run - INFO - Dovetail compliance: ovp.2019.12!
+ 2019-12-04 07:31:13,157 - run - INFO - ================================================
+ 2019-12-04 07:31:13,157 - run - INFO - Build tag: daily-master-0c9184e6-1668-11ea-b1cd-0242ac110002
+ 2019-12-04 07:31:13,610 - run - INFO - >>[testcase]: functest.vping.userdata
+ 2019-12-04 07:31:13,612 - dovetail.test_runner.DockerRunner - WARNING - There is no hosts file /home/ovp/pre_config/hosts.yaml. This may cause some issues with domain name resolution.
+ 2019-12-04 07:31:14,587 - dovetail.test_runner.DockerRunner - INFO - Get hardware info of all nodes list in file /home/ovp/pre_config/pod.yaml ...
+ 2019-12-04 07:31:14,587 - dovetail.test_runner.DockerRunner - INFO - Hardware info of all nodes are stored in file /home/dovetail/results/all_hosts_info.json.
+ 2019-12-04 07:31:14,612 - dovetail.container.Container - WARNING - There is no hosts file /home/ovp/pre_config/hosts.yaml. This may cause some issues with domain name resolution.
+ 2019-12-04 07:32:13,804 - dovetail.report.Report - INFO - Results have been stored with files: ['/home/ovp/results/functest_results.txt'].
+ 2019-12-04 07:32:13,808 - dovetail.report.Report - INFO -
+
+ Dovetail Report
+ Version: 2019.12
+ Build Tag: daily-master-0c9184e6-1668-11ea-b1cd-0242ac110002
+ Test Date: 2019-12-04 07:32:13 UTC
+ Duration: 60.20 s
+
+ Pass Rate: 100.00% (1/1)
+ vping: pass rate 100.00%
+ -functest.vping.userdata PASS
+
When test execution is complete, a tar file with all result and log files is written in
``$DOVETAIL_HOME`` on the Test Host. An example filename is
-``${DOVETAIL_HOME}/logs_20180105_0858.tar.gz``. The file is named using a
-timestamp that follows the convention 'YearMonthDay-HourMinute'. In this case, it was generated
-at 08:58 on January 5th, 2018. This tar file is used to upload to the OVP portal.
+``${DOVETAIL_HOME}/logs_20191204_0732.tar.gz``. The file is named using a timestamp
+that follows the convention 'YearMonthDay_HourMinute'. In this case, it was generated
+at 07:32 on December 4th, 2019. This tar file is used for uploading the logs and
+results to the OVP portal.
Making Sense of OVP Test Results
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
When a tester is performing trial runs, Dovetail stores results in local files on the Test
-Host by default within the directory specified below. Note, that if the '--report' option
-is used to execute tests, results are written to results.json and the files functest_results.txt
-and dovetail_ha_tcXXX.out will not be created.
+Host by default within directory ``$DOVETAIL_HOME/results``.
+ * Log file: dovetail.log
-.. code-block:: bash
+ * Review the dovetail.log to see if all important information has been captured
- cd $DOVETAIL_HOME/results
+ * In default mode without DEBUG.
-#. Local file
+ * Adding option ``-d/--debug`` to change the mode to be DEBUG.
- * Log file: dovetail.log
-
- * Review the dovetail.log to see if all important information has been captured
- - in default mode without DEBUG.
+ * Result file: results.json
* Review the results.json to see all results data including criteria for PASS or FAIL.
- * Example: OpenStack Interoperability test cases
+ * Tempest and security test cases
- * Can see the log details in ``osinterop_logs/dovetail.osinterop.tc001.log``,
+ * Can see the log details in ``tempest_logs/functest.tempest.XXX.html`` and
+ ``security_logs/functest.security.XXX.html`` respectively,
which has the passed, skipped and failed test cases results.
- * The skipped test cases have the reason for the users to see why these test cases skipped.
+ * This kind of files need to be opened with a web browser.
+
+ * The skipped test cases are accompanied with the reason tag for the users to see why these test cases skipped.
- * The failed test cases have rich debug information for the users to see why these test cases fail.
+ * The failed test cases have rich debug information for the users to see why these test cases failed.
- * Example: vping test case example
+ * Vping test cases
- * Its log is stored in dovetail.log.
+ * Its log is stored in ``vping_logs/functest.vping.XXX.log``.
- * Its result is stored in functest_results.txt.
+ * HA test cases
- * Example: ha test case example
+ * Its log is stored in ``ha_logs/yardstick.ha.XXX.log``.
- * Its log is stored in dovetail.log.
+ * Stress test cases
- * Its result is stored in dovetail_ha_tcXXX.out.
+ * Its log is stored in ``stress_logs/bottlenecks.stress.XXX.log``.
- * Example: ipv6, sdnvpn and tempest test cases examples
+ * VNF test cases
+
+ * Its log is stored in ``vnf_logs/functest.vnf.XXX.log``.
- * Can see the log details in ``ipv6_logs/dovetail.ipv6.tcXXX.log``,
- ``sdnvpn_logs/dovetail.sdnvpn.tcXXX.log`` and ``tempest_logs/dovetail.tempest.tcXXX.log``,
- respectively. They all have the passed, skipped and failed test cases results.
OVP Portal Web Interface
------------------------
The OVP portal is a public web interface for the community to collaborate on results
and to submit results for official OPNFV compliance verification. The portal can be used as a
-resource by users and testers to navigate and inspect results more easily than by manually
+resource by users to navigate and inspect results more easily than by manually
inspecting the log files. The portal also allows users to share results in a private manner
until they are ready to submit results for peer community review.
* Web Site URL
- * https://verified.opnfv.org
+ * https://nfvi-verified.lfnetworking.org
* Sign In / Sign Up Links
- * Accounts are exposed through Linux Foundation or OpenStack account credentials.
+ * Accounts are exposed through Linux Foundation.
* If you already have a Linux Foundation ID, you can sign in directly with your ID.
- * If you do not have a Linux Foundation ID, you can sign up for a new one using 'Sign Up'
+ * If you do not have a Linux Foundation ID, you can sign up for a new one using 'Sign Up'.
* My Results Tab
@@ -673,20 +787,25 @@ until they are ready to submit results for peer community review.
* This page lists all results uploaded by you after signing in.
- * You can also upload results on this page with the two steps below.
+ * Following the two steps below, the results are uploaded and in status 'private'.
+
+ * Obtain results tar file located at ``${DOVETAIL_HOME}/``, e.g. ``logs_20180105_0858.tar.gz``.
- * Obtain results tar file located at ``${DOVETAIL_HOME}/``, example ``logs_20180105_0858.tar.gz``
+ * Use the *Choose File* button where a file selection dialog allows you to choose your result file from the hard-disk. Then click the *Upload result* button and see a results ID once your upload succeeds.
- * Use the *Choose File* button where a file selection dialog allows you to choose your result
- file from the hard-disk. Then click the *Upload* button and see a results ID once your
- upload succeeds.
+ * Results are remaining in status 'private' until they are submitted for review.
- * Results are status 'private' until they are submitted for review.
+ * Use the *Operation* column drop-down option *submit to review*, to expose results to
+ OPNFV community peer reviewers. Use the *withdraw submit* option to reverse this action.
- * Use the *Operation* column drop-down option 'submit to review', to expose results to
- OPNFV community peer reviewers. Use the 'withdraw submit' option to reverse this action.
+ * Results status are changed to be 'review' after submit to review.
- * Use the *Operation* column drop-down option 'share with' to share results with other
+ * Use the *View Reviews* to find the review status including reviewers' names and the outcome.
+
+ * The administrator will approve the results which have got 2 positive outcome from 2 reviewers.
+ Then the status will be changed to be 'verified'.
+
+ * Use the *Operation* column drop-down option *share with* to share results with other
users by supplying either the login user ID or the email address associated with
the share target account. The result is exposed to the share target but remains private
otherwise.
@@ -695,18 +814,19 @@ until they are ready to submit results for peer community review.
* This page shows your account info after you sign in.
+ * There are 3 different roles: administrator, user and reviewer.
+
Updating Dovetail or a Test Suite
---------------------------------
Follow the instructions in section `Installing Dovetail on the Test Host`_ and
-`Running the OVP Test Suite`_ by replacing the docker images with new_tags,
+`Running the OVP Test Suite`_ by replacing the docker images with new_tags:
.. code-block:: bash
sudo docker pull opnfv/dovetail:<dovetail_new_tag>
sudo docker pull opnfv/functest:<functest_new_tag>
sudo docker pull opnfv/yardstick:<yardstick_new_tag>
+ sudo docker pull opnfv/bottlenecks:<bottlenecks_new_tag>
This step is necessary if dovetail software or the OVP test suite have updates.
-
-
diff --git a/docs/testing/user/userguide/vnf_test_guide.rst b/docs/testing/user/userguide/vnf_test_guide.rst
new file mode 100644
index 00000000..00c4e4ef
--- /dev/null
+++ b/docs/testing/user/userguide/vnf_test_guide.rst
@@ -0,0 +1,714 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) OPNFV, ONAP, and others.
+
+.. _dovetail-vnf_testers_guide:
+
+===================================
+Conducting ONAP VNF Testing for OVP
+===================================
+
+Overview
+--------
+
+As the LFN verification framework, the Dovetail team has worked with the ONAP VVP, and VTP
+projects to enable VNF testing, results submission, and results review to be completed
+throught the same web portal and processes used for the NFVI testing.
+For more details about VNF SDK and VVP, please refer to `ONAP VNF SDK Compliance Verification Program
+<https://docs.onap.org/en/elalto/submodules/vnfsdk/model.git/docs/files/VNFSDK-LFN-CVC.html>`_
+and `ONAP VVP <https://docs.onap.org/en/elalto/submodules/vvp/documentation.git/docs/index.html>`_.
+
+Testing is available for both HEAT and TOSCA defined VNFs, but the process is different depending
+on the template language. This userguide covers the testing process for both VNF types in the
+two sections below.
+
+
+Definitions and abbreviations
+-----------------------------
+
+- LFN - Linux Foundation Networking
+- ONAP - Open Network Automation Platform
+- OVP - OPNFV Verification Program
+- VNF - Virtual Network Function
+- VNF SDK - VNF Software Development Kit
+- VTP - VNF Test Platform
+- VVP - VNF Validation Program
+
+Testing of HEAT based VNFs
+--------------------------
+
+Environment Preparation
+^^^^^^^^^^^^^^^^^^^^^^^
+
+Prerequisites
+"""""""""""""
+
+- `ONAP ElAlto Release deployed via OOM <https://onap.readthedocs.io/en/latest/submodules/oom.git/docs/oom_quickstart_guide.html>`_
+- An OpenStack deployment is available and privisioned as ONAP's Cloud Site
+- kubectl is installed on the system used to start the testing
+- bash
+- VNF Heat Templates
+- Preload json files
+
+After deploying ONAP, you need to configure ONAP with:
+
+- A cloud owner
+- A cloud region
+- A subscriber
+- A service type
+- A project name
+- An owning entity
+- A platform
+- A line of business
+- A cloud site
+
+If you're not familiar with how to configure ONAP, there are guides that use
+`robot <https://onap.readthedocs.io/en/elalto/submodules/integration.git/docs/docs_robot.html>`_
+or `direct api <https://wiki.onap.org/pages/viewpage.action?pageId=25431491>`_ requests available
+to help, as well as a guide for adding a new OpenStack site to ONAP.
+
+VVP Test Tool Setup
+"""""""""""""""""""
+
+On your local machine, or the system from which you will run the tests, you will need to clone the
+ONAP OOM project repo:
+
+.. code-block:: bash
+
+ git clone --branch 5.0.1-ONAP ssh://<username>@gerrit.onap.org:29418/oom --recurse-submodules
+
+VNF Preparation
+^^^^^^^^^^^^^^^
+
+The vnf lifecycle validation testsuite requires the VNF to be packaged into a specific directory
+hierarchy, shown below.
+
+.. code-block::
+
+ vnf_folder
+ ├── /templates
+ | └── base.yaml
+ | └── base.env
+ | └── incremental_0.yaml
+ | └── incremental_0.env
+ | └── ...
+ ├── /preloads
+ | └── base_preload.json
+ | └── incremental_0_preload.json
+ | └── ...
+ └── vnf-details.json
+
+- The name for vnf_folder is free-form, and can be located anywhere on your computer. The path to this folder will be passed to the testsuite as an argument.
+- /templates should contain your VVP-compliant VNF heat templates.
+- /preloads should contain a preload file for each VNF module (TODO: add link to preload documentation).
+ - For a VNF-API preload: vnf-name, vnf-type, generic-vnf-type, and generic-vnf-name should be empty strings.
+ - For a GR-API preload: vnf-name, vnf-type, vf-module-type, and vf-module-name should be empty strings.
+ - This information will be populated at runtime by the testsuite.
+- vnf-details should be a json file with the information that will be used by ONAP to instantiate the VNF. The structure of vnf-details is shown below.
+- VNF disk image must be uploaded and available in the OpenStack project being managed by ONAP
+- Modules must contain an entry for each module of the VNF. Only one module can be a base module.
+- api_type should match the format of the preloads that are provided in the package.
+- The other information should match what was used to configure ONAP during the pre-requisite section of this guide.
+
+.. code-block:: json
+
+ {
+ "vnf_name": "The Vnf Name",
+ "description": "Description of the VNF",
+ "modules": [
+ {
+ "filename": "base.yaml",
+ "isBase": "true",
+ "preload": "base_preload.json"
+ },
+ {
+ "filename": "incremental_0.yaml",
+ "isBase": "false",
+ "preload": "incremental_0.json"
+ },
+ ...
+ ],
+ "api_type": "[gr_api] or [vnf_api]",
+ "subscriber": "<subscriber name>",
+ "service_type": "<service type>",
+ "tenant_name": "<name of tenant>",
+ "region_id": "<name of region>",
+ "cloud_owner": "<name of cloud owner>",
+ "project_name": "<name of project>",
+ "owning_entity": "<name of owning entity>",
+ "platform": "<name of platform>",
+ "line_of_business": "<name of line of business>",
+ "os_password": "<openstack password>"
+ }
+
+Runnign the HEAT VNF Test
+^^^^^^^^^^^^^^^^^^^^^^^^^
+
+The ONAP OOM Robot framework will run the test, using kubectl to manage the execution. The framework
+will copy your VNF template files to the robot container required to execute the test.
+
+.. code-block:: bash
+
+ cd oom/kubernetes/robot
+ $ ./instantiate-k8s.sh --help
+ ./instantiate-k8s.sh [options]
+
+ required:
+ -n, --namespace <namespace> namespace that robot pod is running under.
+ -f, --folder <folder> path to folder containing heat templates, preloads, and vnf-details.json.
+
+ additional options:
+ -p, --poll some cloud environments (like azure) have a short time out value when executing
+ kubectl. If your shell exits before the testsuite finishes, using this option
+ will poll the testsuite logs every 30 seconds until the test finishes.
+ -t, --tag <tag> robot testcase tag to execute (default is instantiate_vnf).
+
+ This script executes the VNF instantiation robot testsuite.
+ - It copies the VNF folder to the robot container that is part of the ONAP deployment.
+ - It models, distributes, and instantiates a heat-based VNF.
+ - It copies the logs to an output directory, and creates a tarball for upload to the OVP portal.
+
+
+**Sample execution:**
+
+.. code-block:: bash
+
+ $ ./instantiate-k8s.sh --namespace onap --folder /tmp/vnf-instantiation/examples/VNF_API/pass/multi_module/ --poll
+ ...
+ ...
+ ...
+ ...
+ ------------------------------------------------------------------------------
+ Testsuites.Vnf Instantiation :: The main driver for instantiating ... | PASS |
+ 1 critical test, 1 passed, 0 failed
+ 1 test total, 1 passed, 0 failed
+ ==============================================================================
+ Testsuites | PASS |
+ 1 critical test, 1 passed, 0 failed
+ 1 test total, 1 passed, 0 failed
+ ==============================================================================
+ Output: /share/logs/0003_ete_instantiate_vnf/output.xml
+ + set +x
+ testsuite has finished
+ Copying Results from pod...
+ /tmp/vnf-instantiation /tmp/vnf-instantiation
+ a log.html
+ a results.json
+ a stack_report.json
+ a validation-scripts.json
+ /tmp/vnf-instantiation
+ VNF test results: /tmp/vnfdata.46749/vnf_heat_results.tar.gz
+
+The testsuite takes about 10-15 minutes for a simple VNF, and will take longer for a more complicated VNF.
+
+Reporting Results
+"""""""""""""""""
+Once the testsuite is finished, it will create a directory and tarball in /tmp (the name of the directory
+and file is shown at the end of the stdout of the script). There will be a results.json in that directory
+that has the ultimate outcome of the test, in the structure shown below.
+
+**Log Files**
+
+The output tar file will have 4 log files in it.
+
+- results.json: This is high-level results file of all of the test steps, and is consumed by the OVP portal.
+- report.json: This is the output of the vvp validation scripts.
+- stack_report.json: This is the output from querying openstack to validate the heat modules.
+- log.html: This is the robot log, and contains each execution step of the testcase.
+
+If the result is "PASS", that means the testsuite was successful and the tarball is ready for submission
+to the OVP portal.
+
+**results.json**
+
+.. code-block:: json
+
+ {
+ "vnf_checksum": "afc57604a3b3b7401d5b8648328807b594d7711355a2315095ac57db4c334a50",
+ "build_tag": "vnf-validation-53270",
+ "version": "2019.09",
+ "test_date": "2019-09-04 17:50:10.575",
+ "duration": 437.002,
+ "vnf_type": "heat",
+ "testcases_list": [
+ {
+ "mandatory": "true",
+ "name": "onap-vvp.validate.heat",
+ "result": "PASS",
+ "objective": "onap heat template validation",
+ "sub_testcase": [],
+ "portal_key_file": "report.json"
+ },
+ {
+ "mandatory": "true",
+ "name": "onap-vvp.lifecycle_validate.heat",
+ "result": "PASS",
+ "objective": "onap vnf lifecycle validation",
+ "sub_testcase": [
+ {
+ "name": "model-and-distribute",
+ "result": "PASS"
+ },
+ {
+ "name": "instantiation",
+ "result": "PASS"
+ }
+ ],
+ "portal_key_file": "log.html"
+ },
+ {
+ "mandatory": "true",
+ "name": "stack_validation",
+ "result": "PASS",
+ "objective": "onap vnf openstack validation",
+ "sub_testcase": [],
+ "portal_key_file": "stack_report.json"
+ }
+ ]
+ }
+
+
+Additional Resources
+^^^^^^^^^^^^^^^^^^^^
+
+- `ONAP VVP Project <https://wiki.onap.org/display/DW/VNF+Validation+Program+Project>`_
+- `VVP Wiki Users Guide (this will track current ONAP master) <https://wiki.onap.org/pages/viewpage.action?pageId=68546123>`_
+
+Sample VNF templates are available on the VVP Wiki Users Guide page.
+
+Testing of TOSCA based VNFs
+---------------------------
+
+VNF Test Platform (VTP) provides an platform to on-board different test cases required for
+OVP for various VNF testing provided by VNFSDK (for TOSCA) and VVP(for HEAT) projects in
+ONAP. And it generates the test case outputs which would be uploaded into OVP portal for
+VNF badging.
+
+TOSCA VNF Test Environment
+^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+As pre-requestiests steps, Its assumed that, successful ONAP, Vendor VNFM and OpenStack
+cloud are already available. Below installation steps help to setup VTP components and CLI.
+
+.. image:: images/tocsa_vnf_test_environment.png
+ :align: center
+ :scale: 100%
+
+Installation
+^^^^^^^^^^^^
+
+Clone the VNFSDK repo.
+
+.. code-block:: bash
+
+ git clone --branch elalto https://git.onap.org/vnfsdk/refrepo
+
+Install the VTP by using script *refrepo/vnfmarket-be/deployment/install/vtp_install.sh*
+
+Follow the steps as below (in sequence):
+
+- vtp_install.sh --download : It will download all required artifacts into /opt/vtp_stage
+- vtp_install.sh --install : It will install VTP (/opt/controller) and CLI (/opt/oclip)
+- vtp_install.sh --start : It will start VTP controller as tomcat service and CLI as oclip service
+- vtp_install.sh --verify : It will verify the setup is done properly by running some test cases.
+
+Last step (verify) would check the health of VTP components and TOSCA VNF compliance and validation test cases.
+
+Check Available Test Cases
+""""""""""""""""""""""""""
+
+VTP supports to check the compliance of VNF and PNF based on ONAP VNFREQS.
+
+To check:
+
+- Go to command console
+- Run command oclip
+- Now it will provide a command prompt:
+
+*oclip:open-cli>*
+
+Now run command as below and check the supported compliance testcases for VNFREQS.
+
+- csar-validate - Helps to validate given VNF CSAR for all configured VNFREQS.
+- csar-validate-rxxx - Helps to validate given VNF CSAR for a given VNFREQS xxx.
+
+.. code-block:: bash
+
+ oclip:open-cli>schema-list --product onap-dublin --service vnf-compliance
+ +--------------+----------------+------------------------+--------------+----------+------+
+ |product |service |command |ocs-version |enabled |rpc |
+ +--------------+----------------+------------------------+--------------+----------+------+
+ |onap-dublin |vnf-compliance |csar-validate-r10087 |1.0 |true | |
+ +--------------+----------------+------------------------+--------------+----------+------+
+ |onap-dublin |vnf-compliance |csar-validate |1.0 |true | |
+ +--------------+----------------+------------------------+--------------+----------+------+
+ |onap-dublin |vnf-compliance |csar-validate-r26885 |1.0 |true | |
+ +--------------+----------------+------------------------+--------------+----------+------+
+ |onap-dublin |vnf-compliance |csar-validate-r54356 |1.0 |true | |
+ ...
+
+To know the details of each VNFREQS, run as below.
+
+.. code-block:: bash
+
+ oclip:open-cli>use onap-dublin
+ oclip:onap-dublin>csar-validate-r54356 --help
+ usage: oclip csar-validate-r54356
+
+ Data types used by NFV node and is based on TOSCA/YAML constructs specified in draft GS NFV-SOL 001.
+ The node data definitions/attributes used in VNFD MUST comply.
+
+Now run command as below and check the supported validation testcases
+
+.. code-block:: bash
+
+ oclip:onap-dublin>use open-cli
+ oclip:open-cli>schema-list --product onap-dublin --service vnf-validation
+ +--------------+----------------+----------------------+--------------+----------+------+
+ |product |service |command |ocs-version |enabled |rpc |
+ +--------------+----------------+----------------------+--------------+----------+------+
+ |onap-dublin |vnf-validation |vnf-tosca-provision |1.0 |true | |
+ +--------------+----------------+----------------------+--------------+----------+------+
+
+Configure ONAP with required VNFM and cloud details
+"""""""""""""""""""""""""""""""""""""""""""""""""""
+
+**1. Setup the OCOMP profile onap-dublin**
+
+Run following command to configure the ONAP service URL and creadentials as given below, which will be
+used by VTP while executing the test cases
+
+.. code-block:: bash
+
+ oclip:open-cli>use onap-dublin
+ oclip:onap-dublin>profile onap-dublin
+ oclip:onap-dublin>set sdc.onboarding:host-url=http://159.138.8.8:30280
+ oclip:onap-dublin>set sdc.onboarding:host-username=cs0008
+ oclip:onap-dublin>set sdc.onboarding:host-password=demo123456!
+ oclip:onap-dublin>set sdc.catalog:host-url=http://159.138.8.8:30205
+ oclip:onap-dublin>set sdc.catalog:host-password=demo123456\!
+ oclip:onap-dublin>set sdc.catalog:host-username=cs0008
+ oclip:onap-dublin>set sdc.catalog:service-model-approve:host-username=gv0001
+ oclip:onap-dublin>set sdc.catalog:service-model-distribute:host-username=op0001
+ oclip:onap-dublin>set sdc.catalog:service-model-test-start:host-username=jm0007
+ oclip:onap-dublin>set sdc.catalog:service-model-test-accept:host-username=jm0007
+ oclip:onap-dublin>set sdc.catalog:service-model-add-artifact:host-username=ocomp
+ oclip:onap-dublin>set sdc.catalog:vf-model-add-artifact:host-username=ocomp
+ oclip:onap-dublin>set aai:host-url=https://159.138.8.8:30233
+ oclip:onap-dublin>set aai:host-username=AAI
+ oclip:onap-dublin>set aai:host-password=AAI
+ oclip:onap-dublin>set vfc:host-url=http://159.138.8.8:30280
+ oclip:onap-dublin>set multicloud:host-url=http://159.138.8.8:30280
+
+NOTE: Mostly all above entries value would be same execept the IP address used in the
+URL, which would be ONAP k8s cluser IP.
+
+By default, SDC onboarding service does not provide node port, which is available to
+access from external ONAP network. so to enable for external access, register the SDC
+onboarding service into MSB and use MSB url for sdc.onboarding:host-url.
+
+.. code-block:: bash
+
+ oclip:onap-dublin> microservice-create --service-name sdcob --service-version v1.0 --service-url /onboarding-api/v1.0 --path /onboarding-api/v1.0 --node-ip 172.16.1.0 --node-port 8081
+
+NOTE: To find the node-ip and node-port, use the following steps.
+
+Find out SDC onboarding service IP and port details as given here:
+
+.. code-block:: bash
+
+ [root@onap-dublin-vfw-93996-50c1z ~]# kubectl get pods -n onap -o wide | grep sdc-onboarding-be
+ dev-sdc-sdc-onboarding-be-5564b877c8-vpwr5 2/2 Running 0 29d 172.16.1.0 192.168.2.163 <none> <none>
+ dev-sdc-sdc-onboarding-be-cassandra-init-mtvz6 0/1 Completed 0 29d 172.16.0.220 192.168.2.163 <none> <none>
+ [root@onap-dublin-vfw-93996-50c1z ~]#
+
+Note down the IP address for sdc-onboarding-be 172.16.1.0
+
+.. code-block:: bash
+
+ [root@onap-dublin-vfw-93996-50c1z ~]# kubectl get services -n onap -o wide | grep sdc-onboarding-be
+ sdc-onboarding-be ClusterIP 10.247.198.92 <none> 8445/TCP,8081/TCP 29d app=sdc-onboarding-be,release=dev-sdc
+ [root@onap-dublin-vfw-93996-50c1z ~]#
+
+Note down the port for sdc-onboarding-be 8445 8081
+
+Similarly, other service IP and Port could be discovered like above, in case not know earlier :)
+
+Verify these details once by typing 'set'
+
+.. code-block:: bash
+
+ oclip:onap-dublin> set
+
+This profile would be used by user while running the test cases with ONAP setup configured in it, as below
+oclip --profile onap-dublin vnf-tosca-provision ....
+
+**2. Setup SDC consumer**
+
+SDC uses consumer concept to configure required VN model and service model artifacts. So
+following commands required to run, which will create consumer named ocomp, which is
+already configured in onap-dublin profile created in above steps.
+
+.. code-block:: bash
+
+ oclip --product onap-dublin --profile onap-dublin sdc-consumer-create --consumer-name ocomp
+
+NOTE: command oclip could be used in scripting mode as above or in interactive mode as used
+in earlier steps
+
+**3. Update the cloud and vnfm driver details**
+
+In the configuration file /opt/oclip/conf/vnf-tosca-provision.json, update the cloud
+and VNFM details.
+
+.. code-block:: json
+
+ "cloud": {
+ "identity-url": "http://10.12.11.1:5000/v3",
+ "username": "admin",
+ "password": "password",
+ "region": "RegionOVP",
+ "version": "ocata",
+ "tenant": "ocomp"
+ },
+ "vnfm":{
+ "hwvnfmdriver":{
+ "version": "v1.0",
+ "url": "http://159.138.8.8:38088",
+ "username": "admin",
+ "password": "xxxx"
+ },
+ "gvnfmdriver":{
+ "version": "v1.0",
+ "url": "http://159.138.8.8:30280"
+ }
+ }
+
+**4.Configure the decided VNFRES (optional)**
+VTP allows to configure the set of VNFREQS to be considered while running the VNF
+compliance test cases in the configuration file /opt/oclip/conf/vnfreqs.properties.
+
+If not available, please create this file with following entries:
+
+.. code-block:: bash
+
+ vnfreqs.enabled=r02454,r04298,r07879,r09467,r13390,r23823,r26881,r27310,r35851,r40293,r43958,r66070,r77707,r77786,r87234,r10087,r21322,r26885,r40820,r35854,r65486,r17852,r46527,r15837,r54356,r67895,r95321,r32155,r01123,r51347,r787965,r130206
+ pnfreqs.enabled=r10087,r87234,r35854,r15837,r17852,r293901,r146092,r57019,r787965,r130206
+ # ignored all chef and ansible related tests
+ vnferrors.ignored=
+ pnferrors.ignored=
+
+Runnign the TOSCA VNF Test
+^^^^^^^^^^^^^^^^^^^^^^^^^^
+
+Every test provided in VTP is given with guidelines on how to use it. On every execution of test cases, use the following additional arguments based on requirements
+
+- --product onap-dublin - It helps VTP choose the test cases written for onap-dublin version
+- --profile onap-dublin - It helps VTP to use the profile settings provided by admin (optional)
+- --request-id - It helps VTP to track the progress of the test cases execution and user could use this id for same. (optional)
+
+So, final test case execution would be as below. To find the test case arguments details, run second command below.
+
+.. code-block:: bash
+
+ oclip --product onap-dublin --profile onap-dublin --request-id req-1 <test case name> <test case arguments>
+ oclip --product onap-dublin <test case name> --help
+
+Running TOSCA VNF Compliance Testing
+""""""""""""""""""""""""""""""""""""
+
+To run compliance test as below with given CSAR file
+
+.. clode-block:: bash
+
+ oclip --product onap-dublin csar-validate --csar <csar file complete path>
+
+It will produce the result format as below:
+
+.. code-block:: json
+
+ {
+ "date": "Fri Sep 20 17:34:24 CST 2019",
+ "criteria": "PASS",
+ "contact": "ONAP VTP Team onap-discuss@lists.onap.org",
+ "results": [
+ {
+ "description": "V2.4.1 (2018-02)",
+ "passed": true,
+ "vnfreqName": "SOL004",
+ "errors": []
+ },
+ {
+ "description": "If the VNF or PNF CSAR Package utilizes Option 2 for package security, then the complete CSAR file MUST be digitally signed with the VNF or PNF provider private key. The VNF or PNF provider delivers one zip file consisting of the CSAR file, a signature file and a certificate file that includes the VNF or PNF provider public key. The certificate may also be included in the signature container, if the signature format allows that. The VNF or PNF provider creates a zip file consisting of the CSAR file with .csar extension, signature and certificate files. The signature and certificate files must be siblings of the CSAR file with extensions .cms and .cert respectively.\n",
+ "passed": true,
+ "vnfreqName": "r787965",
+ "errors": []
+ }
+ ],
+ "platform": "VNFSDK - VNF Test Platform (VTP) 1.0",
+ "vnf": {
+ "mode": "WITH_TOSCA_META_DIR",
+ "vendor": "ONAP",
+ "name": null,
+ "type": "TOSCA",
+ "version": null
+ }
+ }
+
+In case of errors, the errors section will have list of details as below. Each error block, will be
+given with error code and error details. Error code would be very useful to provide the troubleshooting
+guide in future. Note, to generate the test result in OVP archieve format, its recommended to run this compliance
+test with request-id similar to running validation test as below.
+
+.. code-block:: bash
+
+ [
+ {
+ "vnfreqNo": "R66070",
+ "code": "0x1000",
+ "message": "MissinEntry-Definitions file",
+ "lineNumber": -1
+ }
+ ]
+
+Running TOSCA VNF Validation Testing
+""""""""""""""""""""""""""""""""""""
+VTP provides validation test case with following modes:
+
+.. image:: images/tosca_vnf_test_flow.png
+ :align: left
+ :scale: 100%
+
+
+setup: Create requires Vendor, Service Subscription and VNF cloud in ONAP
+standup: From the given VSP csar, VNF csar and NS csar, it creates VF Model, NS Model and NS service
+cleanup: Remove those entries created during provision
+provision: Runs setup -> standup
+validate: Runs setup -> standup -> cleanup
+checkup: mode helps to verify automation is deployed properly.
+
+For OVP badging, validate mode would be used as below:
+
+.. code-block:: bash
+
+ oclip --request-id WkVVu9fD--product onap-dublin --profile onap-dublin vnf-tosca-provision --vsp <vsp csar> --vnf-csar <v
+
+Validation testing would take for a while to complete the test execution, so user could use the above
+given request-id, to tracking the progress as below:
+
+.. code-block:: bash
+
+ oclip execution-list --request-id WkVVu9fD
+ +------------+------------------------+--------------+------------------+------------------------------+--------------+------------+--------------------------+--------------------------+
+ |request-id |execution-id |product |service |command |profile |status |start-time |end-time |
+ +------------+------------------------+--------------+------------------+------------------------------+--------------+------------+--------------------------+--------------------------+
+ |WkVVu9fD |WkVVu9fD-1568731678753 |onap-dublin |vnf-validation |vnf-tosca-provision | |in-progress |2019-09-17T14:47:58.000 | |
+ +------------+------------------------+--------------+------------------+------------------------------+--------------+------------+--------------------------+--------------------------+
+ |WkVVu9fD |WkVVu9fD-1568731876397 |onap-dublin |sdc.catalog |service-model-test-request |onap-dublin |in-progress |2019-09-17T14:51:16.000 | |
+ +------------+------------------------+--------------+------------------+------------------------------+--------------+------------+--------------------------+--------------------------+
+ |WkVVu9fD |WkVVu9fD-1568731966966 |onap-dublin |sdc.onboarding |vsp-archive |onap-dublin |completed |2019-09-17T14:52:46.000 |2019-09-17T14:52:47.000 |
+ +------------+------------------------+--------------+------------------+------------------------------+--------------+------------+--------------------------+--------------------------+
+ |WkVVu9fD |WkVVu9fD-1568731976982 |onap-dublin |aai |subscription-delete |onap-dublin |completed |2019-09-17T14:52:56.000 |2019-09-17T14:52:57.000 |
+ +------------+------------------------+--------------+------------------+------------------------------+--------------+------------+--------------------------+--------------------------+
+ |WkVVu9fD |WkVVu9fD-1568731785780 |onap-dublin |aai |vnfm-create |onap-dublin |completed |2019-09-17T14:49:45.000 |2019-09-17T14:49:46.000 |
+ ......
+
+While executing the test cases, VTP provides unique execution-id (2nd column) for each step. As you note
+in the example above, some steps are in-progress, while others are completed already. If there is error
+then status will be set to failed.
+
+To find out the foot-print of each step, following commands are available
+
+.. code-block:: bash
+
+ oclip execution-show-out --execution-id WkVVu9fD-1568731785780 - Reports the standard output logs
+ oclip execution-show-err --execution-id WkVVu9fD-1568731785780 - Reports the standard error logs
+ oclip execution-show-debug --execution-id WkVVu9fD-1568731785780 - Reports the debug details like HTTP request and responseoclip execution-show --execution-id WkVVu9fD-1568731785780 - Reports the complete foot-print of inputs, outputs of steps
+
+Track the progress of the vnf-tosca-provision test cases until its completed. Then the out of the validation
+test cases could be retrieved as below:
+
+.. code-block:: bash
+
+ oclip execution-show --execution-id WkVVu9fD-1568731678753 - use vnf tosca test case execution id here
+
+It will provides the output format as below:
+
+.. code-block:: json
+
+ {
+ "output": {
+ "ns-id": null,
+ "vnf-id": "",
+ "vnfm-driver": "hwvnfmdriver",
+ "vnf-vendor-name": "huawei",
+ "onap-objects": {
+ "ns_instance_id": null,
+ "tenant_version": null,
+ "service_type_id": null,
+ "tenant_id": null,
+ "subscription_version": null,
+ "esr_vnfm_id": null,
+ "location_id": null,
+ "ns_version": null,
+ "vnf_status": "active",
+ "entitlement_id": null,
+ "ns_id": null,
+ "cloud_version": null,
+ "cloud_id": null,
+ "vlm_version": null,
+ "esr_vnfm_version": null,
+ "vlm_id": null,
+ "vsp_id": null,
+ "vf_id": null,
+ "ns_instance_status": "active",
+ "service_type_version": null,
+ "ns_uuid": null,
+ "location_version": null,
+ "feature_group_id": null,
+ "vf_version": null,
+ "vsp_version": null,
+ "agreement_id": null,
+ "vf_uuid": null,
+ "ns_vf_resource_id": null,
+ "vsp_version_id": null,
+ "customer_version": null,
+ "vf_inputs": null,
+ "customer_id": null,
+ "key_group_id": null,
+ },
+ "vnf-status": "active",
+ "vnf-name": "vgw",
+ "ns-status": "active"
+ },
+ "input": {
+ "mode": "validate",
+ "vsp": "/tmp/data/vtp-tmp-files/1568731645518.csar",
+ "vnfm-driver": "hwvnfmdriver",
+ "config-json": "/opt/oclip/conf/vnf-tosca-provision.json",
+ "vnf-vendor-name": "huawei",
+ "ns-csar": "/tmp/data/vtp-tmp-files/1568731660745.csar",
+ "onap-objects": "{}",
+ "timeout": "600000",
+ "vnf-name": "vgw",
+ "vnf-csar": "/tmp/data/vtp-tmp-files/1568731655310.csar"
+ },
+ "product": "onap-dublin",
+ "start-time": "2019-09-17T14:47:58.000",
+ "service": "vnf-validation",
+ "end-time": "2019-09-17T14:53:46.000",
+ "request-id": "WkVVu9fD-1568731678753",
+ "command": "vnf-tosca-provision",
+ "status": "completed"
+ }
+
+Reporting Results
+"""""""""""""""""
+
+VTP provides translation tool to migrate the VTP result into OVP portal format and generates the tar file
+for the given test case execution. Please refer `<https://github.com/onap/vnfsdk-refrepo/tree/master/vnfmarket-be/deployment/vtp2ovp>`_ for more details.
+
+Once tar is generated, it can be used to submit into OVP portal `<https://vnf-verified.lfnetworking.org/>`_
+
+.. References
+.. _`OVP VNF portal`: https://vnf-verified.lfnetworking.org