summaryrefslogtreecommitdiffstats
path: root/docs/design
diff options
context:
space:
mode:
authorblsaws <bryan.sullivan@att.com>2016-08-18 22:35:08 -0700
committerblsaws <bryan.sullivan@att.com>2016-08-18 22:35:08 -0700
commit816fc353daf07ffe8f2adc12d565c070fe4f3746 (patch)
tree46c8fa7fca2e7514e5581bf2fd1dc73cbb620201 /docs/design
parent4e504df222285d26bee5365c166d690ae6997353 (diff)
Update docs for Colorado
JIRA: COPPER-1 Change-Id: I2c0b1fa29c7129aceff6779ca48f7fb26f288063 Signed-off-by: blsaws <bryan.sullivan@att.com>
Diffstat (limited to 'docs/design')
-rw-r--r--docs/design/architecture.rst83
-rw-r--r--docs/design/definitions.rst7
-rw-r--r--docs/design/introduction.rst81
-rw-r--r--docs/design/requirements.rst145
-rw-r--r--docs/design/usecases.rst169
5 files changed, 319 insertions, 166 deletions
diff --git a/docs/design/architecture.rst b/docs/design/architecture.rst
index d55c208..8c6d04b 100644
--- a/docs/design/architecture.rst
+++ b/docs/design/architecture.rst
@@ -3,76 +3,93 @@ Architecture
Architectural Concept
---------------------
-The following example diagram illustrates a "relationship diagram" type view of an NFVI platform,
-in which the roles of components focused on policy management, services, and infrastructure are shown.
-This view illustrates that a large-scale deployment of NFVI may leverage multiple components of the same "type"
-(e.g. SDN Controller), which fulfill specific purposes for which they are optimized. For example, a global SDN
-controller and cloud orchestrator can act as directed by a service orchestrator in the provisioning of VNFs per
-intent, while various components at a local and global level handle policy-related events directly and/or feed
-them back through a closed-loop policy design that responds as needed, directly or through the service orchestrator.
+The following example diagram illustrates a "relationship diagram" type view of
+an NFVI platform, in which the roles of components focused on policy management,
+services, and infrastructure are shown.
+
+This view illustrates that a large-scale deployment of NFVI may leverage multiple
+components of the same "type" (e.g. SDN Controller), which fulfill specific
+purposes for which they are optimized. For example, a global SDN controller and
+cloud orchestrator can act as directed by a service orchestrator in the
+provisioning of VNFs per intent, while various components at a local and global
+level handle policy-related events directly and/or feed them back through a
+closed-loop policy design that responds as needed, directly or through the
+service orchestrator.
.. image:: ./images/policy_architecture.png
:width: 700 px
:alt: policy_architecture.png
:align: center
-(source of the diagram above: https://git.opnfv.org/cgit/copper/plain/design_docs/images/policy_architecture.pptx)
+(source of the diagram above:
+https://git.opnfv.org/cgit/copper/plain/design_docs/images/policy_architecture.pptx)
Architectural Aspects
---------------------
* Policies are reflected in two high-level goals
- * Ensure resource requirements of VNFs and services are applied per VNF designer, service, and tenant intent
- * Ensure that generic policies are not violated,
- e.g. *networks connected to VMs must either be public or owned by the VM owner*
+ * Ensure resource requirements of VNFs and services are applied per VNF
+ designer, service, and tenant intent
+ * Ensure that generic policies are not violated, e.g. *networks connected to
+ VMs must either be public or owned by the VM owner*
* Policies are distributed through two main means
- * As part of VNF packages, customized if needed by Service Design tools, expressing intent of the VNF designer and
- service provider, and possibly customized or supplemented by service orchestrators per the intent of specific
- tenants
- * As generic policies provisioned into VIMs (SDN controllers and cloud orchestrators), expressing intent of the
- service provider re what states/events need to be policy-governed independently of specific VNFs
+ * As part of VNF packages, customized if needed by Service Design tools,
+ expressing intent of the VNF designer and service provider, and possibly
+ customized or supplemented by service orchestrators per the intent of
+ specific tenants
+ * As generic policies provisioned into VIMs (SDN controllers and cloud
+ orchestrators), expressing intent of the service provider re what
+ states/events need to be policy-governed independently of specific VNFs
- * Policies are applied locally and in closed-loop systems per the capabilities of the local policy enforcer and
- the impact of the related state/event conditions
+ * Policies are applied locally and in closed-loop systems per the capabilities
+ of the local policy enforcer and the impact of the related state/event conditions
* VIMs should be able to execute most policies locally
* VIMs may need to pass policy-related state/events to a closed-loop system,
- where those events are relevant to other components in the architecture (e.g. service orchestrator),
- or some additional data/arbitration is needed to resolve the state/event condition
+ where those events are relevant to other components in the architecture
+ (e.g. service orchestrator), or some additional data/arbitration is needed
+ to resolve the state/event condition
* Policies are localized as they are distributed/delegated
- * High-level policies (e.g. expressing "intent") can be translated into VNF package elements or generic policies,
- perhaps using distinct syntaxes
- * Delegated policy syntaxes are likely VIM-specific, e.g. Datalog (Congress), YANG (ODL-based SDNC),
- or other schemas specific to other SDNCs (Contrail, ONOS)
+ * High-level policies (e.g. expressing "intent") can be translated into VNF
+ package elements or generic policies, perhaps using distinct syntaxes
+ * Delegated policy syntaxes are likely VIM-specific, e.g. Datalog (Congress)
* Closed-loop policy and VNF-lifecycle event handling are //somewhat// distinct
- * Closed-loop policy is mostly about resolving conditions that can't be handled locally, but as above in some cases
- the conditions may be of relevance and either delivered directly or forwarded to service orchestrators
- * VNF-lifecycle events that can't be handled by the VIM locally are delivered directly to the service orchestrator
+ * Closed-loop policy is mostly about resolving conditions that can't be
+ handled locally, but as above in some cases the conditions may be of
+ relevance and either delivered directly or forwarded to service orchestrators
+ * VNF-lifecycle events that can't be handled by the VIM locally are delivered
+ directly to the service orchestrator
- * Some events/analytics need to be collected into a more "open-loop" system which can enable other actions, e.g.
+ * Some events/analytics need to be collected into a more "open-loop" system
+ which can enable other actions, e.g.
* audits and manual interventions
* machine-learning focused optimizations of policies (largely a future objective)
-Issues to be investigated as part of establishing an overall cohesive/adaptive policy architecture:
+Issues to be investigated as part of establishing an overall cohesive/adaptive
+policy architecture:
- * For the various components which may fulfill a specific purpose, what capabilities (e.g. APIs) do they have/need to
+ * For the various components which may fulfill a specific purpose, what
+ capabilities (e.g. APIs) do they have/need to
* handle events locally
- * enable closed-loop policy handling components to subscribe/optimize policy-related events that are of interest
+ * enable closed-loop policy handling components to subscribe/optimize
+ policy-related events that are of interest
* For global controllers and cloud orchestrators
- * How do they support correlation of events impacting resources in different scopes (network and cloud)
+ * How do they support correlation of events impacting resources in different
+ scopes (network and cloud)
* What event/response flows apply to various policy use cases
* What specific policy use cases can/should fall into each overall class
* locally handled by NFVI components
- * handled by a closed-loop policy system, either VNF/service-specific or VNF-independent
+ * handled by a closed-loop policy system, either VNF/service-specific or
+ VNF-independent
diff --git a/docs/design/definitions.rst b/docs/design/definitions.rst
index 7f0628a..daf7a48 100644
--- a/docs/design/definitions.rst
+++ b/docs/design/definitions.rst
@@ -8,10 +8,13 @@ Definitions
- Meaning
* - State
- - Information that can be used to convey or imply the state of something, e.g. an application, resource, entity, etc. This can include data held inside OPNFV components, "events" that have occurred (e.g. "policy violation"), etc.
+ - Information that can be used to convey or imply the state of something,
+e.g. an application, resource, entity, etc. This can include data held inside
+OPNFV components, "events" that have occurred (e.g. "policy violation"), etc.
* - Event
- - An item of significance to the policy engine, for which the engine has become aware thr ough some method of discovery e.g. polling or notification.
+ - An item of significance to the policy engine, for which the engine has
+become aware through some method of discovery e.g. polling or notification.
Abbreviations
=============
diff --git a/docs/design/introduction.rst b/docs/design/introduction.rst
index a7cbb02..676adab 100644
--- a/docs/design/introduction.rst
+++ b/docs/design/introduction.rst
@@ -9,27 +9,29 @@ Introduction
.. NOTE::
This is the working documentation for the Copper project.
-The `OPNFV Copper <https://wiki.opnfv.org/copper>`_ project aims to help ensure that virtualized infrastructure
-deployments comply with goals of the VNF designer/user, e.g. re affinity and partitioning (e.g. per regulation,
-control/user plane separation, cost...).
-This is a "requirements" project with initial goal to assess "off the shelf" basic OPNFV platform support for policy
-management, using existing open source projects such as OpenStack Congress and OpenDaylight Group-Based Policy (GBP).
-The project will assess what policy-related features are currently supported through research into the related projects
-in OpenStack and ODL, and testing of integrated vanilla distributions of those and other dependent open source projects
-in the OPNFV's NFVI platform scope.
-
-Configuration Policy
---------------------
-
-As focused on by Copper, configuration policy helps ensure that the NFV service environment meets the requirements of
-the variety of stakeholders which will provide or use NFV platforms.
+The `OPNFV Copper <https://wiki.opnfv.org/copper>`_ project aims to help ensure
+that virtualized infrastructure and application deployments comply with goals of
+the NFV service provider or the VNF designer/user.
+
+This is the second ("Colorado") release of the Copper project. The documenation
+provided here focuses on the overall goals of the Copper project, and the
+specific features supported in the Colorado release.
+
+Overall Goals for Configuration Policy
+--------------------------------------
+
+As focused on by Copper, configuration policy helps ensure that the NFV service
+environment meets the requirements of the variety of stakeholders which will
+provide or use NFV platforms.
+
These requirements can be expressed as an *intent* of the stakeholder,
in specific terms or more abstractly, but at the highest level they express:
* what I want
* what I don't want
-Using road-based transportation as an analogy, some examples of this are shown below.
+Using road-based transportation as an analogy, some examples of this are shown
+below.
.. list-table:: Configuration Intent Example
:widths: 10 45 45
@@ -48,12 +50,17 @@ Using road-based transportation as an analogy, some examples of this are shown b
- shoulder warning strips, center media barriers
- speeding, tractors on the freeway
-According to their role, service providers may apply more specific configuration requirements than users,
-since service providers are more likely to be managing specific types of infrastructure capabilities.
+According to their role, service providers may apply more specific configuration
+requirements than users, since service providers are more likely to be managing
+specific types of infrastructure capabilities.
+
Developers and users may also express their requirements more specifically,
based upon the type of application or how the user intends to use it.
-For users, a high-level intent can be also translated into a more or less specific configuration capability
-by the service provider, taking into consideration aspects such as the type of application or its constraints.
+
+For users, a high-level intent can be also translated into a more or less specific
+configuration capability by the service provider, taking into consideration
+aspects such as the type of application or its constraints.
+
Examples of such translation are:
.. list-table:: Intent Translation into Configuration Capability
@@ -77,17 +84,29 @@ Examples of such translation are:
* - resource reclamation
- low-usage monitoring
-Although such intent to capability translation is conceptually useful, it is unclear how it can address the variety of
-aspects that may affect the choice of an applicable configuration capability.
-For that reason, the Copper project will initially focus on more specific configuration requirements as fulfilled by
-specific configuration capabilities, and how those requirements and capabilities are expressed in VNF and service
+Although such intent to capability translation is conceptually useful, it is
+unclear how it can address the variety of aspects that may affect the choice of
+an applicable configuration capability.
+
+For that reason, the Copper project will initially focus on more specific
+configuration requirements as fulfilled by specific configuration capabilities,
+and how those requirements and capabilities are expressed in VNF and service
design and packaging, or as generic poicies for the NFVI.
-Release 1 Scope
----------------
-OPNFV Brahmaputra will be the initial OPNFV release for Copper, with the goals:
- * Add the OpenStack Congress service to OPNFV, through at least one installer project
- * If possible, add Congress support to the OPNFV CI/CD pipeline for all Genesis project installers
- (Apex, Fuel, JOID, Compass)
- * Integrate Congress tests into Functest and develop additional use case tests for post-OPNFV-install
- * Extend with other OpenStack components for testing, as time permits
+Copper Release 1 Scope
+----------------------
+OPNFV Brahmaputra was the initial OPNFV release for Copper, and achieved the
+goals:
+ * Add the OpenStack Congress service to OPNFV, through at least one installer
+project, through post-install configuration.
+ * Provide basis tests scripts and tools to exercise the Congress service
+
+Copper Release 2 Scope
+----------------------
+OPNFV Colorado includes the additional features:
+ * Congress support in the the OPNFV CI/CD pipeline for the JOID and Apex
+installers, through the following projects being upstreamed to OpenStack:
+ * For JOID, a JuJu Charm for Congress
+ * For Apex, a Puppet Module for Congress
+ * Congress use case tests integrated into Functest and as manual tests
+ * Further enhancements of Congress test tools
diff --git a/docs/design/requirements.rst b/docs/design/requirements.rst
index a9b5bc0..a3f32d8 100644
--- a/docs/design/requirements.rst
+++ b/docs/design/requirements.rst
@@ -2,51 +2,77 @@ Requirements
============
This section outlines general requirements for configuration policies,
per the two main aspects in the Copper project scope:
- * Ensuring resource requirements of VNFs and services are applied per VNF designer, service, and tenant intent
+ * Ensuring resource requirements of VNFs and services are applied per VNF
+ designer, service, and tenant intent
* Ensuring that generic policies are not violated,
e.g. *networks connected to VMs must either be public or owned by the VM owner*
Resource Requirements
+++++++++++++++++++++
-Resource requirements describe the characteristics of virtual resources (compute, storage, network) that are needed for
-VNFs and services, and how those resources should be managed over the lifecycle of a VNF/service. Upstream projects
-already include multiple ways in which resource requirements can be expressed and fulfilled, e.g.:
+Resource requirements describe the characteristics of virtual resources (compute,
+storage, network) that are needed for VNFs and services, and how those resources
+should be managed over the lifecycle of a VNF/service. Upstream projects already
+include multiple ways in which resource requirements can be expressed and fulfilled, e.g.:
* OpenStack Nova
- * the `image <http://docs.openstack.org/openstack-ops/content/user_facing_images.html>`_ feature, enabling
- "VM templates" to be defined for NFs, and referenced by name as a specific NF version to be used
- * the `flavor <http://docs.openstack.org/openstack-ops/content/flavors.html>`_ feature, addressing basic compute
- and storage requirements, with extensibility for custom attributes
+ * the `image
+ <http://docs.openstack.org/openstack-ops/content/user_facing_images.html>`_
+ feature, enabling "VM templates" to be defined for NFs, and referenced by
+ name as a specific NF version to be used
+ * the `flavor <http://docs.openstack.org/openstack-ops/content/flavors.html>`_
+ feature, addressing basic compute and storage requirements, with
+ extensibility for custom attributes
* OpenStack Heat
- * the `Heat Orchestration Template <http://docs.openstack.org/developer/heat/template_guide/index.html>`_ feature,
- enabling a variety of VM aspects to be defined and managed by Heat throughout the VM lifecycle, notably
- * alarm handling (requires `Ceilometer <https://wiki.openstack.org/wiki/Ceilometer>`_)
- * attached volumes (requires `Cinder <https://wiki.openstack.org/wiki/Cinder>`_)
- * domain name assignment (requires `Designate <https://wiki.openstack.org/wiki/Designate>`_)
+ * the `Heat Orchestration Template
+ <http://docs.openstack.org/developer/heat/template_guide/index.html>`_
+ feature, enabling a variety of VM aspects to be defined and managed by
+ Heat throughout the VM lifecycle, notably
+ * alarm handling (requires
+ `Ceilometer <https://wiki.openstack.org/wiki/Ceilometer>`_)
+ * attached volumes (requires
+ `Cinder <https://wiki.openstack.org/wiki/Cinder>`_)
+ * domain name assignment (requires
+ `Designate <https://wiki.openstack.org/wiki/Designate>`_)
* images (requires `Glance <https://wiki.openstack.org/wiki/Glance>`_)
* autoscaling
- * software configuration associated with VM "lifecycle hooks (CREATE, UPDATE, SUSPEND, RESUME, DELETE"
+ * software configuration associated with VM "lifecycle hooks (CREATE,
+ UPDATE, SUSPEND, RESUME, DELETE"
* wait conditions and signaling for sequencing orchestration steps
- * orchestration service user management (requires `Keystone <http://docs.openstack.org/developer/keystone/>`_)
+ * orchestration service user management (requires
+ `Keystone <http://docs.openstack.org/developer/keystone/>`_)
* shared storage (requires `Manila <https://wiki.openstack.org/wiki/Manila>`_)
- * load balancing (requires Neutron `LBaaS <http://docs.openstack.org/admin-guide-cloud/content/section_lbaas-overview.html>`_)
- * firewalls (requires Neutron `FWaaS <http://docs.openstack.org/admin-guide-cloud/content/install_neutron-fwaas-agent.html>`_)
+ * load balancing (requires Neutron
+ `LBaaS
+ <http://docs.openstack.org/admin-guide-cloud/content/section_lbaas-overview.html>`_)
+ * firewalls (requires Neutron
+ `FWaaS
+ <http://docs.openstack.org/admin-guide-cloud/content/install_neutron-fwaas-agent.html>`_)
* various Neutron-based network and security configuration items
* Nova flavors
* Nova server attributes including access control
* Nova server group affinity and anti-affinity
- * "Data-intensive application clustering" (requires `Sahara <https://wiki.openstack.org/wiki/Sahara>`_)
+ * "Data-intensive application clustering" (requires
+ `Sahara <https://wiki.openstack.org/wiki/Sahara>`_)
* DBaaS (requires `Trove <http://docs.openstack.org/developer/trove/>`_)
- * "multi-tenant cloud messaging and notification service" (requires `Zaqar <http://docs.openstack.org/developer/zaqar/>`_)
- * OpenStack `Group-Based Policy <https://wiki.openstack.org/wiki/GroupBasedPolicy>`_
- * API-based grouping of endpoints with associated contractual expectations for data flow processing and
- service chaining
+ * "multi-tenant cloud messaging and notification service" (requires
+ `Zaqar <http://docs.openstack.org/developer/zaqar/>`_)
+ * OpenStack `Group-Based Policy
+ <https://wiki.openstack.org/wiki/GroupBasedPolicy>`_
+ * API-based grouping of endpoints with associated contractual expectations
+ for data flow processing and service chaining
* OpenStack `Tacker <https://wiki.openstack.org/wiki/Tacker>`_
- * "a fully functional ETSI MANO based general purpose NFV Orchestrator and VNF Manager for OpenStack"
- * OpenDaylight `Group-Based Policy <https://wiki.opendaylight.org/view/Group_Based_Policy_(GBP)>`_
- * model-based grouping of endpoints with associated contractual expectations for data flow processing
- * OpenDaylight `Service Function Chaining (SFC) <https://wiki.opendaylight.org/view/Service_Function_Chaining:Main>`_
- * model-based management of "service chains" and the infrastucture that enables them
- * Additional projects that are commonly used for configuration management, implemented as client-server frameworks using model-based, declarative, or scripted configuration management data.
+ * "a fully functional ETSI MANO based general purpose NFV Orchestrator and
+ VNF Manager for OpenStack"
+ * OpenDaylight `Group-Based Policy
+ <https://wiki.opendaylight.org/view/Group_Based_Policy_(GBP)>`_
+ * model-based grouping of endpoints with associated contractual expectations
+ for data flow processing
+ * OpenDaylight `Service Function Chaining (SFC)
+ <https://wiki.opendaylight.org/view/Service_Function_Chaining:Main>`_
+ * model-based management of "service chains" and the infrastucture that
+ enables them
+ * Additional projects that are commonly used for configuration management,
+ implemented as client-server frameworks using model-based, declarative, or
+ scripted configuration management data.
* `Puppet <https://puppetlabs.com/puppet/puppet-open-source>`_
* `Chef <https://www.chef.io/chef/>`_
* `Ansible <http://docs.ansible.com/ansible/index.html>`_
@@ -54,36 +80,59 @@ already include multiple ways in which resource requirements can be expressed an
Generic Policy Requirements
+++++++++++++++++++++++++++
-Generic policy requirements address conditions related to resource state and events which need to be monitored for,
-and optionally responded to or prevented. These conditions are typically expected to be VNF/service-independent,
-as VNF/service-dependent condition handling (e.g. scale in/out) are considered to be addressed by VNFM/NFVO/VIM
-functions as described under Resource Requirements or as FCAPS related functions. However the general capabilities
-below can be applied to VNF/service-specific policy handling as well, or in particular to invocation of
-VNF/service-specific management/orchestration actions. The high-level required capabilities include:
+Generic policy requirements address conditions related to resource state and
+events which need to be monitored for, and optionally responded to or prevented.
+These conditions are typically expected to be VNF/service-independent, as
+VNF/service-dependent condition handling (e.g. scale in/out) are considered to
+be addressed by VNFM/NFVO/VIM functions as described under Resource Requirements
+or as FCAPS related functions. However the general capabilities below can be
+applied to VNF/service-specific policy handling as well, or in particular to
+invocation of VNF/service-specific management/orchestration actions. The
+high-level required capabilities include:
* Polled monitoring: Exposure of state via request-response APIs.
* Notifications: Exposure of state via pub-sub APIs.
- * Realtime/near-realtime notifications: Notifications that occur in actual or near realtime.
- * Delegated policy: CRUD operations on policies that are distributed to specific components for local handling,
- including one/more of monitoring, violation reporting, and enforcement.
+ * Realtime/near-realtime notifications: Notifications that occur in actual or
+ near realtime.
+ * Delegated policy: CRUD operations on policies that are distributed to
+ specific components for local handling, including one/more of monitoring,
+ violation reporting, and enforcement.
* Violation reporting: Reporting of conditions that represent a policy violation.
- * Reactive enforcement: Enforcement actions taken in response to policy violation events.
- * Proactive enforcement: Enforcement actions taken in advance of policy violation events,
+ * Reactive enforcement: Enforcement actions taken in response to policy
+ violation events.
+ * Proactive enforcement: Enforcement actions taken in advance of policy
+ violation events,
e.g. blocking actions that could result in a policy violation.
* Compliance auditing: Periodic auditing of state against policies.
-Upstream projects already include multiple ways in which configuration conditions can be monitored and responded to:
- * OpenStack `Congress <https://wiki.openstack.org/wiki/Congress>`_ provides a table-based mechanism for state monitoring and proactive/reactive policy enforcement, including (as of the Kilo release) data obtained from internal databases of Nova, Neutron, Ceilometer, Cinder, Glance, Keystone, and Swift. The Congress design approach is also extensible to other VIMs (e.g. SDNCs) through development of data source drivers for the new monitored state information. See `Stackforge Congress Data Source Translators <https://github.com/stackforge/congress/tree/master/congress/datasources>`_, `congress.readthedocs.org <http://congress.readthedocs.org/en/latest/cloudservices.html#drivers>`_, and the `Congress specs <https://github.com/stackforge/congress-specs>`_ for more info.
- * OpenStack `Ceilometer <https://wiki.openstack.org/wiki/Ceilometer>`_ provides means to trigger alarms upon a wide variety of conditions derived from its monitored OpenStack analytics.
- * `Nagios <https://www.nagios.org/#/>`_ "offers complete monitoring and alerting for servers, switches, applications, and services".
+Upstream projects already include multiple ways in which configuration conditions
+can be monitored and responded to:
+ * OpenStack `Congress <https://wiki.openstack.org/wiki/Congress>`_ provides a
+ table-based mechanism for state monitoring and proactive/reactive policy
+ enforcement, including data obtained from internal databases of OpenStack
+ core and optional services. The Congress design approach is also extensible
+ to other VIMs (e.g. SDNCs) through development of data source drivers for
+ the new monitored state information. See `Stackforge Congress Data Source
+ Translators
+ <https://github.com/stackforge/congress/tree/master/congress/datasources>`_,
+ `congress.readthedocs.org
+ <http://congress.readthedocs.org/en/latest/cloudservices.html#drivers>`_,
+ and the `Congress specs <https://github.com/stackforge/congress-specs>`_ for
+ more info.
+ * OpenStack `Ceilometer <https://wiki.openstack.org/wiki/Ceilometer>`_
+ provides means to trigger alarms upon a wide variety of conditions derived
+ from its monitored OpenStack analytics.
+ * `Nagios <https://www.nagios.org/#/>`_ "offers complete monitoring and
+ alerting for servers, switches, applications, and services".
Requirements Validation Approach
++++++++++++++++++++++++++++++++
-The Copper project will assess the completeness of the upstream project solutions for requirements in scope though
-a process of:
+The Copper project will assess the completeness of the upstream project solutions
+for requirements in scope though a process of:
* developing configuration policy use cases to focus solution assessment tests
* integrating the projects into the OPNFV platform for testing
* executing functional and performance tests for the solutions
- * assessing overall requirements coverage and gaps in the most complete upstream solutions
+ * assessing overall requirements coverage and gaps in the most complete
+ upstream solutions
-Depending upon the priority of discovered gaps, new requirements will be submitted to upstream projects for the next
-available release cycle.
+Depending upon the priority of discovered gaps, new requirements will be
+submitted to upstream projects for the next available release cycle.
diff --git a/docs/design/usecases.rst b/docs/design/usecases.rst
index ef9e82d..ae046f3 100644
--- a/docs/design/usecases.rst
+++ b/docs/design/usecases.rst
@@ -1,21 +1,105 @@
Use Cases
=========
-Resource Requirements
-+++++++++++++++++++++
+Implemented as of this release
+------------------------------
-Workload Placement
-------------------
+DMZ Deployment
+..............
+
+As a service provider, I need to ensure that applications which have not been
+designed for exposure in a DMZ zone, are not attached to DMZ networks.
+
+An example implementation is shown in the Congress use case test "DMZ Placement"
+(dmz.sh) in the Copper repo under the tests folder. This test:
+ * Identifies VMs connected to a DMZ (currently identified through a
+ specifically-named security group)
+ * Identifes VMs connected to a DMZ, which are by policy not allowed to be
+ (currently implemented through an image tag intended to identify images
+ that are "authorized" i.e. tested and secure, to be DMZ-connected)
+ * Reactively enforces the dmz placement rule by pausing VMs found to be in
+ violation of the policy.
+
+As implemented through OpenStack Congress:
+
+.. code::
+
+ dmz_server(x) :-
+ nova:servers(id=x,status='ACTIVE'),
+ neutronv2:ports(id, device_id, status='ACTIVE'),
+ neutronv2:security_group_port_bindings(id, sg),
+ neutronv2:security_groups(sg,name='dmz')"
+
+ dmz_placement_error(id) :-
+ nova:servers(id,name,hostId,status,tenant_id,user_id,image,flavor,az,hh),
+ not glancev2:tags(image,'dmz'),
+ dmz_server(id)"
+
+ execute[nova:servers.pause(id)] :-
+ dmz_placement_error(id),
+ nova:servers(id,status='ACTIVE')"
+
+Configuration Auditing
+......................
+
+As a service provider or tenant, I need to periodically verify that resource
+configuration requirements have not been violated, as a backup means to proactive
+or reactive policy enforcement.
+
+An example implementation is shown in the Congress use case test "SMTP Ingress"
+(smtp_ingress.sh) in the Copper repo under the tests folder. This test:
+ * Detects that a VM is associated with a security group that allows SMTP
+ ingress (TCP port 25)
+ * Adds a policy table row entry for the VM, which can be later investigated
+ for appropriate use of the security group, etc
+
+As implemented through OpenStack Congress:
+
+.. code::
+
+ smtp_ingress(x) :-
+ nova:servers(id=x,status='ACTIVE'),
+ neutronv2:ports(port_id, status='ACTIVE'),
+ neutronv2:security_groups(sg, tenant_id, sgn, sgd),
+ neutronv2:security_group_port_bindings(port_id, sg),
+ neutronv2:security_group_rules(sg, rule_id, tenant_id, remote_group_id,
+ 'ingress', ethertype, 'tcp', port_range_min, port_range_max, remote_ip),
+ lt(port_range_min, 26),
+ gt(port_range_max, 24)
+
+Reserved Resources
+..................
+
+As an NFVI provider, I need to ensure that my admins do not inadvertently
+enable VMs to connect to reserved subnets.
+
+An example implementation is shown in the Congress use case test "Reserved
+Subnet" (reserved_subnet.sh) in the Copper repo under the tests folder. This
+test:
+ * Detects that a subnet has been created in a reserved range
+ * Reactively deletes the subnet
+
+As implemented through OpenStack Congress:
+
+.. code::
+
+ reserved_subnet_error(x) :-
+ neutronv2:subnets(id=x, cidr='10.7.1.0/24')
+
+ execute[neutronv2:delete_subnet(x)] :-
+ reserved_subnet_error(x)
+
+
+For further analysis and implementation
+---------------------------------------
Affinity
........
Ensures that the VM instance is launched "with affinity to" specific resources,
-e.g. within a compute or storage cluster.
-This is analogous to the affinity rules in
-`VMWare vSphere DRS <https://pubs.vmware.com/vsphere-50/topic/com.vmware.vsphere.resmgmt.doc_50/GUID-FF28F29C-8B67-4EFF-A2EF-63B3537E6934.html>`_.
-Examples include: "Same Host Filter", i.e. place on the same compute node as a given set of instances,
-e.g. as defined in a scheduler hint list.
+e.g. within a compute or storage cluster. Examples include: "Same Host Filter",
+i.e. place on the same compute node as a given set of instances, e.g. as defined
+in a scheduler hint list.
As implemented by OpenStack Heat using server groups:
@@ -48,10 +132,10 @@ Anti-Affinity
.............
Ensures that the VM instance is launched "with anti-affinity to" specific resources,
-e.g. outside a compute or storage cluster, or geographic location.
-This filter is analogous to the anti-affinity rules in vSphere DRS.
-Examples include: "Different Host Filter", i.e. ensures that the VM instance is launched
-on a different compute node from a given set of instances, as defined in a scheduler hint list.
+e.g. outside a compute or storage cluster, or geographic location. Examples
+include: "Different Host Filter", i.e. ensures that the VM instance is launched
+on a different compute node from a given set of instances, as defined in a
+scheduler hint list.
As implemented by OpenStack Heat using scheduler hints:
@@ -88,46 +172,27 @@ As implemented by OpenStack Heat using scheduler hints:
- network: {get_param: network}
scheduler_hints: {different_host: {get_resource: serv1}}
-DMZ Deployment
-..............
-As a service provider,
-I need to ensure that applications which have not been designed for exposure in a DMZ zone,
-are not attached to DMZ networks.
-
-Configuration Auditing
-----------------------
-
-As a service provider or tenant,
-I need to periodically verify that resource configuration requirements have not been violated,
-as a backup means to proactive or reactive policy enforcement.
-
-Generic Policy Requirements
-+++++++++++++++++++++++++++
-
-NFVI Self-Service Constraints
------------------------------
-
-As an NFVI provider,
-I need to ensure that my self-service tenants are not able to configure their VNFs in ways
-that would impact other tenants or the reliability, security, etc of the NFVI.
-
Network Access Control
......................
-Networks connected to VMs must be public, or owned by someone in the VM owner's group.
+Networks connected to VMs must be public, or owned by someone in the VM owner's
+group.
This use case captures the intent of the following sub-use-cases:
* Link Mirroring: As a troubleshooter,
- I need to mirror traffic from physical or virtual network ports so that I can investigate trouble reports.
+ I need to mirror traffic from physical or virtual network ports so that I
+ can investigate trouble reports.
* Link Mirroring: As a NFVaaS tenant,
- I need to be able to mirror traffic on my virtual network ports so that I can investigate trouble reports.
+ I need to be able to mirror traffic on my virtual network ports so that I
+ can investigate trouble reports.
* Unauthorized Link Mirroring Prevention: As a NFVaaS tenant,
- I need to be able to prevent other tenants from mirroring traffic on my virtual network ports
- so that I can protect the privacy of my service users.
+ I need to be able to prevent other tenants from mirroring traffic on my
+ virtual network ports so that I can protect the privacy of my service users.
* Link Mirroring Delegation: As a NFVaaS tenant,
- I need to be able to allow my NFVaaS SP customer support to mirror traffic on my virtual network ports
- so that they can assist in investigating trouble reports.
+ I need to be able to allow my NFVaaS SP customer support to mirror traffic
+ on my virtual network ports so that they can assist in investigating trouble
+ reports.
As implemented through OpenStack Congress:
@@ -172,18 +237,17 @@ As implemented through OpenStack Congress:
ldap:group(user1, g),
ldap:group(user2, g)
-Resource Management
--------------------
-
Resource Reclamation
....................
-As a service provider or tenant,
-I need to be informed of VMs that are under-utilized so that I can reclaim the VI resources.
-(example from `RuleYourCloud blog <http://ruleyourcloud.com/2015/03/12/scaling-up-congress.html>`_)
+As a service provider or tenant, I need to be informed of VMs that are
+under-utilized so that I can reclaim the VI resources. (example from
+`RuleYourCloud blog <http://ruleyourcloud.com/2015/03/12/scaling-up-congress.html>`_)
As implemented through OpenStack Congress:
+*Note: untested example...*
+
.. code::
reclaim_server(vm) :-
@@ -198,11 +262,13 @@ As implemented through OpenStack Congress:
Resource Use Limits
...................
-As a tenant or service provider,
-I need to be automatically terminate an instance that has run for a pre-agreed maximum duration.
+As a tenant or service provider, I need to be automatically terminate an
+instance that has run for a pre-agreed maximum duration.
As implemented through OpenStack Congress:
+*Note: untested example...*
+
.. code::
terminate_server(vm) :-
@@ -214,4 +280,3 @@ As implemented through OpenStack Congress:
nova:servers(vm, vm_name, user_id),
keystone:users(user_id, email)
-