1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
|
.. This work is licensed under a Creative Commons Attribution 4.0 International License.
.. SPDX-License-Identifier: CC-BY-4.0
***********************
Testing developer guide
***********************
.. toctree::
:numbered:
:maxdepth: 2
============
Introduction
============
The OPNFV testing ecosystem is wide.
The goal of this guide consists in providing some guidelines for new developers
involved in test areas.
For the description of the ecosystem, see `[1]`_.
=================
Developer journey
=================
Be involved in the testing group
================================
Best practices
==============
Unit tests
----------
Dockerization
-------------
API
---
CLI
---
Traffic generators
------------------
Towards a pseudo micro services approach
----------------------------------------
======================================
Testing group configuration parameters
======================================
Testing categories
==================
The testing group defined several categories also known as tiers. These
categories can be used to group test suites.
+----------------+-------------------------------------------------------------+
| Healthcheck | Simple and quick healthcheck tests case |
+----------------+-------------------------------------------------------------+
| Smoke | Set of smoke test cases/suites to validate the release |
+----------------+-------------------------------------------------------------+
| Features | Test cases that validate a specific feature on top of OPNFV.|
| | Those come from Feature projects and need a bit of support |
| | for integration |
+----------------+-------------------------------------------------------------+
| Components | Tests on a specific component (e.g. OpenStack, OVS, DPDK,..)|
| | It may extend smoke tests |
+----------------+-------------------------------------------------------------+
| Performance | Performance qualification |
+----------------+-------------------------------------------------------------+
| VNF | Test cases related to deploy an open source VNF including |
| | an orchestrator |
+----------------+-------------------------------------------------------------+
| Stress | Stress and robustness tests |
+----------------+-------------------------------------------------------------+
| In Service | In service testing |
+----------------+-------------------------------------------------------------+
Testing domains
===============
The domains deal with the technical scope of the tests. It shall correspond to
domains defined for the certification program:
* compute
* network
* storage
* hypervisor
* container
* vim
* mano
* vnf
* ...
Testing coverage
=================
One of the goals of the testing working group is to identify the poorly covered
areas and avoid testing overlap.
Ideally based on the declaration of the test cases, through the tags, domains
and tier fields, it shall be possible to create heuristic maps.
==============================
Testing group generic enablers
==============================
TestAPI framework
=================
The OPNFV testing group created a test collection database to collect
the test results from CI:
http://testresults.opnfv.org/test/swagger/spec.html
Any test project running on any lab integrated in CI can push the
results to this database.
This database can be used to see the evolution of the tests and compare
the results versus the installers, the scenarios or the labs.
It is used to produce a dashboard with the current test status of the project.
Overall Architecture
--------------------
The Test result management can be summarized as follows::
+-------------+ +-------------+ +-------------+
| | | | | |
| Test | | Test | | Test |
| Project #1 | | Project #2 | | Project #N |
| | | | | |
+-------------+ +-------------+ +-------------+
| | |
V V V
+-----------------------------------------+
| |
| Test Rest API front end |
| |
+-----------------------------------------+
A |
| V
| +-------------------------+
| | |
| | Test Results DB |
| | Mongo DB |
| | |
| +-------------------------+
|
|
+----------------------+
| |
| test Dashboard |
| |
+----------------------+
TestAPI description
-------------------
The TestAPI is used to declare pods, projects, test cases and test
results. Pods are the sets of bare metal or virtual servers and networking
equipments used to run the tests.
The results pushed in the database are related to pods, projects and test cases.
If you try to push results of test done on non referenced pod, the API will
return an error message.
An additional method dashboard has been added to post-process
the raw results in release Brahmaputra (deprecated in Colorado).
The data model is very basic, 5 objects are created:
* Pods
* Projects
* Testcases
* Results
* Scenarios
The code of the API is hosted in the releng repository `[6]`_.
The static documentation of the API can be found at `[7]`_.
The TestAPI has been dockerized and may be installed locally in your
lab. See `[15]`_ for details.
The deployment of the TestAPI has been automated.
A jenkins job manages:
* the unit tests of the TestAPI
* the creation of a new docker file
* the deployment of the new TestAPI
* the archive of the old TestAPI
* the backup of the Mongo DB
TestAPI Authorization
~~~~~~~~~~~~~~~~~~~~~
PUT/DELETE/POST operations of the TestAPI now require token based authorization. The token needs
to be added in the request using a header 'X-Auth-Token' for access to the database.
e.g::
headers['X-Auth-Token']
The value of the header i.e the token can be accessed in the jenkins environment variable
*TestApiToken*. The token value is added as a masked password.
.. code-block:: python
headers['X-Auth-Token'] = os.environ.get('TestApiToken')
The above example is in Python. Token based authentication has been added so that only ci pods
jenkins job can have access to the database.
Please note that currently token authorization is implemented but is not yet enabled.
Reporting
=========
An automatic reporting page has been created in order to provide a
consistent view of the scenarios.
In this page, each scenario is evaluated according to test criteria.
The code for the automatic reporting is available at `[8]`_.
The results are collected from the centralized database every day and,
per scenario. A score is calculated based on the results from the last
10 days.
Dashboard
=========
Dashboard is used to provide a consistent view of the results collected in CI.
The results showed on the dashboard are post processed from the Database,
which only contains raw results.
It can be used in addition of the reporting page (high level view) to allow
the creation of specific graphs according to what the test owner wants to show.
In Brahmaputra, a basic home made dashboard was created in Functest.
In Colorado, Yardstick adopted Grafana (time based graphs) and ELK (complex
graphs).
Since Danube, the testing community decided to adopt ELK framework and to rely
on bitergia. It was not implemented for Danube but it is planned for Euphrates.
Bitergia already provides a dashboard for code and infrastructure.
A new Test tab will be added. The dataset will be built by consuming
the TestAPI.
See `[3]`_ for details.
=======
How TOs
=======
Where can I find information on the different test projects?
===========================================================
How can I contribute to a test project?
=======================================
Where can I find hardware resources?
====================================
How do I integrate my tests in CI?
==================================
How to declare my tests in the test Database?
=============================================
How to push your results into the Test Database?
================================================
The test database is used to collect test results. By default it is
enabled only for CI tests from Production CI pods.
Please note that it is possible to create your own local database.
A dedicated database is for instance created for each plugfest.
The architecture and associated API is described in previous chapter.
If you want to push your results from CI, you just have to call the API
at the end of your script.
You can also reuse a python function defined in functest_utils.py `[5]`_
Where can I find the documentation on the test API?
===================================================
http://artifacts.opnfv.org/releng/docs/testapi.html
I have tests, to which category should I declare them?
======================================================
The main ambiguity could be between features and VNF.
In fact sometimes you have to spawn VMs to demonstrate the capabilities of the
feature you introduced.
We recommend to declare your test in the feature category.
VNF category is really dedicated to test including:
* creation of resources
* deployement of an orchestrator/VNFM
* deployment of the VNF
* test of the VNFM
* free resources
The goal is not to study a particular feature on the infrastructure but to have
a whole end to end test of a VNF automatically deployed in CI.
Moreover VNF are run in weekly jobs (one a week), feature tests are in daily
jobs and use to get a scenario score.
Where are the logs of CI runs?
==============================
Logs and configuration files can be pushed to artifact server from the CI under
http://artifacts.opnfv.org/<project name>
==========
References
==========
_`[1]`: http://docs.opnfv.org/en/stable-danube/testing/ecosystem/overview.html
_`[2]`: http://www.opnfv.org
_`[3]`: https://wiki.opnfv.org/display/testing/Result+alignment+for+ELK+post-processing
_`[4]`: https://wiki.opnfv.org/display/INF/CI+Scenario+Naming
_`[5]`: https://git.opnfv.org/functest/tree/functest/utils/functest_utils.py#176
_`[6]`: https://git.opnfv.org/functest/tree/releng
_`[7]`: http://artifacts.opnfv.org/releng/docs/testapi.html
IRC support chan: #opnfv-testperf
|