diff options
author | Morgan Richomme <morgan.richomme@orange.com> | 2016-07-08 08:29:51 +0000 |
---|---|---|
committer | Gerrit Code Review <gerrit@172.30.200.206> | 2016-07-08 08:29:51 +0000 |
commit | 7e650ef91b3f1a121ce92e8e9263b04c3a3e2008 (patch) | |
tree | 0b84a28c6fe279ef4ccd0b2ed9ac346125a18f2e /docs/devguide | |
parent | 19c88ef768e167465bc1c834e25e0f4cbac873b6 (diff) | |
parent | ed5bd81651d7af42605ece8d70235b7283c5a655 (diff) |
Merge "Update developer guide documentation"
Diffstat (limited to 'docs/devguide')
-rw-r--r-- | docs/devguide/index.rst | 788 |
1 files changed, 504 insertions, 284 deletions
diff --git a/docs/devguide/index.rst b/docs/devguide/index.rst index da134bc76..88ae232bc 100644 --- a/docs/devguide/index.rst +++ b/docs/devguide/index.rst @@ -11,11 +11,31 @@ OPNFV FUNCTEST developer guide Introduction ============ -This document describes how feature projects aiming to run functional tests can -be integrated into FuncTest framework. +Functest is a project dealing with functional testing. +Functest produces its own internal test cases but can also be considered +as a framework to support feature project testing suite integration. +Functest developed a test API and defined a Test collection framework +that can be used by any OPNFV project. + +Therefore there are many ways to contribute to Functest. You can: + + * develop new internal test cases + * integrate the tests from your feature project + * develop the framework to ease the integration of external test cases + * develop the API / Test collection framework + * develop dashboards or automatic reporting portals + +This document describes how, as a developer, you may interact with the +Functest project. The first section details the main working areas of +the project. The Second part is a list of "How to" to help you to join +the Functest family whatever your field of interest is. + + +======================== +Functest developer areas +======================== -================================ Functest High level architecture ================================ @@ -38,266 +58,85 @@ Functest can be described as follow:: | | +----------------------+ -Functest deals with internal and external test cases. -The Internal test cases in Brahmaputra are: - - * vPing_SSH - * vPing_userdata - * ODL - * Tempest - * vIMS - * Rally - +Functest internal test cases +============================ +The internal test cases in Colorado are: + + * vping_ssh + * vping_userdata + * odl + * tempest_smoke_serial + * rally_sanity + * tempest_full_parallel + * rally_full + * vims + +By internal, we mean that this particular test cases have been +developped by functest contributors and the associated code is hosted in +the Functest repository. +The structure of this repository is detailed in `[1]`_. +The main internal test cases are int the testcases subfolder of the +repository, the internal test cases are: + + * Controllers: odl, onos + * OpenStack: vping_ssh, vping_userdata, tempest_*, rally_* + * vIMS: vims + +If you want to create a new test cases you will have to create a new +folder under the testcases directory + +Functest external test cases +============================ The external tescases are: - * Promise - * Doctor - * Onos - * BGPVPN - -see `[2]`_ for details. - -Functest can also be considered as a framework that may include external OPNFV -projects. -This framework will ease the integration of the feature test suite to the CI. - -================== -How Functest works -================== - -The installation and the launch of the Functest docker image is described in -`[1]`_. - -The Functest docker directories are:: - - home - | - `-- opnfv - |-- functest - | |-- conf - | |-- data - | `-- results - `-- repos - |-- bgpvpn - |-- doctor - |-- functest - |-- odl_integration - |-- onos - |-- ovno - |-- promise - |-- rally - |-- releng - `-- vims-test - -:: - - +-----------+-------------------+---------------------------------------------+ - | Directory | Subdirectory | Comments | - +-----------+-------------------+---------------------------------------------+ - | | <project>/conf | All the useful configuration file(s) for | - | | | <project> the openstack creds are put there | - | | | for CI | - | | | It is recommended to push it there when | - | | | passing the credentials to container through| - | | | the -v option | - | +-------------------+---------------------------------------------+ - | opnfv | <project>/data | Usefull data, images for <projects> | - | | | By default we put a cirros image: | - | | | cirros-0.3.4-x86_64-disk.img | - | | | This image can be used by any projects | - | +-------------------+---------------------------------------------+ - | | <project>/results | Local result directory of project <project> | - +-----------+-------------------+---------------------------------------------+ - | | bgpvpn | | - | +-------------------+ + - | repos | doctor | | - | +-------------------+ + - | | functest | | - | +-------------------+ + - | | odl_integration | | - | +-------------------+ Clone of the useful repositories + - | | onos | These repositories may include: | - | +-------------------+ - tooling + - | | promise | - scenario | - | +-------------------+ - scripts + - | | rally | | - | +-------------------+ + - | | releng | | - | +-------------------+ + - | | vims-test | | - | +-------------------+ + - | | <your project> | | - +-----------+-------------------+---------------------------------------------+ - -Before running the test suite, you must prepare the environement by running:: - - $ /home/opnfv/repos/functest/docker/prepare_env.sh - -By running prepare_env.sh, you build the test environement required by the tests -including the retrieval and sourcing of OpenStack credentials. -This is an example of the env variables we have in the docker container: - - * HOSTNAME=373f77816eb0 - * INSTALLER_TYPE=fuel - * repos_dir=/home/opnfv/repos - * INSTALLER_IP=10.20.0.2 - * PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin - * PWD=/home/opnfv - * SHLVL=1 - * HOME=/home/opnfv - * NODE_NAME=opnfv-jump-2 - * creds=/home/opnfv/functest/conf/openstack.creds - * _=/usr/bin/env - -The prepare_env.sh will source the credentials, so once run you should have -access to the following env variables:: - - root@373f77816eb0:~# env|grep OS_ - OS_REGION_NAME=RegionOne - OS_PROJECT_NAME=admin - OS_PASSWORD=admin - OS_AUTH_STRATEGY=keystone - OS_AUTH_URL=http://172.30.10.70:5000/v2.0 - OS_USERNAME=admin - OS_TENANT_NAME=admin - OS_ENDPOINT_TYPE=internalURL - OS_NO_CACHE=true - - -Then you may run the test suite by running:: - - $ /home/opnfv/repos/functest/docker/run_tests.sh -t <your project> - -see `[2]`_ for details. - - - -========================= -How to integrate Functest -========================= - -The files of the Functest repository you must modify to integrate Functest are: - - * functest/docker/Dockerfile - * functest/docker/common.sh - * functest/docker/requirements.pip - * functest/docker/run_tests.sh - * functest/docker/prepare_env.sh - * functest/config_funtest.yaml - - -Dockerfile -========== - -This file lists the repositories to be cloned in the Functest container. -The repositories can be internal or external:: - - RUN git clone https://gerrit.opnfv.org/gerrit/<your porject> ${repos_dir}/<your project> - - -Common.sh -========== - -This file can be used to declare the branch and commit variables of your -projects:: - - <YOUR_PROJECT>_BRANCH=$(cat $config_file | grep -w <your project>_branch | awk 'END {print $NF}') - <YOUR_PROJECT>_COMMIT=$(cat $config_file | grep -w <your project>_commit | awk 'END {print $NF}') - - echo "<YOUR_PROJECT>_BRANCH=${<YOUR_PROJECT>_BRANCH}" - echo "<YOUR_PROJECT>_COMMIT=${<YOUR_PROJECT>_COMMIT}" - - -requirements.pip -================ - -This file can be used to preloaded all the needed Python libraries (and avoid -that each project does it) -The current libraries used in Functest container are:: - - # cat requirements.pip - pyyaml==3.10 - gitpython==1.0.1 - python-neutronclient==2.6.0 - python-novaclient==2.28.1 - python-glanceclient==1.1.0 - python-cinderclient==1.4.0 - python-ceilometerclient==1.5.1 - python-keystoneclient==1.6.0 - virtualenv==1.11.4 - pexpect==4.0 - requests==2.8.0 - robotframework==2.9.1 - robotframework-requests==0.3.8 - robotframework-sshlibrary==2.1.1 - configObj==5.0.6 - Flask==0.10.1 - xmltodict==0.9.2 - scp==0.10.2 - paramiko==1.16.0 - - -prepare_env.sh -============== + * promise + * doctor + * onos + * bgpvpn + * copper + * moon + * security_scan + * sfc-odl + * sfc-onos -This script can be adapted if you need to set up a specific environment before -running the tests. -run_tests.sh -============ - -This script is used to run the tests. You must thus complete the cases with your -own project:: - - ;; - "promise") - info "Running PROMISE test case..." - # TODO - ;; - "doctor") - info "Running Doctor test..." - python ${FUNCTEST_REPO_DIR}/testcases/features/doctor.py - ;; - "<your project>") - info "Running <your project> test..." - # your script that launchs your tests... - ;; +Note that security_scan has been bootstraped in Functest but will be +considered as an external test cases as soon as it will get its own +repository. -And do not forget to update also the help line:: +The code to run these test cases may be directly in the repository of +the project. We have also a features sub directory under testcases +directory that may be used (it can be usefull if you want to reuse +Functest library). - -t|--test run specific set of tests - <test_name> one or more of the following separated by comma: - vping_ssh,vping_userdata,odl,rally,tempest,vims,onos,promise,ovno +Functest framework +================== -config_funtest.yaml -=================== +Functest can be considered as a framework. +Functest is release a a docker file, including tools, scripts and a CLI +to prepare the environement and run tests. +It simplifies the integration of external test suite in CI pipeline +and provide commodity tools to collect and display results. -This file is the main configuration file of Functest. You must add the -references to your project:: +Since Colorado, test categories also known as tiers have been created to +group similar tests, provide consistant sub-lists and at the end optimize +test duration for CI (see How To section). - general: - directories: - dir_repo_<your project>: /home/opnfv/repos/<your project> - repositories: - # branch and commit ID to which the repos will be reset (HEAD) - <your project>_branch: master - <your project>_commit: latest +see `[2]`_ for details. +Test API +======== -==================== -Test Dashboard & API -==================== +The OPNFV testing group created a test collection database to collect +the test results from CI. +Any test project running on any lab integrated in CI can push the +results to this database. +This database can be used afterwards to see the evolution of the tests +and compare the results versus the installers, the scenario or the labs. -The OPNFV testing group created a test collection database to collect the test -results from CI. -Any test project running on any lab integrated in CI can push the results to -this database. -This database can be used afterwards to see the evolution of the tests and -compare the results versus the installers, the scenario or the labs. - -You can find more information about the dashboard from Testing Dashboard wiki -page `[3]`_. +You can find more information about the dashboard from Testing Dashboard +wiki page `[3]`_. Overall Architecture ==================== @@ -335,8 +174,10 @@ The Test result management can be summarized as follows:: Test API description ==================== -The Test API is used to declare pods, projects, test cases and test results. An -additional method dashboard has been added to post-process the raw results in release Brahmaputra. +The Test API is used to declare pods, projects, test cases and test +results. An additional method dashboard has been added to post-process +the raw results in release Brahmaputra. + The data model is very basic, 4 objects are created: * Pods @@ -588,11 +429,16 @@ Dashboard: | GET |/dashboard/v1/projects? | Get all the dashboard ready projects | +--------+----------------------------+-----------------------------------------+ -Dashboard description -===================== +The code of the API is hosted in the releng repository `[6]`_. + +Dashboard +========= +Dashboard is used to provide a consistant view of the results collected +in CI. The results with dashboard method are post-processed from raw results. -Please note that dashboard results are not stored. Only raw results are stored. +Please note that dashboard results are not stored. Only raw results are +stored. Release Brahmaputra ------------------- @@ -601,43 +447,409 @@ Dashboard url: http://testresults.opnfv.org/dashboard/ Release Colorado ---------------- +Since Colorado, it was decided to adopt ELK framework. Mongo DB results +are extracted to feed Elasticsearch database (`[7]`_). Dashboard url: http://testresults.opnfv.org/kibana_dashboards/ Credentials for a guest account: opnfvuser/kibana -=============================================== + +Automatic reporting +=================== + +OPNFV release is scenario centric. An automatic reporting page has been +created in order to provide a consistant view of the scenarios. +In this page each scenario is evaluated according to test criteria. +The code for the automatic reporting is available at `[8]`_. + +Every day, we collect the results from the centralized database and, per +scenario we calculate a score. This score is the addition of individual +tests considered on the last 10 runs according to a defined criteria. + +If we consider for instance a scenario os-odl_l2-nofeature-ha, we will +consider for the scoring all the runnable tests from the first test +categories (tiers healthcheck, smoke, controller and feature) +corresponding to this scenario. + + + +---------------------+---------+---------+---------+---------+ + | Test | Apex | Compass | Fuel | Joid | + +=====================+=========+=========+=========+=========+ + | vPing_ssh | X | X | X | X | + +---------------------+---------+---------+---------+---------+ + | vPing_userdata | X | X | X | X | + +---------------------+---------+---------+---------+---------+ + | tempest_smoke_serial| X | X | X | X | + +---------------------+---------+---------+---------+---------+ + | rally_sanity | X | X | X | X | + +---------------------+---------+---------+---------+---------+ + | odl | X | X | X | X | + +---------------------+---------+---------+---------+---------+ + | promise | | | X | X | + +---------------------+---------+---------+---------+---------+ + | security_scan | X | | | | + +---------------------+---------+---------+---------+---------+ + | copper | X | | | X | + +---------------------+---------+---------+---------+---------+ + +All the testcases listed in the table are runnable on +os-odl_l2-nofeature scenarios. +If no results are available or if all the results are failed, the test +case get 0 point. +If it was succesfull at least once but no anymore during the 4 runs, +the case get 1 point (it worked once). +If at least 3 of the last 4 runs were successful, the case get 2 points. +If the last 4 runs of the test are successful, the test get 3 points. + +In the example above, the target score for fuel/os-odl_l2-nofeature-ha +is 3x6 = 18 points. + +The scenario is validated per installer when we got 3 points for all +individual test cases (e.g 18/18). +Please note that complex or long duration tests are not considered for +the scoring. Such cases will be also get points, but these points will +be indicative and not used for the scenario validation. + +A web page is automatically generated every day to display the status. +This page can be found at `[9]`_. For the status, click on Status menu, +you may also get feedback for vims and tempest_smoke_serial test cases. + +Any validated scenario is stored in a local file on the web server. In +fact as we are using a sliding windows to get results, it may happen +that a successfull scenarios is not more run (because considered as +stable)and then the number of iterations (need 4) would not be +sufficient to get the green status. + +Please note that other test cases e.g. sfc_odl, bgpvpn, moon) need also +ODL configuration addons and as a consequence specific scenario. +There are not considered as runnable on the generic odl_l2 scenario. + +======= +How TOs +======= + +How Functest works? +=================== + +The installation and configuration of the Functest docker image is +described in `[1]`_. + +The procedure to start tests is described in `[2]`_ + + +How can I contribute to Functest? +================================= + +If you are already a contributor of any OPNFV project, you can +contribute to functest. If you are totally new to OPNFV, you must first +create your Linux Foundation account, then contact us in order to +declare you in the repository database. + +We distinguish 2 levels of contributors: the standard contributor can +push patch and vote +1/0/-1 on any Functest patch. The commitor can vote +-2/-1/0/+1/+2 and merge. +Functest commitors are promoted by the Functest contributors. + + +Where can I find some help to start? +==================================== + +This guide is made for you. You can also have a look at the project wiki +page `[10]`_. +There are references on documentation, video tutorials, tips... + +You can also directly contact us by mail with [Functest] prefix at +opnfv-tech-discuss@lists.opnfv.org or on the IRC chan #opnfv-functest. + + +What kind of testing do you do in Functest? +=========================================== + +Functest is focusing on Functional testing. The results must be PASS or +FAILED. We do not deal with performance and/or qualification tests. +We consider OPNFV as a black box and execute our tests from the jumphost +according to Pharos reference technical architecture. + +Upstream test suites are integrated (Rally/Tempest/ODL/ONOS,...). +If needed Functest may bootstrap temporarily testing activities if they +are identified but not covered yet by an existing testing project (e.g +security_scan before the creation of the security reporistory) + + +How test constraints are defined? +================================= + +Test constraints are defined according to 2 paramaters: + + * the scenario (DEPLOY_SCENARIO env variable) + * the installer (INSTALLER_TYPE env variable) + +A scenario is a formal description of the system under test. +The rules to define a scenario are described in `[4]`_ + +These 2 constraints are considered to determinate if the test is runnable +or not (e.g. no need to run onos suite on odl scenario). + +In the test declaration for CI, the test owner shall indicate these 2 +constraints. The file testcases.yaml `[5]`_ must be patched in git to +include new test cases. A more elaborated system based on template is +planned for next releases + +For each dependency, it is possible to define a regex:: + + name: promise + criteria: 'success_rate == 100%' + description: >- + Test suite from Promise project. + dependencies: + installer: '(fuel)|(joid)' + scenario: '' + +In the example above, it means that promise test will be runnable only +with joid or fuel installers on any scenario. + +The vims criteria means any installer and exclude onos and odl with +bgpvpn scenarios:: + + name: vims + criteria: 'status == "PASS"' + description: >- + This test case deploys an OpenSource vIMS solution from Clearwater + using the Cloudify orchestrator. It also runs some signaling traffic. + dependencies: + installer: '' + scenario: '(ocl)|(nosdn)|^(os-odl)((?!bgpvpn).)*$' + + + + +How to write and check constaint regex? +======================================= + +Regex are standard regex. You can have a look at `[11]`_ + +You caa also easily test your regex via an online regex checker such as + `[12]`_. +Put your scenario in the TEST STRING window (e.g. os-odl_l3-ovs-ha), put +your regex in the REGULAR EXPRESSION window, then you can test your rule +. + + +How to know which test I can run? +================================= + +You can use the API `[13]`_. The static declaration is in git `[5]`_ + +If you are in a Functest docker container (assuming that the +environement has been prepared): just use the new CLI. + +You can get the list per Test cases or by Tier:: + + # functest testcase list + healthcheck + vping_ssh + vping_userdata + tempest_smoke_serial + rally_sanity + odl + doctor + security_scan + tempest_full_parallel + rally_full + vims + # functest tier list + - 0. healthcheck: + ['healthcheck'] + - 1. smoke: + ['vping_ssh', 'vping_userdata', 'tempest_smoke_serial', 'rally_sanity'] + - 2. sdn_suites: + ['odl'] + - 3. features: + ['doctor', 'security_scan'] + - 4. openstack: + ['tempest_full_parallel', 'rally_full'] + - 5. vnf: + ['vims'] + + +How to manually start Functest test? +==================================== + +Assuming that you are connected on the jumhost and that the system is +"Pharos compliant", i.e the technical architecture is compatible with +the one defined in the Pharos project:: + + # docker pull opnfv/functest:latest + # envs="-e INSTALLER_TYPE=fuel -e INSTALLER_IP=10.20.0.2 -e DEPLOY_SCENARIO=os-odl_l2-nofeature-ha -e CI_DEBUG=true" + # sudo docker run --privileged=true -id ${envs} opnfv/functest:latest /bin/bash + + +Then you must connect to the docker container and source the +credentials:: + + # docker ps (copy the id) + # docker exec -ti <container_id> bash + # source $creds + + +You must first check if the environment is ready:: + + # functest env status + Functest environment ready to run tests. + + +If not ready, prepare the env by launching:: + + # functest env prepare + Functest environment ready to run tests. + +Once the Functest env is ready, you can use the CLI to start tests. + +You can run test cases per test case or per tier: + # functest testcase run <case name> or # functest tier run <tier name> + + +e.g:: + + # functest testcase run tempest_smoke_serial + # functest tier run features + + +If you want to run all the tests you can type:: + + # functest testcase run all + + +If you want to run all the tiers (same at the end that running all the +test cases) you can type:: + + # functest tier run all + + +How to declare my tests in Functest? +==================================== + +If you want to add new internal test cases, you can submit patch under +the testcases directory of Functest reporsitory. + +For feature test integration, the code can be kept into your own +repository. Then the files of the Functest repository you must modify to +integrate Functest are: + + * functest/docker/Dockerfile: get your code in Functest container + * functest/ci/testcases.yaml: reference your test and its associated constraints + * functest/ci/exec_test.sh: run your test from CI and CLI + + +Dockerfile +---------- + +This file lists the repositories to be cloned in the Functest container. +The repositories can be internal or external:: + + RUN git clone https://gerrit.opnfv.org/gerrit/<your project> ${repos_dir}/<your project> + +testcases.yaml +-------------- + +All the test cases that must be run from CI / CLI must be declared in +co/testcases.yaml. +This file is used to get the constraintes related to the test:: + + name: <my_super_test_case> + criteria: <not used yet in Colorado, could be > 'PASS', 'rate > 90%' + description: >- + <the description of your super test suite> + dependencies: + installer: regex related to installer e.g. 'fuel', '(apex)||(joid)' + scenario: regex related to the scenario e.g. 'ovs*no-ha' + + +You must declare your test case in one of the category (tier). + +If you are integrating test suites from a feature project, the category +must be features. + + +exec_test.sh +------------ + +This file is used to start your test. It is used in CI but also by the +CLI. + +You just patch the file in git and add a line:: + + + "<my_super_test_case>") + python ${FUNCTEST_REPO_DIR}/testcases/features/mycase.py + ;; + + +Note you can use python or bash scripts. + + +How to select my list of tests for CI? +====================================== + +Functest can be run automatically from CI, a jenkins job is usually +called after an OPNFV fresh installation. +By default we try to run all the possible tests (see `[14]` called from +Functest jenkins job):: + + cmd="python ${FUNCTEST_REPO_DIR}/ci/run_tests.py -t all ${flags}" + + +Each case can be configured as daily and/or weekly task. +When executing run_tests.py a check based on the jenkins build tag will +be considered to detect whether it is a daily and/or a weekly test. + +in your CI you can customize the list of test you want to run by case or +by tier, just change the line:: + + cmd="python ${FUNCTEST_REPO_DIR}/ci/run_tests.py -t <whatever you want> ${flags}" + +e.g.:: + + cmd="python ${FUNCTEST_REPO_DIR}/ci/run_tests.py -t healthcheck,smoke ${flags}" + +This command will run all the test cases of the first 2 tiers, i.e. +healthcheck, vping_ssh, vping_userdata, tempest_smoke_serial and +rally_sanity. + + How to push your results into the Test Database =============================================== -The test database is used to collect test results. By default it is enabled only -in Continuous Integration. -The architecture and associated API is described in `[2]`_. -If you want to push your results from CI, you just have to use the API at the -end of your script. +The test database is used to collect test results. By default it is +enabled only for CI tests from Production CI pods. + +The architecture and associated API is described in previous chapter. +If you want to push your results from CI, you just have to use the API +at the end of your script. You can also reuse a python function defined in functest_utils.py:: - def push_results_to_db(db_url, case_name, logger, pod_name,version, payload): - """ - POST results to the Result target DB - """ - url = db_url + "/results" - installer = get_installer_type(logger) - params = {"project_name": "functest", "case_name": case_name, - "pod_name": pod_name, "installer": installer, - "version": version, "details": payload} - - headers = {'Content-Type': 'application/json'} - try: - r = requests.post(url, data=json.dumps(params), headers=headers) - if logger: - logger.debug(r) - return True - except Exception, e: - print "Error [push_results_to_db('%s', '%s', '%s', '%s', '%s')]:" \ - % (db_url, case_name, pod_name, version, payload), e - return False + def push_results_to_db(db_url, case_name, logger, pod_name,version, payload): + """ + POST results to the Result target DB + """ + url = db_url + "/results" + installer = get_installer_type(logger) + params = {"project_name": "functest", "case_name": case_name, + "pod_name": pod_name, "installer": installer, + "version": version, "details": payload} + + headers = {'Content-Type': 'application/json'} + try: + r = requests.post(url, data=json.dumps(params), headers=headers) + if logger: + logger.debug(r) + return True + except Exception, e: + print "Error [push_results_to_db('%s', '%s', '%s', '%s', '%s')]:" \ + % (db_url, case_name, pod_name, version, payload), e + return False ========== @@ -647,8 +859,17 @@ References .. _`[1]`: http://artifacts.opnfv.org/functest/docs/configguide/index.html Functest configuration guide URL .. _`[2]`: http://artifacts.opnfv.org/functest/docs/userguide/index.html functest user guide URL .. _`[3]`: https://wiki.opnfv.org/opnfv_test_dashboard - - +.. _`[4]`: https://wiki.opnfv.org/display/INF/CI+Scenario+Naming +.. _`[5]`: https://git.opnfv.org/cgit/functest/tree/ci/testcases.yaml +.. _`[6]`: https://git.opnfv.org/cgit/releng/tree/utils/test/result_collection_api +.. _`[7]`: https://git.opnfv.org/cgit/releng/tree/utils/test/scripts +.. _`[8]`: https://git.opnfv.org/cgit/releng/tree/utils/test/reporting/functest +.. _`[9]`: http://testresults.opnfv.org/reporting/ +.. _`[10]`: https://wiki.opnfv.org/opnfv_functional_testing +.. _`[11]`: https://docs.python.org/2/howto/regex.html +.. _`[12]`: https://regex101.com/ +.. _`[13]`: http://testresults.opnfv.org/test/api/v1/projects/functest/cases +.. _`[14]`: https://git.opnfv.org/cgit/releng/tree/jjb/functest/functest-daily.sh OPNFV main site: opnfvmain_. OPNFV functional test page: opnfvfunctest_. @@ -661,4 +882,3 @@ IRC support chan: #opnfv-testperf .. _`Rally installation procedure`: https://rally.readthedocs.org/en/latest/tutorial/step_0_installation.html .. _`config_test.py` : https://git.opnfv.org/cgit/functest/tree/testcases/config_functest.py .. _`config_functest.yaml` : https://git.opnfv.org/cgit/functest/tree/testcases/config_functest.yaml - |