diff options
author | Morgan Richomme <morgan.richomme@orange.com> | 2016-05-19 14:41:23 +0200 |
---|---|---|
committer | Morgan Richomme <morgan.richomme@orange.com> | 2016-05-19 14:44:36 +0200 |
commit | d9d3645559462a4139ee33f0cca3be12e37b2d36 (patch) | |
tree | 020870700567e85cf77e36d697c05a2cfa8be4c0 /utils/test | |
parent | 80014d7bb5932998e95ca37729b68938fb28fe9f (diff) |
Adapt reporting after Functest refactoring
Get test cases from cases declares in Tiers
Consider only Tier 0-3 to validate scenario
Display results for Tier > 3
Change-Id: I581702bd7f2cc323d38b82a2404b301fb8fd7840
Signed-off-by: Morgan Richomme <morgan.richomme@orange.com>
Diffstat (limited to 'utils/test')
-rw-r--r-- | utils/test/reporting/functest/__init__.py | 0 | ||||
-rw-r--r-- | utils/test/reporting/functest/default.css | 23 | ||||
-rw-r--r-- | utils/test/reporting/functest/reporting-status.py | 388 | ||||
-rw-r--r-- | utils/test/reporting/functest/reportingUtils.py | 135 | ||||
-rw-r--r-- | utils/test/reporting/functest/scenarioResult.py | 20 | ||||
-rw-r--r-- | utils/test/reporting/functest/template/index-status-tmpl.html | 18 | ||||
-rw-r--r-- | utils/test/reporting/functest/testCase.py | 106 |
7 files changed, 405 insertions, 285 deletions
diff --git a/utils/test/reporting/functest/__init__.py b/utils/test/reporting/functest/__init__.py new file mode 100644 index 000000000..e69de29bb --- /dev/null +++ b/utils/test/reporting/functest/__init__.py diff --git a/utils/test/reporting/functest/default.css b/utils/test/reporting/functest/default.css index 0e330e965..897c3b12b 100644 --- a/utils/test/reporting/functest/default.css +++ b/utils/test/reporting/functest/default.css @@ -53,4 +53,25 @@ td{ .btn-more { color: white; background-color: #0095a2; -}
\ No newline at end of file +} + +h1 { + display: block; + font-size: 2em; + margin-top: 0.67em; + margin-bottom: 0.67em; + margin-left: 0; + margin-right: 0; + font-weight: bold; +} + +h2 { + display: block; + font-size: 1.5em; + margin-top: 0.83em; + margin-bottom: 0.83em; + margin-left: 0; + margin-right: 0; + font-weight: bold; + color:rgb(128, 128, 128) +} diff --git a/utils/test/reporting/functest/reporting-status.py b/utils/test/reporting/functest/reporting-status.py index 9271717bb..e6ee8d321 100644 --- a/utils/test/reporting/functest/reporting-status.py +++ b/utils/test/reporting/functest/reporting-status.py @@ -1,328 +1,159 @@ -from urllib2 import Request, urlopen, URLError +#!/usr/bin/python +# +# This program and the accompanying materials +# are made available under the terms of the Apache License, Version 2.0 +# which accompanies this distribution, and is available at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# import datetime -import json import jinja2 import os -import re import requests import sys import time import yaml -# Declaration of the variables -functest_test_list = ['vPing', 'vPing_userdata', - 'Tempest', 'Rally', - 'ODL', 'ONOS', 'vIMS'] -# functest_test_list = ['vPing'] -companion_test_list = ['doctor/doctor-notification', 'promise/promise'] -# companion_test_list = [] -installers = ["apex", "compass", "fuel", "joid"] -# installers = ["fuel"] -versions = ["brahmaputra", "master"] -# versions = ["master"] -PERIOD = 10 -MAX_SCENARIO_CRITERIA = 18 +import reportingUtils as utils +import reportingConf as conf +import testCase as tc +import scenarioResult as sr -# Correspondance between the name of the test case and the name in the DB -# ideally we should modify the DB to avoid such interface.... -# '<name in the DB':'<name in the config'> -# I know it is uggly... -test_match_matrix = {'healthcheck': 'healthcheck', - 'vPing': 'vping_ssh', - 'vPing_userdata': 'vping_userdata', - 'ODL': 'odl', - 'ONOS': 'onos', - 'Tempest': 'tempest', - 'Rally': 'rally', - 'vIMS': 'vims', - 'doctor-notification': 'doctor', - 'promise': 'promise'} - - -class TestCase(object): - def __init__(self, name, project, criteria=-1, isRunnable=True): - self.name = name - self.project = project - self.criteria = criteria - self.isRunnable = isRunnable - - def getName(self): - return self.name - - def getProject(self): - return self.project - - def getCriteria(self): - return self.criteria - - def setCriteria(self, criteria): - self.criteria = criteria - - def setIsRunnable(self, isRunnable): - self.isRunnable = isRunnable - - def checkRunnable(self, installer, scenario, config): - # Re-use Functest declaration - # Retrieve Functest configuration file functest_config.yaml - is_runnable = True - config_test = "" - TEST_ENV = functest_yaml_config.get("test-dependencies") - - # print " *********************** " - # print TEST_ENV - # print " ---------------------- " - # print "case = " + self.name - # print "installer = " + installer - # print "scenario = " + scenario - # print "project = " + self.project - - # Retrieve test constraints - case_name_formated = test_match_matrix[self.name] - - try: - config_test = TEST_ENV[self.project][case_name_formated] - except KeyError: - # if not defined in dependencies => no dependencies - config_test = TEST_ENV[case_name_formated] - except Exception, e: - print "Error [getTestEnv]:", e - - # Retrieve test execution param - test_execution_context = {"installer": installer, - "scenario": scenario} - # By default we assume that all the tests are always runnable... - # if test_env not empty => dependencies to be checked - if config_test is not None and len(config_test) > 0: - # possible criteria = ["installer", "scenario"] - # consider test criteria from config file - # compare towards CI env through CI en variable - for criteria in config_test: - if re.search(config_test[criteria], - test_execution_context[criteria]) is None: - # print "Test "+ test + " cannot be run on the environment" - is_runnable = False - # print is_runnable - self.isRunnable = is_runnable - - -class ScenarioResult(object): - def __init__(self, status, score=0): - self.status = status - self.score = score - - def getStatus(self): - return self.status - - def getScore(self): - return self.score - -# ***************************************************************************** - - -def getApiResults(case, installer, scenario, version): - case = case.getName() - results = json.dumps([]) - # to remove proxy (to be removed at the end for local test only) - # proxy_handler = urllib2.ProxyHandler({}) - # opener = urllib2.build_opener(proxy_handler) - # urllib2.install_opener(opener) - # url = "http://127.0.0.1:8000/results?case=" + case + \ - # "&period=30&installer=" + installer - url = ("http://testresults.opnfv.org/testapi/results?case=" + case + - "&period=" + str(PERIOD) + "&installer=" + installer + - "&scenario=" + scenario + "&version=" + version) - request = Request(url) - - try: - response = urlopen(request) - k = response.read() - results = json.loads(k) - except URLError, e: - print 'No kittez. Got an error code:', e - - return results - - -def getScenarios(case, installer, version): - - case = case.getName() - url = "http://testresults.opnfv.org/testapi/results?case=" + case + \ - "&period=" + str(PERIOD) + "&installer=" + installer + \ - "&version=" + version - request = Request(url) - - try: - response = urlopen(request) - k = response.read() - results = json.loads(k) - except URLError, e: - print 'Got an error code:', e - - test_results = results['test_results'] - - if test_results is not None: - test_results.reverse() - - scenario_results = {} - - for r in test_results: - # Retrieve all the scenarios per installer - if not r['scenario'] in scenario_results.keys(): - scenario_results[r['scenario']] = [] - scenario_results[r['scenario']].append(r) - - return scenario_results - - -def getScenarioStats(scenario_results): - scenario_stats = {} - for k, v in scenario_results.iteritems(): - scenario_stats[k] = len(v) - - return scenario_stats - - -def getNbtestOk(results): - nb_test_ok = 0 - for r in results: - for k, v in r.iteritems(): - try: - if "passed" in v: - nb_test_ok += 1 - except: - print "Cannot retrieve test status" - return nb_test_ok - - -def getResult(testCase, installer, scenario, version): - - # retrieve raw results - results = getApiResults(testCase, installer, scenario, version) - # let's concentrate on test results only - test_results = results['test_results'] - - # if results found, analyze them - if test_results is not None: - test_results.reverse() - - scenario_results = [] - - # print " ---------------- " - # print test_results - # print " ---------------- " - # print "nb of results:" + str(len(test_results)) - - for r in test_results: - # print r["creation_date"] - # print r["criteria"] - scenario_results.append({r["creation_date"]: r["criteria"]}) - # sort results - scenario_results.sort() - # 4 levels for the results - # 3: 4+ consecutive runs passing the success criteria - # 2: <4 successful consecutive runs but passing the criteria - # 1: close to pass the success criteria - # 0: 0% success, not passing - test_result_indicator = 0 - nbTestOk = getNbtestOk(scenario_results) - # print "Nb test OK:"+ str(nbTestOk) - # check that we have at least 4 runs - if nbTestOk < 1: - test_result_indicator = 0 - elif nbTestOk < 2: - test_result_indicator = 1 - else: - # Test the last 4 run - if (len(scenario_results) > 3): - last4runResults = scenario_results[-4:] - if getNbtestOk(last4runResults): - test_result_indicator = 3 - else: - test_result_indicator = 2 - else: - test_result_indicator = 2 - print " >>>> Test indicator:" + str(test_result_indicator) - return test_result_indicator - -# ****************************************************************************** -# ****************************************************************************** -# ****************************************************************************** -# ****************************************************************************** -# ****************************************************************************** +testCases4Validation = [] +otherTestCases = [] # init just tempest to get the list of scenarios # as all the scenarios run Tempest -tempest = TestCase("Tempest", "functest", -1) +tempest = tc.TestCase("Tempest", "functest", -1) # Retrieve the Functest configuration to detect which tests are relevant # according to the installer, scenario -cf = "https://git.opnfv.org/cgit/functest/plain/ci/config_functest.yaml" +# cf = "https://git.opnfv.org/cgit/functest/plain/ci/config_functest.yaml" +cf = "https://git.opnfv.org/cgit/functest/plain/ci/testcases.yaml" response = requests.get(cf) functest_yaml_config = yaml.load(response.text) print "****************************************" print "* Generating reporting..... *" -print ("* Data retention = %s days *" % PERIOD) +print ("* Data retention = %s days *" % conf.PERIOD) print "* *" print "****************************************" +# Retrieve test cases of Tier 1 (smoke) +config_tiers = functest_yaml_config.get("tiers") + +# we consider Tier 1 (smoke),2 (sdn suites) and 3 (features) +# to validate scenarios +# Tier > 4 are not used to validate scenarios but we display the results anyway +# tricky thing for the API as some tests are Functest tests +# other tests are declared directly in the feature projects +for tier in config_tiers: + if tier['order'] > 0 and tier['order'] < 3: + for case in tier['testcases']: + testCases4Validation.append(tc.TestCase(case['name'], + "functest", + case['dependencies'])) + elif tier['order'] == 3: + for case in tier['testcases']: + testCases4Validation.append(tc.TestCase(case['name'], + case['name'], + case['dependencies'])) + elif tier['order'] > 3: + for case in tier['testcases']: + otherTestCases.append(tc.TestCase(case['name'], + "functest", + case['dependencies'])) + # For all the versions -for version in versions: +for version in conf.versions: # For all the installers - for installer in installers: + for installer in conf.installers: # get scenarios - scenario_results = getScenarios(tempest, installer, version) - scenario_stats = getScenarioStats(scenario_results) + scenario_results = utils.getScenarios(tempest, installer, version) + scenario_stats = utils.getScenarioStats(scenario_results) items = {} scenario_result_criteria = {} # For all the scenarios get results for s, s_result in scenario_results.items(): - testCases = [] # Green or Red light for a given scenario nb_test_runnable_for_this_scenario = 0 scenario_score = 0 - # For each scenario declare the test cases - # Functest cases - for test_case in functest_test_list: - testCases.append(TestCase(test_case, "functest")) - - # project/case - for test_case in companion_test_list: - test_split = test_case.split("/") - test_project = test_split[0] - test_case = test_split[1] - testCases.append(TestCase(test_case, test_project)) - + testCases2BeDisplayed = [] # Check if test case is runnable / installer, scenario + # for the test case used for Scenario validation try: - for test_case in testCases: - test_case.checkRunnable(installer, s, functest_yaml_config) - # print "testcase %s is %s" % (test_case.getName(), - # test_case.isRunnable) + print ("---------------------------------") print ("installer %s, version %s, scenario %s:" % (installer, version, s)) - for testCase in testCases: + + # 1) Manage the test cases for the scenario validation + # concretely Tiers 0-3 + for test_case in testCases4Validation: + test_case.checkRunnable(installer, s, + test_case.getConstraints()) + print ("testcase %s is %s" % (test_case.getName(), + test_case.isRunnable)) time.sleep(1) - if testCase.isRunnable: + if test_case.isRunnable: + dbName = test_case.getDbName() + name = test_case.getName() + project = test_case.getProject() nb_test_runnable_for_this_scenario += 1 print (" Searching results for case %s " % - (testCase.getName())) - result = getResult(testCase, installer, s, version) - testCase.setCriteria(result) - items[s] = testCases + (dbName)) + result = utils.getResult(dbName, installer, s, version) + print " >>>> Test result=" + str(result) + test_case.setCriteria(result) + test_case.setIsRunnable(True) + testCases2BeDisplayed.append(tc.TestCase(name, + project, + "", + result, + True, + 1)) scenario_score = scenario_score + result + + # 2) Manage the test cases for the scenario qualification + # concretely Tiers > 3 + for test_case in otherTestCases: + test_case.checkRunnable(installer, s, + test_case.getConstraints()) + print ("testcase %s is %s" % (test_case.getName(), + test_case.isRunnable)) + time.sleep(1) + if test_case.isRunnable: + dbName = test_case.getDbName() + name = test_case.getName() + project = test_case.getProject() + print (" Searching results for case %s " % + (dbName)) + result = utils.getResult(dbName, installer, s, version) + test_case.setCriteria(result) + test_case.setIsRunnable(True) + testCases2BeDisplayed.append(tc.TestCase(name, + project, + "", + result, + True, + 4)) + + items[s] = testCases2BeDisplayed except: - print ("installer %s, version %s, scenario %s" % + print ("Error: installer %s, version %s, scenario %s" % (installer, version, s)) print "No data available , error %s " % (sys.exc_info()[0]) + # ********************************************** + # Evaluate the results for scenario validation + # ********************************************** # the validation criteria = nb runnable tests x 3 + # because each test case = 0,1,2 or 3 scenario_criteria = nb_test_runnable_for_this_scenario * 3 # if 0 runnable tests set criteria at a high value if scenario_criteria < 1: - scenario_criteria = MAX_SCENARIO_CRITERIA + scenario_criteria = conf.MAX_SCENARIO_CRITERIA s_score = str(scenario_score) + "/" + str(scenario_criteria) s_status = "KO" @@ -333,14 +164,15 @@ for version in versions: else: print ">>>>> scenario OK, save the information" s_status = "OK" - with open("./release/" + version + - "/validated_scenario_history.txt", "a") as f: + path_validation_file = ("./release/" + version + + "/validated_scenario_history.txt") + with open(path_validation_file, "a") as f: time_format = "%Y-%m-%d %H:%M" info = (datetime.datetime.now().strftime(time_format) + ";" + installer + ";" + s + "\n") f.write(info) - scenario_result_criteria[s] = ScenarioResult(s_status, s_score) + scenario_result_criteria[s] = sr.ScenarioResult(s_status, s_score) print "--------------------------" templateLoader = jinja2.FileSystemLoader(os.path.dirname @@ -355,7 +187,7 @@ for version in versions: scenario_results=scenario_result_criteria, items=items, installer=installer, - period=PERIOD, + period=conf.PERIOD, version=version) with open("./release/" + version + diff --git a/utils/test/reporting/functest/reportingUtils.py b/utils/test/reporting/functest/reportingUtils.py new file mode 100644 index 000000000..f8d64aa88 --- /dev/null +++ b/utils/test/reporting/functest/reportingUtils.py @@ -0,0 +1,135 @@ +#!/usr/bin/python +# +# This program and the accompanying materials +# are made available under the terms of the Apache License, Version 2.0 +# which accompanies this distribution, and is available at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +from urllib2 import Request, urlopen, URLError +import json +import reportingConf + + +def getApiResults(case, installer, scenario, version): + results = json.dumps([]) + # to remove proxy (to be removed at the end for local test only) + # proxy_handler = urllib2.ProxyHandler({}) + # opener = urllib2.build_opener(proxy_handler) + # urllib2.install_opener(opener) + # url = "http://127.0.0.1:8000/results?case=" + case + \ + # "&period=30&installer=" + installer + url = ("http://testresults.opnfv.org/testapi/results?case=" + case + + "&period=" + str(reportingConf.PERIOD) + "&installer=" + installer + + "&scenario=" + scenario + "&version=" + version) + request = Request(url) + + try: + response = urlopen(request) + k = response.read() + results = json.loads(k) + except URLError, e: + print 'No kittez. Got an error code:', e + + return results + + +def getScenarios(case, installer, version): + + case = case.getName() + url = ("http://testresults.opnfv.org/testapi/results?case=" + case + + "&period=" + str(reportingConf.PERIOD) + "&installer=" + installer + + "&version=" + version) + request = Request(url) + + try: + response = urlopen(request) + k = response.read() + results = json.loads(k) + except URLError, e: + print 'Got an error code:', e + + test_results = results['test_results'] + + if test_results is not None: + test_results.reverse() + + scenario_results = {} + + for r in test_results: + # Retrieve all the scenarios per installer + if not r['scenario'] in scenario_results.keys(): + scenario_results[r['scenario']] = [] + scenario_results[r['scenario']].append(r) + + return scenario_results + + +def getScenarioStats(scenario_results): + scenario_stats = {} + for k, v in scenario_results.iteritems(): + scenario_stats[k] = len(v) + + return scenario_stats + + +def getNbtestOk(results): + nb_test_ok = 0 + for r in results: + for k, v in r.iteritems(): + try: + if "passed" in v: + nb_test_ok += 1 + except: + print "Cannot retrieve test status" + return nb_test_ok + + +def getResult(testCase, installer, scenario, version): + + # retrieve raw results + results = getApiResults(testCase, installer, scenario, version) + # let's concentrate on test results only + test_results = results['test_results'] + + # if results found, analyze them + if test_results is not None: + test_results.reverse() + + scenario_results = [] + + # print " ---------------- " + # print test_results + # print " ---------------- " + # print "nb of results:" + str(len(test_results)) + + for r in test_results: + # print r["creation_date"] + # print r["criteria"] + scenario_results.append({r["creation_date"]: r["criteria"]}) + # sort results + scenario_results.sort() + # 4 levels for the results + # 3: 4+ consecutive runs passing the success criteria + # 2: <4 successful consecutive runs but passing the criteria + # 1: close to pass the success criteria + # 0: 0% success, not passing + test_result_indicator = 0 + nbTestOk = getNbtestOk(scenario_results) + # print "Nb test OK:"+ str(nbTestOk) + # check that we have at least 4 runs + if nbTestOk < 1: + test_result_indicator = 0 + elif nbTestOk < 2: + test_result_indicator = 1 + else: + # Test the last 4 run + if (len(scenario_results) > 3): + last4runResults = scenario_results[-4:] + if getNbtestOk(last4runResults): + test_result_indicator = 3 + else: + test_result_indicator = 2 + else: + test_result_indicator = 2 + return test_result_indicator diff --git a/utils/test/reporting/functest/scenarioResult.py b/utils/test/reporting/functest/scenarioResult.py new file mode 100644 index 000000000..743346adc --- /dev/null +++ b/utils/test/reporting/functest/scenarioResult.py @@ -0,0 +1,20 @@ +#!/usr/bin/python +# +# This program and the accompanying materials +# are made available under the terms of the Apache License, Version 2.0 +# which accompanies this distribution, and is available at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# + + +class ScenarioResult(object): + def __init__(self, status, score=0): + self.status = status + self.score = score + + def getStatus(self): + return self.status + + def getScore(self): + return self.score diff --git a/utils/test/reporting/functest/template/index-status-tmpl.html b/utils/test/reporting/functest/template/index-status-tmpl.html index 7a0656b74..89a1d1527 100644 --- a/utils/test/reporting/functest/template/index-status-tmpl.html +++ b/utils/test/reporting/functest/template/index-status-tmpl.html @@ -74,20 +74,25 @@ <table class="table"> <tr> {% for test in items[scenario] -%} - <th>{{test.getName() }}</th> + <th> + {% if test.getCriteria() > -1 -%} + {{test.getDbName() }} + {%- endif %} + {% if test.getTier() > 3 -%} + * + {%- endif %} + </th> {%- endfor %} </tr> <tr class="tr-weather-weather"> - {% for test in items[scenario] -%} - {% if test.isRunnable is sameas false -%} - <td>N.R</td> - {% elif test.getCriteria() > 2 -%} + {% for test in items[scenario] -%} + {% if test.getCriteria() > 2 -%} <td><img src="../../img/weather-clear.png"></td> {%- elif test.getCriteria() > 1 -%} <td><img src="../../img/weather-few-clouds.png"></td> {%- elif test.getCriteria() > 0 -%} <td><img src="../../img/weather-overcast.png"></td> - {%- else -%} + {%- elif test.getCriteria() > -1 -%} <td><img src="../../img/weather-storm.png"></td> {%- endif %} {%- endfor %} @@ -96,6 +101,7 @@ </div> </div> {%- endfor %} + *: not used for scenario validation </div> <div class="col-md-1"></div> </div> diff --git a/utils/test/reporting/functest/testCase.py b/utils/test/reporting/functest/testCase.py new file mode 100644 index 000000000..f6ab95a05 --- /dev/null +++ b/utils/test/reporting/functest/testCase.py @@ -0,0 +1,106 @@ +#!/usr/bin/python +# +# This program and the accompanying materials +# are made available under the terms of the Apache License, Version 2.0 +# which accompanies this distribution, and is available at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +import re + + +class TestCase(object): + + def __init__(self, name, project, constraints, + criteria=-1, isRunnable=True, tier=-1): + self.name = name + self.project = project + self.constraints = constraints + self.criteria = criteria + self.isRunnable = isRunnable + self.tier = tier + + def getName(self): + return self.name + + def getProject(self): + return self.project + + def getConstraints(self): + return self.constraints + + def getCriteria(self): + return self.criteria + + def getTier(self): + return self.tier + + def setCriteria(self, criteria): + self.criteria = criteria + + def setIsRunnable(self, isRunnable): + self.isRunnable = isRunnable + + def checkRunnable(self, installer, scenario, config): + # Re-use Functest declaration + # Retrieve Functest configuration file functest_config.yaml + is_runnable = True + config_test = config + # print " *********************** " + # print TEST_ENV + # print " ---------------------- " + # print "case = " + self.name + # print "installer = " + installer + # print "scenario = " + scenario + # print "project = " + self.project + + # Retrieve test constraints + # Retrieve test execution param + test_execution_context = {"installer": installer, + "scenario": scenario} + + # By default we assume that all the tests are always runnable... + # if test_env not empty => dependencies to be checked + if config_test is not None and len(config_test) > 0: + # possible criteria = ["installer", "scenario"] + # consider test criteria from config file + # compare towards CI env through CI en variable + for criteria in config_test: + if re.search(config_test[criteria], + test_execution_context[criteria]) is None: + # print "Test "+ test + " cannot be run on the environment" + is_runnable = False + # print is_runnable + self.isRunnable = is_runnable + + def toString(self): + testcase = ("Name=" + self.name + ";Criteria=" + str(self.criteria) + + ";Project=" + self.project + ";Constraints=" + + str(self.constraints) + ";IsRunnable" + + str(self.isRunnable)) + return testcase + + def getDbName(self): + # Correspondance name of the test case / name in the DB + # ideally we should modify the DB to avoid such interface.... + # '<name in the config>':'<name in the DB>' + # I know it is uggly... + test_match_matrix = {'healthcheck': 'healthcheck', + 'vping_ssh': 'vPing', + 'vping_userdata': 'vPing_userdata', + 'odl': 'ODL', + 'onos': 'ONOS', + 'ovno': 'ovno', + 'tempest_smoke_serial': 'Tempest', + 'tempest_full_parallel': 'tempest_full_parallel', + 'rally_sanity': 'Rally', + 'bgpvpn': 'bgpvpn', + 'rally_full': 'rally_full', + 'vims': 'vIMS', + 'doctor': 'doctor-notification', + 'promise': 'promise' + } + try: + return test_match_matrix[self.name] + except: + return "unknown" |