summaryrefslogtreecommitdiffstats
path: root/utils
diff options
context:
space:
mode:
Diffstat (limited to 'utils')
-rw-r--r--utils/test/result_collection_api/dashboard/dashboard_utils.py2
-rw-r--r--utils/test/result_collection_api/dashboard/doctor2Dashboard.py105
-rw-r--r--utils/test/result_collection_api/dashboard/functest2Dashboard.py30
3 files changed, 135 insertions, 2 deletions
diff --git a/utils/test/result_collection_api/dashboard/dashboard_utils.py b/utils/test/result_collection_api/dashboard/dashboard_utils.py
index 3252c3af6..472bbc727 100644
--- a/utils/test/result_collection_api/dashboard/dashboard_utils.py
+++ b/utils/test/result_collection_api/dashboard/dashboard_utils.py
@@ -28,6 +28,8 @@ from qtip2Dashboard import format_qtip_for_dashboard, \
check_qtip_case_exist
from promise2Dashboard import format_promise_for_dashboard, \
check_promise_case_exist
+from doctor2Dashboard import format_doctor_for_dashboard, \
+ check_doctor_case_exist
# any project test project wishing to provide dashboard ready values
# must include at least 2 methods
diff --git a/utils/test/result_collection_api/dashboard/doctor2Dashboard.py b/utils/test/result_collection_api/dashboard/doctor2Dashboard.py
new file mode 100644
index 000000000..eba35b57b
--- /dev/null
+++ b/utils/test/result_collection_api/dashboard/doctor2Dashboard.py
@@ -0,0 +1,105 @@
+ #!/usr/bin/python
+#
+# Copyright (c) 2015 Orange
+# morgan.richomme@orange.com
+#
+# This program and the accompanying materials
+# are made available under the terms of the Apache License, Version 2.0
+# which accompanies this distribution, and is available at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# This script is used to build dashboard ready json results
+# It may be used for all the test case of the Doctor project
+# a new method format_<Test_case>_for_dashboard(results)
+#
+import re
+import datetime
+
+
+def get_doctor_cases():
+ """
+ get the list of the supported test cases
+ TODO: update the list when adding a new test case for the dashboard
+ """
+ return ["doctor-notification","doctor-mark-down"]
+
+
+def format_doctor_for_dashboard(case, results):
+ """
+ generic method calling the method corresponding to the test case
+ check that the testcase is properly declared first
+ then build the call to the specific method
+ """
+
+ if check_doctor_case_exist(case):
+ # note we add _case because testcase and project had the same name
+ # TODO refactoring...looks fine at the beginning wit only 1 project
+ # not very ugly now and clearly not optimized...
+ cmd = "format_" + case.replace('-','_') + "_case_for_dashboard(results)"
+ res = eval(cmd)
+ else:
+ res = []
+ return res
+
+
+def check_doctor_case_exist(case):
+ """
+ check if the testcase exists
+ if the test case is not defined or not declared in the list
+ return False
+ """
+ doctor_cases = get_doctor_cases()
+
+ if (case is None or case not in doctor_cases):
+ return False
+ else:
+ return True
+
+
+def format_doctor_mark_down_case_for_dashboard(results):
+ """
+ Post processing for the doctor test case
+ """
+ test_data = [{'description': 'doctor-mark-down results for Dashboard'}]
+ return test_data
+
+
+def format_doctor_notification_case_for_dashboard(results):
+ """
+ Post processing for the doctor-notification test case
+ """
+ test_data = [{'description': 'doctor results for Dashboard'}]
+ # Graph 1: (duration)=f(time)
+ # ***************************************
+ new_element = []
+
+ # default duration 0:00:08.999904
+ # consider only seconds => 09
+ for data in results:
+ t = data['details']['duration']
+ new_element.append({'x': data['creation_date'],
+ 'y': t})
+
+ test_data.append({'name': "doctor-notification duration ",
+ 'info': {'type': "graph",
+ 'xlabel': 'time (s)',
+ 'ylabel': 'duration (s)'},
+ 'data_set': new_element})
+
+ # Graph 2: bar
+ # ************
+ nbTest = 0
+ nbTestOk = 0
+
+ for data in results:
+ nbTest += 1
+ if data['details']['status'] == "OK":
+ nbTestOk += 1
+
+ test_data.append({'name': "doctor-notification status",
+ 'info': {"type": "bar"},
+ 'data_set': [{'Nb tests': nbTest,
+ 'Nb Success': nbTestOk}]})
+
+ return test_data
diff --git a/utils/test/result_collection_api/dashboard/functest2Dashboard.py b/utils/test/result_collection_api/dashboard/functest2Dashboard.py
index a817f7597..3f4e1a2db 100644
--- a/utils/test/result_collection_api/dashboard/functest2Dashboard.py
+++ b/utils/test/result_collection_api/dashboard/functest2Dashboard.py
@@ -14,8 +14,8 @@
# a new method format_<Test_case>_for_dashboard(results)
# v0.1: basic example with methods for odl, Tempest, Rally and vPing
#
-import re
import datetime
+import re
def get_functest_cases():
@@ -343,6 +343,33 @@ def format_Rally_for_dashboard(results):
Post processing for the Rally test case
"""
test_data = [{'description': 'Rally results for Dashboard'}]
+ # Graph 1: Test_Duration = f(time)
+ # ********************************
+ new_element = []
+ for data in results:
+ summary_cursor = len(data)
+ new_element.append({'x': data['creation_date'],
+ 'y': int(data['details'][summary_cursor]['summary']['duration'])})
+
+ test_data.append({'name': "rally duration",
+ 'info': {'type': "graph",
+ 'xlabel': 'time',
+ 'ylabel': 'duration (s)'},
+ 'data_set': new_element})
+
+ # Graph 2: Success rate = f(time)
+ # ********************************
+ new_element = []
+ for data in results:
+ new_element.append({'x': data['creation_date'],
+ 'y': float(data['details'][summary_cursor]['summary']['nb success'])})
+
+ test_data.append({'name': "rally success rate",
+ 'info': {'type': "graph",
+ 'xlabel': 'time',
+ 'ylabel': 'success rate (%)'},
+ 'data_set': new_element})
+
return test_data
@@ -418,4 +445,3 @@ def format_vPing_userdata_for_dashboard(results):
'Nb Success': nbTestOk}]})
return test_data
-