summaryrefslogtreecommitdiffstats
path: root/utils/test/reporting
diff options
context:
space:
mode:
Diffstat (limited to 'utils/test/reporting')
-rwxr-xr-xutils/test/reporting/docker/reporting.sh2
-rwxr-xr-xutils/test/reporting/reporting/functest/reporting-status.py24
-rw-r--r--utils/test/reporting/reporting/qtip/reporting-status.py19
-rw-r--r--utils/test/reporting/reporting/qtip/template/index-status-tmpl.html8
-rw-r--r--utils/test/reporting/reporting/reporting.yaml2
5 files changed, 31 insertions, 24 deletions
diff --git a/utils/test/reporting/docker/reporting.sh b/utils/test/reporting/docker/reporting.sh
index 25d3ef117..2cb438d00 100755
--- a/utils/test/reporting/docker/reporting.sh
+++ b/utils/test/reporting/docker/reporting.sh
@@ -59,8 +59,6 @@ if [ -z "$1" ]; then
report_project $i $i "status"
sleep 5
done
- report_project "QTIP" "qtip" "status"
-
echo "Functest reporting vIMS..."
report_project "functest" "functest" "vims"
diff --git a/utils/test/reporting/reporting/functest/reporting-status.py b/utils/test/reporting/reporting/functest/reporting-status.py
index 267803e6c..c71e00f3b 100755
--- a/utils/test/reporting/reporting/functest/reporting-status.py
+++ b/utils/test/reporting/reporting/functest/reporting-status.py
@@ -230,12 +230,26 @@ for version in versions:
# Evaluate the results for scenario validation
# **********************************************
# the validation criteria = nb runnable tests x 3
- # because each test case = 0,1,2 or 3
- scenario_criteria = nb_test_runnable_for_this_scenario * 3
- # if 0 runnable tests set criteria at a high value
- if scenario_criteria < 1:
- scenario_criteria = 50 # conf.MAX_SCENARIO_CRITERIA
+ # because each test case can get
+ # 0 point (never PASS)
+ # 1 point at least (PASS once over the time window)
+ # 2 points (PASS more than once but 1 FAIL on the last 4)
+ # 3 points PASS on the last 4 iterations
+ # e.g. 1 scenario = 10 cases
+ # 1 iteration : max score = 10 (10x1)
+ # 2 iterations : max score = 20 (10x2)
+ # 3 iterations : max score = 20
+ # 4 or more iterations : max score = 30 (1x30)
+ if len(s_result) > 3:
+ k_score = 3
+ elif len(s_result) < 2:
+ k_score = 1
+ else:
+ k_score = 2
+
+ scenario_criteria = nb_test_runnable_for_this_scenario*k_score
+ # score for reporting
s_score = str(scenario_score) + "/" + str(scenario_criteria)
s_score_percent = rp_utils.getScenarioPercent(
scenario_score,
diff --git a/utils/test/reporting/reporting/qtip/reporting-status.py b/utils/test/reporting/reporting/qtip/reporting-status.py
index 56f9e0aee..524338ca0 100644
--- a/utils/test/reporting/reporting/qtip/reporting-status.py
+++ b/utils/test/reporting/reporting/qtip/reporting-status.py
@@ -45,27 +45,22 @@ def prepare_profile_file(version):
def profile_results(results, installer, profile_fd):
result_criterias = {}
for s_p, s_p_result in results.iteritems():
- ten_criteria = len(s_p_result)
- ten_score = sum(s_p_result)
+ avg_last_ten = sum([int(s) for s in s_p_result]) / len(s_p_result)
LASTEST_TESTS = rp_utils.get_config(
'general.nb_iteration_tests_success_criteria')
- four_result = s_p_result[:LASTEST_TESTS]
- four_criteria = len(four_result)
- four_score = sum(four_result)
-
- s_four_score = str(four_score / four_criteria)
- s_ten_score = str(ten_score / ten_criteria)
+ last_four = s_p_result[-LASTEST_TESTS:]
+ avg_last_four = sum([int(s) for s in last_four]) / len(last_four)
info = '{},{},{},{},{}\n'.format(reportingDate,
s_p,
installer,
- s_ten_score,
- s_four_score)
+ '',
+ avg_last_four)
profile_fd.write(info)
result_criterias[s_p] = sr.ScenarioResult('OK',
- s_four_score,
- s_ten_score,
+ avg_last_four,
+ avg_last_ten,
'100')
logger.info("--------------------------")
diff --git a/utils/test/reporting/reporting/qtip/template/index-status-tmpl.html b/utils/test/reporting/reporting/qtip/template/index-status-tmpl.html
index 92f3395dc..da8424a22 100644
--- a/utils/test/reporting/reporting/qtip/template/index-status-tmpl.html
+++ b/utils/test/reporting/reporting/qtip/template/index-status-tmpl.html
@@ -47,10 +47,10 @@
<ul class="nav nav-justified">
<li class="active"><a href="http://testresults.opnfv.org/reporting/index.html">Home</a></li>
<li><a href="status-apex.html">Apex</a></li>
- <li><a href="status-compass.html">Compass</a></li>
- <li><a href="status-daisy.html">Daisy</a></li>
- <li><a href="status-fuel.html">Fuel</a></li>
- <li><a href="status-joid.html">Joid</a></li>
+ <!--<li><a href="status-compass.html">Compass</a></li>-->
+ <!--<li><a href="status-daisy.html">Daisy</a></li>-->
+ <!--<li><a href="status-fuel.html">Fuel</a></li>-->
+ <!--<li><a href="status-joid.html">Joid</a></li>-->
</ul>
</nav>
</div>
diff --git a/utils/test/reporting/reporting/reporting.yaml b/utils/test/reporting/reporting/reporting.yaml
index dae832ab5..2a4aa492c 100644
--- a/utils/test/reporting/reporting/reporting.yaml
+++ b/utils/test/reporting/reporting/reporting.yaml
@@ -56,7 +56,7 @@ storperf:
qtip:
log_level: ERROR
- period: 1
+ period: 10
bottlenecks:
test_list: