summaryrefslogtreecommitdiffstats
path: root/docs
diff options
context:
space:
mode:
Diffstat (limited to 'docs')
-rw-r--r--docs/results/fuel-os-nosdn-nofeature-ha.rst109
1 files changed, 101 insertions, 8 deletions
diff --git a/docs/results/fuel-os-nosdn-nofeature-ha.rst b/docs/results/fuel-os-nosdn-nofeature-ha.rst
index 23e3ac3ab..eb8b14741 100644
--- a/docs/results/fuel-os-nosdn-nofeature-ha.rst
+++ b/docs/results/fuel-os-nosdn-nofeature-ha.rst
@@ -14,25 +14,118 @@ Test Results for fuel-os-nosdn-nofeature-ha
Details
=======
-.. after this doc is filled, remove all comments and include the scenario in
-.. results.rst by removing the comment on the file name.
-
+.. _Grafana: http://130.211.154.108/grafana/dashboard/db/yardstick-main
+.. _POD2: https://wiki.opnfv.org/pharos?&#community_test_labs
Overview of test results
------------------------
-.. general on metrics collected, number of iterations
+See Grafana_ for viewing test result metrics for each respective test case. It
+is possible to chose which specific scenarios to look at, and then to zoom in
+on the details of each run test scenario as well.
+
+All of the test case results below are based on 5 consecutive scenario test
+runs, each run on the Ericsson POD2_ between February 13 and 18 in 2016. The
+best would be to have more runs to draw better conclusions from, but these are
+the only runs available at the time of OPNFV R2 release.
+
+TC002
+-----
+The round-trip-time (RTT) between 2 VMs on different blades is measured using
+ping. The measurements are on average varying between 0.5 and 1.1 ms
+with a first 2 - 2.5 ms RTT spike in the beginning of each run (This could be
+because of normal ARP handling). The 2 last runs are very similar in their
+results. But, to be able to draw any further conclusions more runs should be
+made. There is one measurement taken on February 16 that does not have the
+first RTT spike, and less variations to the RTT. The reason for this is
+unknown. There is a discussion on another test measurement made Feb. 16 in
+TC037_.
+SLA set to 10 ms. The SLA value is used as a reference, it has not
+been defined by OPNFV.
+
+TC005
+-----
+The IO read bandwidth look similar between different test runs, with an
+average at approx. 160-170 MB/s. Within each run the results vary much,
+minimum 2 MB/s and maximum 630 MB/s on the totality. Most runs have a
+minimum of 3 MB/s (one run at 2 MB/s). The maximum BW varies much more in
+absolute numbers, between 566 and 630 MB/s.
+SLA set to 400 MB/s. The SLA value is used as a reference, it has not been
+defined by OPNFV.
+
+TC010
+-----
+The measurements for memory latency are consistent among test runs and results
+in approx. 1.2 ns. The variations between runs are similar, between
+1.215 and 1.219 ns. One exception is February 16, where the varation is
+greater, between 1.22 and 1.28 ns. SLA set to 30 ns. The SLA value is used as
+a reference, it has not been defined by OPNFV.
+
+TC011
+-----
+For this scenario no results are available to report on. Probable reason is
+an integer/floating point issue regarding how InfluxDB is populated with
+result data from the test runs.
+
+TC012
+-----
+The average measurements for memory bandwidth are consistent among most of the
+different test runs at 17.2 - 17.3 GB/s. The very first test run averages at
+17.7 GB/s. Within each run the results vary, with a minimal BW of 15.4
+GB/s and maximum of 18.2 GB/s of the totality.
+SLA set to 15 GB/s. The SLA value is used as a reference, it has not been
+defined by OPNFV.
+
+TC014
+-----
+The Unixbench processor single and parallel speed scores show similar results
+at approx. 3200. The runs vary between scores 3160 and 3240.
+No SLA set.
+
+TC037
+-----
+The amount of packets per second (PPS) and round trip times (RTT) between 2 VMs
+on different blades are measured when increasing the amount of UDP flows sent
+between the VMs using pktgen as packet generator tool.
+
+Round trip times and packet throughput between VMs are typically affected by
+the amount of flows set up and result in higher RTT and less PPS
+throughput.
+
+When running with less than 10000 flows the results are flat and consistent.
+RTT is then approx. 30 ms and the number of PPS remains flat at approx.
+250000 PPS. Beyond approx. 10000 flows and up to 1000000 (one million) there
+is an even drop in RTT and PPS performance, eventually ending up at approx.
+150-250 ms and 40000 PPS respectively.
+
+There is one measurement made February 16 that has slightly worse results
+compared to the other 4 measurements. The reason for this is unknown. For
+instance anyone being logged onto the POD can be of relevance for such a
+disturbance.
Detailed test results
---------------------
+The scenario was run on Ericsson POD2_ with:
+Fuel 8.0
+OpenStack Liberty
+OVS 2.3.1
-.. info on lab, installer, scenario
+No SDN controller installed
Rationale for decisions
-----------------------
-.. result analysis, pass/fail
+Pass
+
+Tests were successfully executed and metrics collects (apart from TC011_).
+No SLA was verified. To be decided on in next release of OPNFV.
Conclusions and recommendations
-------------------------------
-
-.. did the expected behavior occured?
+The pktgen test configuration has a relatively large base effect on RTT in
+TC037 compared to TC002, where there is no background load at all (30 ms
+compared to 1 ms or less, which is more than a 3000 percentage different
+in RTT results). The larger amounts of flows in TC037 generate worse
+RTT results, in the magnitude of several hundreds of milliseconds. It would
+be interesting to also make and compare all these measurements to completely
+(optimized) bare metal machines running native Linux with all other relevant
+tools available, e.g. lmbench, pktgen etc.