aboutsummaryrefslogtreecommitdiffstats
path: root/docs/testing/developer/devguide/storage-qpi.rst
diff options
context:
space:
mode:
Diffstat (limited to 'docs/testing/developer/devguide/storage-qpi.rst')
-rw-r--r--docs/testing/developer/devguide/storage-qpi.rst112
1 files changed, 112 insertions, 0 deletions
diff --git a/docs/testing/developer/devguide/storage-qpi.rst b/docs/testing/developer/devguide/storage-qpi.rst
new file mode 100644
index 00000000..8a82241e
--- /dev/null
+++ b/docs/testing/developer/devguide/storage-qpi.rst
@@ -0,0 +1,112 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. (c) 2017 ZTE Corp.
+
+
+***********
+Storage QPI
+***********
+
+The storage QPI gives user an overall score for storage performance.
+
+The measurement is done by `StorPerf`_.
+
+.. _StorPerf: https://wiki.opnfv.org/display/storperf
+
+System Information
+==================
+
+System Information are environmental parameters and factors may affect storage performance:
+
++--------------------------------+-------------------------------------------------+-----------------------------------------------------------------------------+
+| System Factors | Detail | Extraction Method |
++--------------------------------+-------------------------------------------------+-----------------------------------------------------------------------------+
+| Ceph Node List | List of nodes which has ceph-osd roles. | Getting from return result of installer node list CLI command. |
+| | For example [node-2, node-3, node-4]. | |
++--------------------------------+-------------------------------------------------+-----------------------------------------------------------------------------+
+| Ceph Client RDB Cache Mode | Values: "None", "write-through", "write-back". | Getting from value of "rbd cache" and "rbd cache max dirty" keys |
+| | | in client section of ceph configuration; To enable write-through mode, |
+| | | set rbd cache max dirty to 0. |
++--------------------------------+-------------------------------------------------+-----------------------------------------------------------------------------+
+| Ceph Client RDB Cache Size | The RBD cache size in bytes. Default is 32 MiB. | Getting from value of "rdb cache size" key in client section |
+| | | of ceph configuration. |
++--------------------------------+-------------------------------------------------+-----------------------------------------------------------------------------+
+| Ceph OSD Tier Cache Mode | Values: "None", "Write-back", "Readonly". | Getting from ceph CLI "ceph report" output info. |
++--------------------------------+-------------------------------------------------+-----------------------------------------------------------------------------+
+| Use SSD Backed OSD Cache | Values: "Yes", "No". | Getting from POD description and CEPH CLI "ceph-disk list" output info. |
++--------------------------------+-------------------------------------------------+-----------------------------------------------------------------------------+
+| Use SSD For Journal | Values: "Yes", "No". | Getting from POD description and CEPH CLI "ceph-disk list" output info. |
++--------------------------------+-------------------------------------------------+-----------------------------------------------------------------------------+
+| Ceph Cluster Network Bandwidth | Values: "1G", "10G", "40G". | Getting from physical interface information in POD description, |
+| | | "ifconfig" output info on ceph osd node, |
+| | | and value of "cluster network" key in global section of ceph configuration. |
++--------------------------------+-------------------------------------------------+-----------------------------------------------------------------------------+
+
+Test Condition
+==============
+
++-----------------------------+------------------------------------------+-------------------------------------------------------------+
+| Test Condition | Detail | Extraction Method |
++-----------------------------+------------------------------------------+-------------------------------------------------------------+
+| Number of Testing VMs | Number of VMs which are created, | It equals the number of Cinder nodes of the SUT. |
+| | during running Storperf test case. | |
++-----------------------------+------------------------------------------+-------------------------------------------------------------+
+| Distribution of Testing VMS | Number of VMs on each computer node, | Recording the distribution when runing Storperf test case. |
+| | for example [(node-2: 1), (node-3: 2))]. | |
++-----------------------------+------------------------------------------+-------------------------------------------------------------+
+
+Baseline
+========
+
+Baseline is established by testing with a set of work loads:
+
+- `Queue depth`_ (1, 2, 8)
+- `Block size`_ (2KB, 8KB, 16KB)
+- `Read write`_
+ - sequential read
+ - sequential write
+ - random read
+ - random write
+ - random mixed read write 70/30
+
+.. _Queue depth: http://fio.readthedocs.io/en/latest/fio_man.html#cmdoption-arg-iodepth
+.. _Block size: http://fio.readthedocs.io/en/latest/fio_man.html#cmdoption-arg-blocksize
+.. _Read write: http://fio.readthedocs.io/en/latest/fio_man.html#cmdoption-arg-readwrite
+
+Metrics
+=======
+
+- Throughput: data transfer rate
+- IOPS: I/O operations per second
+- Latency: response time
+
+Workload Scores
+===============
+
+For each test run, if an equivalent work load in baseline is available, a score will be calculated by comparing the
+result to baseline.
+
+Section Scores
+==============
+
++-----------------+--------------------------------------------------------+-----------------------------------------+
+| Section | Detail | Indication |
++=================+========================================================+=========================================+
+| IOPS | Read write I/O Operation per second under steady state | Important for frequent storage access |
+| | Workloads : random read/write | such as event sinks |
++-----------------+--------------------------------------------------------+-----------------------------------------+
+| Throughput | Read write data transfer rate under steady state | Important for high throughput services |
+| | Workloads: sequential read/write, block size 16KB | such as video server |
++-----------------+--------------------------------------------------------+-----------------------------------------+
+| Latency | Average response latency under steady state | Important for real time applications |
+| | Workloads: all | |
++-----------------+--------------------------------------------------------+-----------------------------------------+
+
+Section score is the `geometric mean <https://en.wikipedia.org/wiki/Geometric_mean>`_ of all
+workload score.
+
+Storage QPI
+===========
+
+Storage QPI is the `weighted arithmetic mean <https://en.wikipedia.org/wiki/Weighted_arithmetic_mean>`_ of all section
+scores.