From 971a7c98515a9d83661f5e423f7e8390f35dca59 Mon Sep 17 00:00:00 2001 From: MofassirArif Date: Thu, 21 Jan 2016 06:42:23 -0800 Subject: bug fix: result collection bug fix for docker images Change-Id: Ia4ea09b90c7a4f4e3699af456c6d66e85661cc0b Signed-off-by: MofassirArif --- docs/network_testcases.rst | 41 +++++++++++++++++++++++++++++++++++++++++ 1 file changed, 41 insertions(+) create mode 100644 docs/network_testcases.rst (limited to 'docs/network_testcases.rst') diff --git a/docs/network_testcases.rst b/docs/network_testcases.rst new file mode 100644 index 00000000..ac68b11b --- /dev/null +++ b/docs/network_testcases.rst @@ -0,0 +1,41 @@ +NETWORK THROUGHPUT TESTCASE + +QTIP uses IPerf3 as the main tool for testing the network throughput. +There are two tests that are run through the QTIP framework. + +Network Throughput for VMs +Network Throughput for Compute Nodes + +For the throughout of the compute nodes we simply go into the systems-under-test +and install iperf3 on the nodes. One of the SUTs is used a server and the other as a +client. The client pushes traffic to the server for a duration specified by the user +configuration file for iperf. These files can be found in the test_cases/{POD}/network/ +directory. The bandwidth is limited only by the physical link layer speed available to the server. +The result file inlcudes the b/s bandwidth and the CPU usage for both the client and server. + +For the VMs we are running two topologies through the framework. + +1: VMs on the same compute nodes +2: VMs on different compute nodes + +QTIP framework sets up a stack with a private network, security groups, routers and attaches the VMs to this network. Iperf3 is installed +on the VMs and one is assigned the role of client while other serves as a server. Traffic is pushed +over the QTIP private network between the VMs. A closer look in needed to see how the traffic actually +flows between the VMs in this configuration to understand what is happening to the packet as traverses +the openstack network. + +The packet originates from VM1 and its sent to the linux bridge via a tap interface where the security groups +are written. Afterwards the packet is forwarded to the Integration bridge via a patch port. Since VM2 is also connected +to the Integration bridge in a similar manner as VM1 so the packet gets forwarded to the linux bridge connecting +VM2. After the linux bridge the packet is sent to VM2 and is recieved by the Iperf3 server. Since no physical link is +involved in this topology, only the OVS (Integration bridge) is being benchmarked and we are seeing bandwidth in the range +of 14-15 Gbps. + +For the topology where the VMs are spawned on different compute nodes, the path the packet takes becomes more cumbersome. +The packet leaves a VM and makes its way to the Integration Bridge as in the first topology however the integration bridge +forwards the packet to the physical link through the ethernet bridge. The packet then gets a VLAN/Tunnel depending on the network +and is forwarded to the particular Compute node where the second VM is spwaned. The packets enter the compute node through the physical +ethernet port and makes its way to the VM through the integration bridge and linux bridge. As seen here the path is much more involved +even when discussed without the mention of overheads faced at all the internfaces so we are seeing the results in the range of 2 Gbps. + + -- cgit 1.2.3-korg