From 77348164208df28ca8545e552cc9c06eefd05256 Mon Sep 17 00:00:00 2001 From: mbeierl Date: Fri, 31 Mar 2017 03:01:17 -0400 Subject: Fixed paths to match standard Change-Id: I733d34eb8bbdf0201b06a21d0ceeda50a06dd49f Signed-off-by: mbeierl --- docs/testing/user/index.rst | 17 +++ docs/testing/user/installation.rst | 129 ++++++++++++++++++++ docs/testing/user/introduction.rst | 101 +++++++++++++++ docs/testing/user/test-usage.rst | 244 +++++++++++++++++++++++++++++++++++++ 4 files changed, 491 insertions(+) create mode 100644 docs/testing/user/index.rst create mode 100755 docs/testing/user/installation.rst create mode 100644 docs/testing/user/introduction.rst create mode 100644 docs/testing/user/test-usage.rst (limited to 'docs/testing') diff --git a/docs/testing/user/index.rst b/docs/testing/user/index.rst new file mode 100644 index 0000000..e9054dc --- /dev/null +++ b/docs/testing/user/index.rst @@ -0,0 +1,17 @@ +.. _storperf-userguide: + +.. This work is licensed under a Creative Commons Attribution 4.0 International +.. License. +.. http://creativecommons.org/licenses/by/4.0 +.. (c) OPNFV, Dell EMC and others. + +====================== +StorPerf User Guide +====================== + +.. toctree:: + :maxdepth: 2 + + introduction.rst + installation.rst + test-usage.rst diff --git a/docs/testing/user/installation.rst b/docs/testing/user/installation.rst new file mode 100755 index 0000000..0effb2f --- /dev/null +++ b/docs/testing/user/installation.rst @@ -0,0 +1,129 @@ +.. This work is licensed under a Creative Commons Attribution 4.0 International License. +.. http://creativecommons.org/licenses/by/4.0 +.. (c) OPNFV, Dell EMC and others. + +=========================== +StorPerf Installation Guide +=========================== + +OpenStack Prerequisites +=========================== +If you do not have an Ubuntu 16.04 image in Glance, you will need to add one. +There are scripts in storperf/ci directory to assist, or you can use the follow +code snippets: + +.. code-block:: bash + + # Put an Ubuntu Image in glance + wget -q https://cloud-images.ubuntu.com/releases/16.04/release/ubuntu-16.04-server-cloudimg-amd64-disk1.img + openstack image create "Ubuntu 16.04 x86_64" --disk-format qcow2 --public \ + --container-format bare --file ubuntu-16.04-server-cloudimg-amd64-disk1.img + + # Create StorPerf flavor + openstack flavor create storperf \ + --id auto \ + --ram 8192 \ + --disk 4 \ + --vcpus 2 + + +Planning +=========================== + +StorPerf is delivered as a `Docker container +`__. There are two possible +methods for installation in your environment: + 1. Run container on Jump Host + 2. Run container in a VM + + +Running StorPerf on Jump Host +============================= + +Requirements: + + * Docker must be installed + * Jump Host must have access to the OpenStack Controller API + * Jump Host must have internet connectivity for downloading docker image + * Enough floating IPs must be available to match your agent count + +Running StorPerf in a VM +======================== + +Requirements: + + * VM has docker installed + * VM has OpenStack Controller credentials and can communicate with the Controller API + * VM has internet connectivity for downloading the docker image + * Enough floating IPs must be available to match your agent count + +VM Creation +~~~~~~~~~~~ + +The following procedure will create the VM in your environment + +.. code-block:: console + + # Create the StorPerf VM itself. Here we use the network ID generated by OPNFV FUEL. + ADMIN_NET_ID=`neutron net-list | grep 'admin_internal_net ' | awk '{print $2}'` + + nova boot --nic net-id=$ADMIN_NET_ID --flavor m1.small --key-name=StorPerf --image 'Ubuntu 14.04' 'StorPerf Master' + +At this point, you may associate a floating IP with the StorPerf master VM. + +VM Docker Installation +~~~~~~~~~~~~~~~~~~~~~~ + +The following procedure will install Docker on Ubuntu 14.04. + +.. code-block:: console + + sudo apt-key adv --keyserver hkp://p80.pool.sks-keyservers.net:80 --recv-keys 58118E89F3A912897C070ADBF76221572C52609D + cat << EOF | sudo tee /etc/apt/sources.list.d/docker.list + deb https://apt.dockerproject.org/repo ubuntu-trusty main + EOF + + sudo apt-get update + sudo apt-get install -y docker-engine + sudo usermod -aG docker ubuntu + +Pulling StorPerf Container +========================== + +Danube +~~~~~~ + +The tag for the latest stable Danube will be: + +.. code-block:: bash + + docker pull opnfv/storperf:danube.1.0 + +Colorado +~~~~~~~~ + +The tag for the latest stable Colorado release is: + +.. code-block:: bash + + docker pull opnfv/storperf:colorado.0.1 + +Brahmaputra +~~~~~~~~~~~ + +The tag for the latest stable Brahmaputra release is: + +.. code-block:: bash + + docker pull opnfv/storperf:brahmaputra.1.2 + +Development +~~~~~~~~~~~ + +The tag for the latest development version is: + +.. code-block:: bash + + docker pull opnfv/storperf:master + + diff --git a/docs/testing/user/introduction.rst b/docs/testing/user/introduction.rst new file mode 100644 index 0000000..a40750f --- /dev/null +++ b/docs/testing/user/introduction.rst @@ -0,0 +1,101 @@ +.. This work is licensed under a Creative Commons Attribution 4.0 International License. +.. http://creativecommons.org/licenses/by/4.0 +.. (c) OPNFV, Dell EMC and others. + +================================== +StorPerf Container Execution Guide +================================== + +Planning +======== + +There are some ports that the container can expose: + + * 22 for SSHD. Username and password are root/storperf. This is used for CLI access only + * 5000 for StorPerf ReST API. + * 8000 for StorPerf's Graphite Web Server + +OpenStack Credentials +~~~~~~~~~~~~~~~~~~~~~ + +You must have your OpenStack Controller environment variables defined and passed to +the StorPerf container. The easiest way to do this is to put the rc file contents +into a clean file the looks similar to this for V2 authentication: + +.. code-block:: console + + OS_AUTH_URL=http://10.13.182.243:5000/v2.0 + OS_TENANT_ID=e8e64985506a4a508957f931d1800aa9 + OS_TENANT_NAME=admin + OS_PROJECT_NAME=admin + OS_USERNAME=admin + OS_PASSWORD=admin + OS_REGION_NAME=RegionOne + +For V3 authentication, use the following: + +.. code-block:: console + + OS_AUTH_URL=http://10.13.182.243:5000/v3 + OS_PROJECT_ID=32ae78a844bc4f108b359dd7320463e5 + OS_PROJECT_NAME=admin + OS_USER_DOMAIN_NAME=Default + OS_USERNAME=admin + OS_PASSWORD=admin + OS_REGION_NAME=RegionOne + OS_INTERFACE=public + OS_IDENTITY_API_VERSION=3 + +Additionally, if you want your results published to the common OPNFV Test Results + DB, add the following: + +.. code-block:: console + + TEST_DB_URL=http://testresults.opnfv.org/testapi + +Running StorPerf Container +========================== + +You might want to have the local disk used for storage as the default size of the docker +container is only 10g. This is done with the -v option, mounting under +/opt/graphite/storage/whisper + +.. code-block:: console + + mkdir -p ~/carbon + sudo chown 33:33 ~/carbon + +The recommended method of running StorPerf is to expose only the ReST and Graphite +ports. The command line below shows how to run the container with local disk for +the carbon database. + +.. code-block:: console + + docker run -t --env-file admin-rc -p 5000:5000 -p 8000:8000 -v ~/carbon:/opt/graphite/storage/whisper --name storperf opnfv/storperf + + +Docker Exec +~~~~~~~~~~~ + +Instead of exposing port 5022 externally, you can use the exec method in docker. This +provides a slightly more secure method of running StorPerf container without having to +expose port 22. + +If needed, the container can be entered with docker exec. This is not normally required. + +.. code-block:: console + + docker exec -it storperf bash + +Container with SSH +~~~~~~~~~~~~~~~~~~ + +Running the StorPerf Container with all ports open and a local disk for the result +storage. This is not recommended as the SSH port is open. + +.. code-block:: console + + docker run -t --env-file admin-rc -p 5022:22 -p 5000:5000 -p 8000:8000 -v ~/carbon:/opt/graphite/storage/whisper --name storperf opnfv/storperf + +This will then permit ssh to localhost port 5022 for CLI access. + diff --git a/docs/testing/user/test-usage.rst b/docs/testing/user/test-usage.rst new file mode 100644 index 0000000..2beae69 --- /dev/null +++ b/docs/testing/user/test-usage.rst @@ -0,0 +1,244 @@ +.. This work is licensed under a Creative Commons Attribution 4.0 International License. +.. http://creativecommons.org/licenses/by/4.0 +.. (c) OPNFV, Dell EMC and others. + +============================= +StorPerf Test Execution Guide +============================= + +Prerequisites +============= + +This guide requires StorPerf to be running and have its ReST API accessible. If +the ReST API is not running on port 5000, adjust the commands provided here as +needed. + +Interacting With StorPerf +========================= + +Once the StorPerf container has been started and the ReST API exposed, you can +interact directly with it using the ReST API. StorPerf comes with a Swagger +interface that is accessible through the exposed port at: + +.. code-block:: console + + http://StorPerf:5000/swagger/index.html + +The typical test execution follows this pattern: + +#. Configure the environment +#. Initialize the cinder volumes +#. Execute one or more performance runs +#. Delete the environment + +Configure The Environment +========================= + +The following pieces of information are required to prepare the environment: + +- The number of VMs/Cinder volumes to create +- The Glance image that holds the VM operating system to use. StorPerf has + only been tested with Ubuntu 16.04 +- The name of the public network that agents will use +- The size, in gigabytes, of the Cinder volumes to create + +The ReST API is a POST to http://StorPerf:5000/api/v1.0/configurations and +takes a JSON payload as follows. + +.. code-block:: json + + { + "agent_count": int, + "agent_image": string, + "public_network": string, + "volume_size": int + } + +This call will block until the stack is created, at which point it will return +the OpenStack heat stack id. + +Initialize the Cinder Volumes +============================= +Before executing a test run for the purpose of measuring performance, it is +necessary to fill the Cinder volume with random data. Failure to execute this +step can result in meaningless numbers, especially for read performance. Most +Cinder drivers are smart enough to know what blocks contain data, and which do +not. Uninitialized blocks return "0" immediately without actually reading from +the volume. + +Initiating the data fill looks the same as a regular performance test, but uses +the special workload called "_warm_up". StorPerf will never push _warm_up +data to the OPNFV Test Results DB, nor will it terminate the run on steady state. +It is guaranteed to run to completion, which fills 100% of the volume with +random data. + +The ReST API is a POST to http://StorPerf:5000/api/v1.0/jobs and +takes a JSON payload as follows. + +.. code-block:: json + + { + "workload": "_warm_up" + } + +This will return a job ID as follows. + +.. code-block:: json + + { + "job_id": "edafa97e-457e-4d3d-9db4-1d6c0fc03f98" + } + +This job ID can be used to query the state to determine when it has completed. +See the section on querying jobs for more information. + +Execute a Performance Run +========================= +Performance runs can execute either a single workload, or iterate over a matrix +of workload types, block sizes and queue depths. + +Workload Types +~~~~~~~~~~~~~~ +rr + Read, Random. 100% read of random blocks +rs + Read, Sequential. 100% read of sequential blocks of data +rw + Read / Write Mix, Random. 70% random read, 30% random write +wr + Write, Random. 100% write of random blocks +ws + Write, Sequential. 100% write of sequential blocks. + +Block Sizes +~~~~~~~~~~~ +A comma delimited list of the different block sizes to use when reading and +writing data. Note: Some Cinder drivers (such as Ceph) cannot support block +sizes larger than 16k (16384). + +Queue Depths +~~~~~~~~~~~~ +A comma delimited list of the different queue depths to use when reading and +writing data. The queue depth parameter causes FIO to keep this many I/O +requests outstanding at one time. It is used to simulate traffic patterns +on the system. For example, a queue depth of 4 would simulate 4 processes +constantly creating I/O requests. + +Deadline +~~~~~~~~ +The deadline is the maximum amount of time in minutes for a workload to run. If +steady state has not been reached by the deadline, the workload will terminate +and that particular run will be marked as not having reached steady state. Any +remaining workloads will continue to execute in order. + +.. code-block:: json + + { + "block_sizes": "2048,16384, + "deadline": 20, + "queue_depths": "2,4", + "workload": "wr,rr,rw", + } + +Metadata +~~~~~~~~ +A job can have metadata associated with it for tagging. The following metadata +is required in order to push results to the OPNFV Test Results DB: + +.. code-block:: json + + "metadata": { + "disk_type": "HDD or SDD", + "pod_name": "OPNFV Pod Name", + "scenario_name": string, + "storage_node_count": int, + "version": string, + "build_tag": string, + "test_case": "snia_steady_state" + } + + + +Query Jobs Information +====================== + +By issuing a GET to the job API http://StorPerf:5000/api/v1.0/jobs?job_id=, +you can fetch information about the job as follows: + +- &type=status: to report on the status of the job. +- &type=metrics: to report on the collected metrics. +- &type=metadata: to report back any metadata sent with the job ReST API + +Status +~~~~~~ +The Status field can be: +- Running to indicate the job is still in progress, or +- Completed to indicate the job is done. This could be either normal completion + or manually terminated via HTTP DELETE call. + +Workloads can have a value of: +- Pending to indicate the workload has not yet started, +- Running to indicate this is the active workload, or +- Completed to indicate this workload has completed. + +This is an example of a type=status call. + +.. code-block:: json + + { + "Status": "Running", + "TestResultURL": null, + "Workloads": { + "eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.1.block-size.16384": "Pending", + "eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.1.block-size.4096": "Pending", + "eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.1.block-size.512": "Pending", + "eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.4.block-size.16384": "Running", + "eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.4.block-size.4096": "Pending", + "eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.4.block-size.512": "Pending", + "eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.8.block-size.16384": "Completed", + "eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.8.block-size.4096": "Pending", + "eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.8.block-size.512": "Pending" + } + } + +Metrics +~~~~~~~ +Metrics can be queried at any time during or after the completion of a run. +Note that the metrics show up only after the first interval has passed, and +are subject to change until the job completes. + +This is a sample of a type=metrics call. + +.. code-block:: json + + { + "rw.queue-depth.1.block-size.512.read.bw": 52.8, + "rw.queue-depth.1.block-size.512.read.iops": 106.76199999999999, + "rw.queue-depth.1.block-size.512.read.lat.mean": 93.176, + "rw.queue-depth.1.block-size.512.write.bw": 22.5, + "rw.queue-depth.1.block-size.512.write.iops": 45.760000000000005, + "rw.queue-depth.1.block-size.512.write.lat.mean": 21764.184999999998 + } + +Abort a Job +=========== +Issuing an HTTP DELETE to the job api http://StorPerf:5000/api/v1.0/jobs will +force the termination of the whole job, regardless of how many workloads +remain to be executed. + +.. code-block:: bash + + curl -X DELETE --header 'Accept: application/json' http://StorPerf:5000/api/v1.0/jobs + +Delete the Environment +====================== +After you are done testing, you can have StorPerf delete the Heat stack by +issuing an HTTP DELETE to the configurations API. + +.. code-block:: bash + + curl -X DELETE --header 'Accept: application/json' http://StorPerf:5000/api/v1.0/configurations + +You may also want to delete an environment, and then create a new one with a +different number of VMs/Cinder volumes to test the impact of the number of VMs +in your environment. -- cgit 1.2.3-korg