1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
|
.. This work is licensed under a Creative Commons Attribution 4.0 International License.
.. http://creativecommons.org/licenses/by/4.0
.. (c) OPNFV, Dell EMC and others.
=============================
StorPerf Test Execution Guide
=============================
Prerequisites
=============
This guide requires StorPerf to be running and have its ReST API accessible. If
the ReST API is not running on port 5000, adjust the commands provided here as
needed.
Interacting With StorPerf
=========================
Once the StorPerf container has been started and the ReST API exposed, you can
interact directly with it using the ReST API. StorPerf comes with a Swagger
interface that is accessible through the exposed port at:
.. code-block:: console
http://StorPerf:5000/swagger/index.html
The typical test execution follows this pattern:
#. Configure the environment
#. Initialize the cinder volumes
#. Execute one or more performance runs
#. Delete the environment
OpenStack or Stackless
======================
StorPerf provides the option of controlling the OpenStack environment
via a Heat Stack, or it can run in stackless mode, where it connects
directly to the IP addresses supplied, regardless of how the slave
was created or even if it is an OpenStack VM.
Configure The Environment for OpenStack Usage
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
The following pieces of information are required to prepare the environment:
- The number of VMs/Cinder volumes to create.
- The Cinder volume type (optional) to create
- The Glance image that holds the VM operating system to use.
- The OpenStack flavor to use when creating the VMs.
- The name of the public network that agents will use.
- The size, in gigabytes, of the Cinder volumes to create.
- The number of the Cinder volumes to attach to each VM.
- The availability zone (optional) in which the VM is to be launched. Defaults to **nova**.
- The username (optional) if we specify a custom image.
- The password (optional) for the above image.
**Note**: on ARM based platforms there exists a bug in the kernel which can prevent
VMs from properly attaching Cinder volumes. There are two known workarounds:
#. Create the environment with 0 Cinder volumes attached, and after the VMs
have finished booting, modify the stack to have 1 or more Cinder volumes.
See section on Changing Stack Parameters later in this guide.
#. Add the following image metadata to Glance. This will cause the Cinder
volume to be mounted as a SCSI device, and therefore your target will be
/dev/sdb, etc, instead of /dev/vdb. You will need to specify this in your
warm up and workload jobs.
.. code-block:
--property hw_disk_bus=scsi --property hw_scsi_model=virtio-scsi
The ReST API is a POST to http://StorPerf:5000/api/v1.0/configurations and
takes a JSON payload as follows.
.. code-block:: json
{
"agent_count": int,
"agent_flavor": "string",
"agent_image": "string",
"availability_zone": "string",
"password": "string",
"public_network": "string",
"username": "string",
"volume_count": int,
"volume_size": int,
"volume_type": "string"
}
This call will block until the stack is created, at which point it will return
the OpenStack heat stack id as well as the IP addresses of the slave agents.
Configure The Environment for Stackless Usage
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
To configure the environment for stackless usage, the slaves must be
fully operational (ie: a Linux operating system is running, are reachable
via TCP/IP address or hostname).
It is not necessary to use the Configurations API, but instead define the
stack name as 'null' in any of the other APIs. This instructs StorPerf not to
gather information about the stack from OpenStack, and to simply use the
supplied IP addresses and credentials to communicate with the slaves.
A slave can be a container (provided we can SSH to it), a VM running in any
hypervisor, or even a bare metal server. In the bare metal case, it even
allows for performing RADOS or RDB performance tests using the appropriate
FIO engine.
If the slave SSH server is listening to a port other than 22, the port number
can be specified as part of the address as follows:
IPv4 example for port 2222:
.. code-block::
192.168.1.10:2222
IPv6 example for port 2222:
.. code-block::
[1fe80::58bb:c8b:f2f2:c888]:2222
Helper Container Image for Workloads
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
A new docker container is provided with StorPerf that can be used to test
under docker or Kubernetes environments. It has hard coded credentials
of root/password with an SSH server built it, so be cautious about security
concerns when using this image. It listens internally on port 22, so that
port must be exposed to a free port on the host in order for StorPerf to
reach the synthetic workload container.
.. code-block:: bash
docker run --name=storperf-workloadagent -p 2222:22
opnfv/storperf-workloadagent:latest
Initialize the Target Volumes
=============================
Before executing a test run for the purpose of measuring performance, it is
necessary to fill the volume or file with random data. Failure to execute this
step can result in meaningless numbers, especially for read performance. Most
Cinder drivers are smart enough to know what blocks contain data, and which do
not. Uninitialized blocks return "0" immediately without actually reading from
the volume.
Initiating the data fill behave similarly to a regular performance run, but
will tag the data with a special workload name called "_warm_up". It is
designed to run to completion, filling 100% of the specified target with
random data.
The ReST API is a POST to http://StorPerf:5000/api/v1.0/initializations and
takes a JSON payload as follows. The body is optional unless your target
is something other than /dev/vdb. For example, if you want to profile a
glance ephemeral storage file, you could specify the target as "/filename.dat",
which is a file that then gets created on the root filesystem.
.. code-block:: json
{
"target": "/dev/vdb"
}
This will return a job ID as follows.
.. code-block:: json
{
"job_id": "edafa97e-457e-4d3d-9db4-1d6c0fc03f98"
}
This job ID can be used to query the state to determine when it has completed.
See the section on querying jobs for more information.
Authentication and Slave Selection
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
It is possible to run the Initialization API against a subset of the slaves
known to the stack, or to run it in stackless mode, where StorPerf
connects directly to the IP addresses supplied via SSH. The following
keys are available:
slave_addresses
(optional) A list of IP addresses or hostnames to use as targets. If
omitted, and StorPerf is not running in stackless mode, the full list of
IP addresses from the OpenStack Heat stack is used.
stack_name
(optional) Either the name of the stack in Heat to use, or null if running
in stackless mode.
username
(optional) The username to supply to SSH when logging in. This defaults to
'storperf' if not supplied.
password
(optional) The password to supply to SSH when logging in. If omitted, the
SSH key is used instead.
ssh_private_key
(optional) The SSH private key to supply to SSH when logging in. If omitted,
the default StorPerf private key is used.
This shows an example of stackless mode going against a single bare metal
server reachable by IP address:
.. code-block:: json
{
"username": "labadmin",
"ssh_private_key": "-----BEGIN RSA PRIVATE KEY----- \nMIIE...X0=\n-----END RSA PRIVATE KEY-----",
"slave_addresses": [
"172.17.108.44"
],
"stack_name": null,
}
Filesystems and Mounts
~~~~~~~~~~~~~~~~~~~~~~
It is also possible to instruct StorPerf to create a file system on a device
and mount that as the target directory. The filesystem can be anything
supported by the target slave OS and it is possible to pass specific arguments
to the mkfs command. The following additional keys are available in the
Initializations API for file system control:
mkfs
The type and arguments to pass for creating a filesystem
mount_device
The target device on which to make the file system. The file system will
be mounted on the target specified.
The following example shows the forced creation (-f) of an XFS filesystem
on device /dev/sdb, and mounting that device on /storperf/filesystem.
**Note** If any of the commands (mkfs, mount) fail for any reason, the
Initializations API will return with a 400 code and the body of the response
will contain the error message.
.. code-block:: json
{
"target": "/storperf/filesystem",
"mkfs": "xfs -f",
"mount_device": "/dev/sdb",
}
Initializing Filesystems
~~~~~~~~~~~~~~~~~~~~~~~~
Just like we need to fill Cinder volumes with data, if we want to profile
files on a mounted file system, we need to initialize the file sets with
random data prior to starting a performance run. The Initializations API
can also be used to create test data sets.
**Note** be sure to use the same parameters for the number of files, sizes
and jobs in both the Initializations API and the Jobs API, or you will end
up with possibly incorrect results in the Job performance run.
The following keys are available in the Initializations API for file creation:
filesize
The size of each file to be created and filled with random data.
nrfiles
The number of files per job to create.
numjobs
The number of independent instances of FIO to launch.
Example:
.. code-block:: json
{
"target": "/storperf/filesystem",
"filesize": "2G",
"nrfiles": 10,
"numjobs": 10
}
This would create 100 (10 nrfiles x 10 numjobs) 2G files in the directory
/storperf/filesystem.
.. code-block:: json
{
"username": "labadmin",
"ssh_private_key": "-----BEGIN RSA PRIVATE KEY----- \nMIIE...X0=\n-----END RSA PRIVATE KEY-----",
"slave_addresses": [
"172.17.108.44"
],
"stack_name": null,
"target": "/storperf/filesystem",
"mkfs": "ext4",
"mount_device": "/dev/sdb",
"filesize": "2G",
"nrfiles": 10,
"numjobs": 10
}
Execute a Performance Run
=========================
Performance runs can execute either a single workload, or iterate over a matrix
of workload types, block sizes and queue depths.
Workload Types
~~~~~~~~~~~~~~
rr
Read, Random. 100% read of random blocks
rs
Read, Sequential. 100% read of sequential blocks of data
rw
Read / Write Mix, Sequential. 70% random read, 30% random write
wr
Write, Random. 100% write of random blocks
ws
Write, Sequential. 100% write of sequential blocks.
Custom Workload Types
~~~~~~~~~~~~~~~~~~~~~
New in Gambia (7.0), you can specify custom workload parameters for StorPerf
to pass on to FIO. This is available in the /api/v2.0/jobs API, and takes
a different format than the default v1.0 API.
The format is as follows:
.. code-block:: json
"workloads": {
"name": {
"fio argument": "fio value"
}
}
The name is used the same way the 'rr', 'rs', 'rw', etc is used, but can be
any arbitrary alphanumeric string. This is for you to identify the job later.
Following the name is a series of arguments to pass on to FIO. The most
important on of these is the actual I/O operation to perform. From the `FIO
manual`__, there are a number of different workloads:
.. _FIO_IOP: http://git.kernel.dk/cgit/fio/tree/HOWTO#n985
__ FIO_IOP_
* read
* write
* trim
* randread
* etc
This is an example of how the original 'ws' workload looks in the new format:
.. code-block:: json
"workloads": {
"ws": {
"rw": "write"
}
}
Using this format, it is now possible to initiate any combination of IO
workload type. For example, a mix of 60% reads and 40% writes scattered
randomly throughout the volume being profiled would be:
.. code-block:: json
"workloads": {
"6040randrw": {
"rw": "randrw",
"rwmixread": "60"
}
}
Additional arguments can be added as needed. Here is an example of random
writes, with 25% duplicated blocks, followed by a second run of 75/25% mixed
reads and writes. This can be used to test the deduplication capabilities
of the underlying storage driver.
.. code-block:: json
"workloads": {
"dupwrite": {
"rw": "randwrite",
"dedupe_percentage": "25"
},
"7525randrw": {
"rw": "randrw",
"rwmixread": "75",
"dedupe_percentage": "25"
}
}
There is no limit on the number of workloads and additional FIO arguments
that can be specified.
Note that as in v1.0, the list of workloads will be iterated over with the
block sizes and queue depths specified.
StorPerf will also do a verification of the arguments given prior to returning
a Job ID from the ReST call. If an argument fails validation, the error
will be returned in the payload of the response.
File System Profiling
~~~~~~~~~~~~~~~~~~~~~
As noted in the Initializations API, files in a file system should be
initialized prior to executing a performance run, and the number of jobs,
files and size of files should match the initialization. Given the following
Initializations API call:
.. code-block:: json
{
"username": "labadmin",
"ssh_private_key": "-----BEGIN RSA PRIVATE KEY----- \nMIIE...X0=\n-----END RSA PRIVATE KEY-----",
"slave_addresses": [
"172.17.108.44"
],
"stack_name": null,
"target": "/storperf/filesystem",
"mkfs": "ext4",
"mount_device": "/dev/sdb",
"filesize": "2G",
"nrfiles": 10,
"numjobs": 10
}
The corresponding call to the Jobs API would appear as follows:
.. code-block:: json
{
"username": "labadmin",
"ssh_private_key": "-----BEGIN RSA PRIVATE KEY----- \nMIIE...X0=\n-----END RSA PRIVATE KEY-----",
"slave_addresses": [
"172.17.108.44"
],
"stack_name": null,
"target": "/storperf/filesystem",
"block_sizes": "4k",
"queue_depths": "8",
"workloads": {
"readwritemix": {
"rw": "rw",
"filesize": "2G",
"nrfiles": "10",
"numjobs": "10"
}
}
}
**Note** the queue depths and block sizes as well as the I/O pattern (rw)
can change, but the filesize, nrfiles, numjobs and slave addresses must
match the initialization or the performance run could contain skewed results
due to disk initialization. StorPerf explicitly allows for the mismatch
of these so that it is possible to visualize performance when the files
or disks have not been properly initialized.
Block Sizes
~~~~~~~~~~~
A comma delimited list of the different block sizes to use when reading and
writing data. Note: Some Cinder drivers (such as Ceph) cannot support block
sizes larger than 16k (16384).
Queue Depths
~~~~~~~~~~~~
A comma delimited list of the different queue depths to use when reading and
writing data. The queue depth parameter causes FIO to keep this many I/O
requests outstanding at one time. It is used to simulate traffic patterns
on the system. For example, a queue depth of 4 would simulate 4 processes
constantly creating I/O requests.
Deadline
~~~~~~~~
The deadline is the maximum amount of time in minutes for a workload to run. If
steady state has not been reached by the deadline, the workload will terminate
and that particular run will be marked as not having reached steady state. Any
remaining workloads will continue to execute in order.
.. code-block:: json
{
"block_sizes": "2048,16384",
"deadline": 20,
"queue_depths": "2,4",
"workload": "wr,rr,rw"
}
Metadata
~~~~~~~~
A job can have metadata associated with it for tagging. The following metadata
is required in order to push results to the OPNFV Test Results DB:
.. code-block:: json
"metadata": {
"disk_type": "HDD or SDD",
"pod_name": "OPNFV Pod Name",
"scenario_name": string,
"storage_node_count": int,
"version": string,
"build_tag": string,
"test_case": "snia_steady_state"
}
Changing Stack Parameters
~~~~~~~~~~~~~~~~~~~~~~~~~
While StorPerf currently does not support changing the parameters of the
stack directly, it is possible to change the stack using the OpenStack client
library. The following parameters can be changed:
- agent_count: to increase or decrease the number of VMs.
- volume_count: to change the number of Cinder volumes per VM.
- volume_size: to increase the size of each volume. Note: Cinder cannot shrink volumes.
Increasing the number of agents or volumes, or increasing the size of the volumes
will require you to kick off a new _warm_up job to initialize the newly
allocated volumes.
The following is an example of how to change the stack using the heat client:
.. code-block::
heat stack-update StorPerfAgentGroup --existing -P "volume_count=2"
Query Jobs Information
======================
By issuing a GET to the job API http://StorPerf:5000/api/v1.0/jobs?job_id=<ID>,
you can fetch information about the job as follows:
- &type=status: to report on the status of the job.
- &type=metrics: to report on the collected metrics.
- &type=metadata: to report back any metadata sent with the job ReST API
Status
~~~~~~
The Status field can be:
- Running to indicate the job is still in progress, or
- Completed to indicate the job is done. This could be either normal completion
or manually terminated via HTTP DELETE call.
Workloads can have a value of:
- Pending to indicate the workload has not yet started,
- Running to indicate this is the active workload, or
- Completed to indicate this workload has completed.
This is an example of a type=status call.
.. code-block:: json
{
"Status": "Running",
"TestResultURL": null,
"Workloads": {
"eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.1.block-size.16384": "Pending",
"eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.1.block-size.4096": "Pending",
"eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.1.block-size.512": "Pending",
"eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.4.block-size.16384": "Running",
"eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.4.block-size.4096": "Pending",
"eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.4.block-size.512": "Pending",
"eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.8.block-size.16384": "Completed",
"eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.8.block-size.4096": "Pending",
"eeb2e587-5274-4d2f-ad95-5c85102d055e.ws.queue-depth.8.block-size.512": "Pending"
}
}
If the `job_id` is not provided along with `type` status, then all jobs are returned along with their status.
Metrics
~~~~~~~
Metrics can be queried at any time during or after the completion of a run.
Note that the metrics show up only after the first interval has passed, and
are subject to change until the job completes.
This is a sample of a type=metrics call.
.. code-block:: json
{
"rw.queue-depth.1.block-size.512.read.bw": 52.8,
"rw.queue-depth.1.block-size.512.read.iops": 106.76199999999999,
"rw.queue-depth.1.block-size.512.read.lat_ns.mean": 93.176,
"rw.queue-depth.1.block-size.512.write.bw": 22.5,
"rw.queue-depth.1.block-size.512.write.iops": 45.760000000000005,
"rw.queue-depth.1.block-size.512.write.lat_ns.mean": 21764.184999999998
}
Abort a Job
===========
Issuing an HTTP DELETE to the job api http://StorPerf:5000/api/v1.0/jobs will
force the termination of the whole job, regardless of how many workloads
remain to be executed.
.. code-block:: bash
curl -X DELETE --header 'Accept: application/json' http://StorPerf:5000/api/v1.0/jobs
List all Jobs
=============
A list of all Jobs can also be queried. You just need to issue a GET request without any
Job ID.
.. code-block:: bash
curl -X GET --header 'Accept: application/json' http://StorPerf/api/v1.0/jobs
Delete the Environment
======================
After you are done testing, you can have StorPerf delete the Heat stack by
issuing an HTTP DELETE to the configurations API.
.. code-block:: bash
curl -X DELETE --header 'Accept: application/json' http://StorPerf:5000/api/v1.0/configurations
You may also want to delete an environment, and then create a new one with a
different number of VMs/Cinder volumes to test the impact of the number of VMs
in your environment.
Viewing StorPerf Logs
=====================
Logs are an integral part of any application as they help debugging the application. The user just
needs to issue an HTTP request. To view the entire logs
.. code-block:: bash
curl -X GET --header 'Accept: application/json' http://StorPerf:5000/api/v1.0/logs?lines=all
Alternatively, one can also view a certain amount of lines by specifying the number in the
request. If no lines are specified, then last 35 lines are returned
.. code-block:: bash
curl -X GET --header 'Accept: application/json' http://StorPerf:5000/api/v1.0/logs?lines=12
|