Age | Commit message (Collapse) | Author | Files | Lines |
|
|
|
JIRA: YARDSTICK-622
Bug log: https://build.opnfv.org/ci/view/armband/job/yardstick-fuel-armband-baremetal-daily-danube/24/console
This error is because the file changes done in the load_images.sh is not updated.
After file modification, pip install -U . is required to tack changes into effect.
Change-Id: I8be3ffc963e3d5cec861b68a33efc2fa5bdabbd2
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
|
|
This allows calculating throughput in Gbps.
Add: test_pktgen.py
Change-Id: I7e2132ec71985c14570ef18b952b2264ebf7abbc
JIRA: YARKSTICK-611
Signed-off-by: Jing Zhang <jing.c.zhang@nokia.com>
|
|
JIRA: YARDSTICK-617
Apex has a new scenario 'opnfv_os-odl_l3-ovs-ha_daily' in CI.
This patch add yardstick test suite for this scenario.
Change-Id: I50eb84fb506e7cf87cbd6bcb67f3df6d185d8b6b
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
JIRA: YARDSTICK-597
Currently yardstick can not run offline.
In some case this will block the process, causing yardstick job failed
The reason is yardstick will pull the latest code when run in CI.
Actually we need not pull the latest code because the code in yardstick
is already the latest code.
So I remove them.
Also I check the environment variable and yardstick-image when fetch
openrc file and load images.
Change-Id: I797da272932ab5292aa40585c82c119aa71b8849
Signed-off-by: chenjiankun <chenjiankun1@huawei.com>
|
|
|
|
|
|
JIRA: YARDSTICK-607
some scenarios come with a Cirros-0.3.5 image, Yardstick's load_images.sh will
load a second cirros-0.3.5 image, causing the following error:
2017-03-27 07:42:58,317 yardstick.orchestrator.heat heat.py:449 INFO Creating stack 'demo-2ada98f3'
error: failed to deploy stack: 'ERROR: Property error: : resources.ares.demo-2ada98f3.properties.image: : Error validating value 'cirros-0.3.5': No images unique match found for
{'name': u'cirros-0.3.5'}
This patch fixs this error by skipping load cirros image if there already exist
one in the glance.
Change-Id: I4d43d82d331f6575efd4560be9b6be175e03cd1b
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
|
|
|
|
JIRA: YARDSTICK-602
change HA test cases to rely on dynamical generated pod.yaml in fuel pods
Change-Id: I0efb01489eb2dd94babf1a45bff2fbf6adef5f6e
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
JIRA: YARDSTICK-604
the storperf container now uses a flavor called "storperf", if this flavor is
absent, stack creating process will fail. also the openstack variables that
passed to the storperf container also needs a update
Change-Id: Iab4be6491cc9b24526948e5571f0476ec885a93d
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
Change-Id: Iffa1e4529e5b6e9664ac503a5815ea9a034d7515
Signed-off-by: Michael Polenchuk <mpolenchuk@mirantis.com>
|
|
remove redundnant parens
replace type with isinstance
fix import order
Change-Id: I0407cbbf6993290f392f17c398827603b5bc1ebd
Signed-off-by: Ross Brattain <ross.b.brattain@intel.com>
|
|
JIRA: YARDSTICK-601
This error is cause because the TC 002 configuration file has been modified in
image section, the old sed command in load_images.sh
"sed -i 's/image: cirros-0.3.3/image: TestVM/g' tests/opnfv/test_cases/opnfv_yardstick_tc002.yaml samples/ping.yaml" didn't work properly.
Change-Id: Ie897213622d05bb267f94b466eb2529e797d484e
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
|
|
|
|
|
|
|
|
|
|
Get the arch from a compute node
Also move code that depends on YARD_IMG_ARCH,
after YARD_IMG_ARCH gets set
JIRA: YARDSTICK-591
Change-Id: I5ed3ef9cfa92e5bf9e6ada21396d18643cf1a55c
Signed-off-by: Matei Valeanu <Matei.Valeanu@enea.com>
|
|
|
|
Vivid is end of life, and thus not supported.
Change-Id: Ia293fcd6fd2a3279a642a6dbbb7ad1811334c3e1
Signed-off-by: Ross Brattain <ross.b.brattain@intel.com>
|
|
JIRA: YARDSTICK-598
As we have replaced 'glance image-create' command with 'openstack image create'
command, we also need to update the disk-format for openstack CLI. the
'root-tar' is no longer a valid choice. (choose from 'ami', 'ari', 'aki',
'vhd', 'vmdk', 'raw', 'qcow2', 'vhdx', 'vdi', 'iso', 'ploop')
Change-Id: I53f839f658c5a6f216f5bb942b1c4f4dcdbf9814
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
JIRA: YARDSTICK-589
Change-Id: I1690d4e2bd741fe1514c8496ee6ff10b72158859
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
|
|
Currently Yardstick doesnt have any UI which gives detail analysis
of the test-results.
This commit generates a HTML page after the execution of a command
"yardstick report generate <task-ID> <TC-name>" which intern
can be executed after the execution of test-case.
Used: Highcharts.js for the graphs.
JIRA: YARDSTICK-280
Change-Id: Ic98cc348719f3922bff178f52e7944a4a931763a
Signed-off-by: Rajesh K <4k.rajesh@gmail.com>
|
|
JIRA: YARDSTICK-589
Some new scenarios, e.g, opnfv_k8-nosdn-nofeature-noha_daily, have been added
in the Danube release, test suite files for these scenarios should be added.
Change-Id: If3f4ef233b7b1516e182ed28f15feece1728e6d9
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
JIRA: YARDSTICK-573
Since the yardstick framework now has supported an improved node type context,
this patch adds support for VM vcpu pinning ability in the node type context.
It provides several scripts that can be used to configurate the controller and
compute nodes.
Change-Id: If2c6e7b1b85ff78b9d2a5997bf03bdc6877aaf74
Signed-off-by: JingLu5 <lvjing5@huawei.com>
|
|
|
|
load_images.sh arm64 change had issue with shell
load_images.sh: line 34: [: amd64=: unary operator expected
https://build.opnfv.org/ci/view/yardstick/job/yardstick-compass-baremetal-daily-master/894/console
Change-Id: I14f0b8628db7ac90a4966f85e9e7e27a0e0fae02
Signed-off-by: Ross Brattain <ross.b.brattain@intel.com>
|
|
JIRA: YARDSTICK-585
In CI when run tc070, there is a error: Failed executing command: 'free -s 1 -c 10'
Here it is the log:
Traceback (most recent call last):
File
"/usr/local/lib/python2.7/dist-packages/yardstick/benchmark/runners/duration.py",
line 69, in _worker_process
method(data)
File
"/usr/local/lib/python2.7/dist-packages/yardstick/benchmark/scenarios/compute/memload.py",
line 126, in run
result.update(self._get_mem_usage())
File
"/usr/local/lib/python2.7/dist-packages/yardstick/benchmark/scenarios/compute/memload.py",
line 116, in _get_mem_usage
result = self._execute_command(cmd)
File
"/usr/local/lib/python2.7/dist-packages/yardstick/benchmark/scenarios/compute/memload.py",
line 70, in _execute_command
cmd, stderr)
RuntimeError: ('Failed executing command: ', 'free -s 1 -c 10',u"free: seconds argument `1' failed\n")
And it is a bug of free. the -c option should in front of -s, so change the position will solve
this problem.
Also it has another bug:
'KeyError', there no 'cached' keyword, so I change it to 'buff/cache'.
Change-Id: I0ca16e8d8cc11c6a3b2f364cadbdb3ea367eee53
Signed-off-by: chenjiankun <chenjiankun1@huawei.com>
|
|
We should have run shellcheck on this.
This is why we don't use the shell, too many
obscure gotchas.
shellcheck: load_images.sh:25:75: warning: sudo doesn't affect redirects. Use .. | sudo tee -a file [SC2024]
sudo echo foo >> /etc/foo won't work,
the >> will be run in the current shell,
replace with:
echo foo | sudo tee -a /etc/foo
shellcheck: load_images.sh:47:15: warning: Declare and assign separately to avoid masking return values. [SC2155]
replace local cmd='a' with two line version
shellcheck: load_images.sh:141:32: error: [ .. ] can't match globs. Use [[ .. ]] or grep. [SC2081]
use [[ ]]
Change-Id: I14d9768612ea147eaf139603f7eb8753e979ba2c
Signed-off-by: Ross Brattain <ross.b.brattain@intel.com>
|
|
in parallel
JIRA: YARDSTICK-575
Currently yardstick will create stacks with the same name when run using
API in parallel.
The reason is there is a global variable in context base and the core
will always deploy the first context in Context.list. When run in
parallel, it will run in the one process. So yardstick will deploy
stacks with the same name.
The solution is do not use Context.list in yardstick core. And using a
local variable instead.
BTW, if we use API to call yardstick core, we can not config the output
way. So I parse yardstick.conf when task start. And I think we can
include scenario_cfg, context_cfg, yardstick_cfg in one config object
later so that we can get all config in one object.
Change-Id: I1ada4ef486bd252e78c3a2e49c6a39b3f8f16a7c
Signed-off-by: chenjiankun <chenjiankun1@huawei.com>
|
|
|
|
Add support for downloading and building an uefi single part image.
Keep three part image support if needed.
Keep modifications required for arm64, when using vivid.
Install and use parted required for GPT partitioning.
Change-Id: I1a9a20701e2fe1f6f644f2c687c5da2e09b69d89
Signed-off-by: Catalina Focsa <catalina.focsa@enea.com>
|
|
|
|
|
|
JIRA: YARDSTICK-576
Currently the reporting_target reuse the http_target.
But if we want to custom reporting_target, the http_target is also change,
and the yardstick result will store in mongoDB.
So I add reporting_target to split them.
Change-Id: Ib0e4e8ec9a55f49027f9776931fdaa7dc194654b
Signed-off-by: chenjiankun <chenjiankun1@huawei.com>
|
|
Change-Id: I46e4e21787a9b0a3104b2a684d62f50681258205
Signed-off-by: Romanos Skiadas <rski@intracom-telecom.com>
|
|
|
|
JIRA: YARDSTICK-566
Current Heat context support affinity and availability arguments
but not support affinity and anti-affinity.
Enhance Heat context to support affinity and anti-affinity:
1. can create heat server group with affinity/anti-affinity
2. each server could be specified which server group they are in
Change-Id: I46e7376fd116c6e109cb5dcb1c168460918e6d43
Signed-off-by: rexlee8776 <limingjiang@huawei.com>
|
|
JIRA: YARDSTICK-569
We have much redundancy code when using nova/neutron/glance... python client.
So I write this code in openstack_utils for calling.
Change-Id: I322b7577de4933246a15e6742ae5a28bea16eb02
Signed-off-by: chenjiankun <chenjiankun1@huawei.com>
|
|
|
|
Change-Id: I063fd37fe25754c94d164ae5a209d15b69322093
Signed-off-by: Deepak S <deepak.s@linux.intel.com>
|
|
|
|
due to https://gerrit.opnfv.org/gerrit/#/c/27605/
os-odl_l2-bgpvpn-ha in apex has already been renamed to
be os-odl-bgpvpn-ha, so add a new one.
Change-Id: I167bf295b67b2dddbd3da2aa5a2dca666b3fa152
Signed-off-by: rexlee8776 <limingjiang@huawei.com>
|
|
new scenario os-odl-gluon-noha in apex pod
Change-Id: I0151ce227e2a3ff476044f0b5c8d4ed61d20b21a
Signed-off-by: rexlee8776 <limingjiang@huawei.com>
|
|
|