summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--samples/ping-multiple-context.yaml56
-rw-r--r--yardstick/benchmark/context/model.py21
-rwxr-xr-xyardstick/main.py54
-rw-r--r--yardstick/orchestrator/heat.py19
4 files changed, 118 insertions, 32 deletions
diff --git a/samples/ping-multiple-context.yaml b/samples/ping-multiple-context.yaml
new file mode 100644
index 000000000..8e1b1b700
--- /dev/null
+++ b/samples/ping-multiple-context.yaml
@@ -0,0 +1,56 @@
+---
+# Sample benchmark task config file to measure network latency using ping
+# Client/server parts are located in different contexts(stacks)
+# => longer communication path and higher latency
+
+schema: "yardstick:task:0.1"
+
+scenarios:
+-
+ type: Ping
+ options:
+ packetsize: 100
+ host: client.demo1
+ target: server.demo2
+ runner:
+ type: Duration
+ duration: 60
+ interval: 1
+ sla:
+ max_rtt: 15
+ action: monitor
+
+contexts:
+-
+ name: demo1
+ image: cirros-0.3.3
+ flavor: m1.tiny
+ user: cirros
+ placement_groups:
+ pgrp1:
+ policy: "availability"
+ servers:
+ client:
+ floating_ip: true
+ placement: "pgrp1"
+ networks:
+ test:
+ cidr: '10.0.1.0/24'
+ external_network: "net04_ext"
+-
+ name: demo2
+ image: cirros-0.3.3
+ flavor: m1.tiny
+ user: cirros
+ placement_groups:
+ pgrp1:
+ policy: "availability"
+ servers:
+ server:
+ floating_ip: true
+ placement: "pgrp1"
+ networks:
+ test:
+ cidr: '10.0.1.0/24'
+ external_network: "net04_ext"
+
diff --git a/yardstick/benchmark/context/model.py b/yardstick/benchmark/context/model.py
index 768c1d381..afb2d56dd 100644
--- a/yardstick/benchmark/context/model.py
+++ b/yardstick/benchmark/context/model.py
@@ -123,6 +123,7 @@ class Server(Object):
self.stack_name = context.name + "-" + self.name
self.keypair_name = context.keypair_name
self.secgroup_name = context.secgroup_name
+ self.context = context
if attrs is None:
attrs = {}
@@ -372,8 +373,8 @@ class Context(object):
except Exception as err:
sys.exit("error: failed to deploy stack: '%s'" % err)
- # copy some vital stack output into context
- for server in Server.list:
+ # Iterate the servers in this context and copy out needed info
+ for server in self.servers:
for port in server.ports.itervalues():
port["ipaddr"] = self.stack.outputs[port["stack_name"]]
@@ -391,6 +392,16 @@ class Context(object):
self.stack = None
print "Context undeployed"
- def get_server(self, name):
- '''lookup server object by name from context'''
- return self._server_map[name]
+ @staticmethod
+ def get_server(dn):
+ '''lookup server object by DN
+
+ dn is a distinguished name including the context name'''
+ if "." not in dn:
+ raise ValueError("dn '%s' is malformed" % dn)
+
+ for context in Context.list:
+ if dn in context._server_map:
+ return context._server_map[dn]
+
+ return None
diff --git a/yardstick/main.py b/yardstick/main.py
index 050a56436..5669fde10 100755
--- a/yardstick/main.py
+++ b/yardstick/main.py
@@ -23,7 +23,6 @@ import pkg_resources
from yardstick.benchmark.context.model import Context
from yardstick.benchmark.runners import base as base_runner
from yardstick.cmdparser import CmdParser
-from yardstick.orchestrator.heat import HeatStack
class TaskParser(object):
@@ -43,43 +42,58 @@ class TaskParser(object):
if cfg["schema"] != "yardstick:task:0.1":
sys.exit("error: file %s has unknown schema %s" % (self.path,
cfg["schema"]))
- context = Context()
- context.init(cfg["context"])
+
+ # TODO: support one or many contexts? Many would simpler and precise
+ if "context" in cfg:
+ context_cfgs = [cfg["context"]]
+ else:
+ context_cfgs = cfg["contexts"]
+
+ for cfg_attrs in context_cfgs:
+ context = Context()
+ context.init(cfg_attrs)
run_in_parallel = cfg.get("run_in_parallel", False)
# TODO we need something better here, a class that represent the file
- return cfg["scenarios"], run_in_parallel, context
+ return cfg["scenarios"], run_in_parallel
def atexit_handler():
'''handler for process termination'''
base_runner.Runner.terminate_all()
- if HeatStack.stacks_exist():
- print "Deleting all stacks"
- HeatStack.delete_all()
+ if len(Context.list) > 0:
+ print "Undeploying all contexts"
+ for context in Context.list:
+ context.undeploy()
-def run_one_scenario(scenario_cfg, context, output_file):
+def run_one_scenario(scenario_cfg, output_file):
'''run one scenario using context'''
key_filename = pkg_resources.resource_filename(
'yardstick.resources', 'files/yardstick_key')
- host = context.get_server(scenario_cfg["host"])
+ host = Context.get_server(scenario_cfg["host"])
runner_cfg = scenario_cfg["runner"]
runner_cfg['host'] = host.floating_ip["ipaddr"]
- runner_cfg['user'] = context.user
+ runner_cfg['user'] = host.context.user
runner_cfg['key_filename'] = key_filename
runner_cfg['output_filename'] = output_file
- target = context.get_server(scenario_cfg["target"])
+ # TODO target should be optional to support single VM scenarios
+ target = Context.get_server(scenario_cfg["target"])
if target.floating_ip:
runner_cfg['target'] = target.floating_ip["ipaddr"]
- # TODO hardcoded name below, a server can be attached to several nets
- scenario_cfg["ipaddr"] = target.ports["test"]["ipaddr"]
+ # TODO scenario_cfg["ipaddr"] is bad, "dest_ip" is better
+ if host.context != target.context:
+ # target is in another context, get its public IP
+ scenario_cfg["ipaddr"] = target.floating_ip["ipaddr"]
+ else:
+ # TODO hardcoded name below, a server can be attached to several nets
+ scenario_cfg["ipaddr"] = target.ports["test"]["ipaddr"]
runner = base_runner.Runner.get(runner_cfg)
@@ -105,17 +119,18 @@ def main():
prog_args = CmdParser().parse_args()
parser = TaskParser(prog_args.taskfile[0])
- scenarios, run_in_parallel, context = parser.parse()
+ scenarios, run_in_parallel = parser.parse()
if prog_args.parse_only:
sys.exit(0)
- context.deploy()
+ for context in Context.list:
+ context.deploy()
runners = []
if run_in_parallel:
for scenario in scenarios:
- runner = run_one_scenario(scenario, context, prog_args.output_file)
+ runner = run_one_scenario(scenario, prog_args.output_file)
runners.append(runner)
# Wait for runners to finish
@@ -125,15 +140,16 @@ def main():
else:
# run serially
for scenario in scenarios:
- runner = run_one_scenario(scenario, context, prog_args.output_file)
+ runner = run_one_scenario(scenario, prog_args.output_file)
runner_join(runner)
print "Runner ended, output in", prog_args.output_file
if prog_args.keep_deploy:
# keep deployment, forget about stack (hide it for exit handler)
- context.stack = None
+ Context.list = []
else:
- context.undeploy()
+ for context in Context.list:
+ context.undeploy()
print "Done, exiting"
diff --git a/yardstick/orchestrator/heat.py b/yardstick/orchestrator/heat.py
index 9c0d0f1a4..ddab89640 100644
--- a/yardstick/orchestrator/heat.py
+++ b/yardstick/orchestrator/heat.py
@@ -66,9 +66,9 @@ class HeatStack(HeatObject):
''' Represents a Heat stack (deployed template) '''
stacks = []
- def __init__(self, uuid, name):
+ def __init__(self, name):
super(HeatStack, self).__init__()
- self.uuid = uuid
+ self.uuid = None
self.name = name
self.outputs = None
HeatStack.stacks.append(self)
@@ -80,6 +80,9 @@ class HeatStack(HeatObject):
def _delete(self):
'''deletes a stack from the target cloud using heat'''
+ if self.uuid is None:
+ return
+
log.info("Deleting stack '%s', uuid:%s", self.name, self.uuid)
heat = self._get_heat_client()
template = heat.stacks.get(self.uuid)
@@ -131,7 +134,7 @@ class HeatStack(HeatObject):
@staticmethod
def delete_all():
- for stack in HeatStack.stacks:
+ for stack in HeatStack.stacks[:]:
stack.delete()
def update(self):
@@ -394,14 +397,14 @@ class HeatTemplate(HeatObject):
returns a dict with the requested output values from the template'''
log.info("Creating stack '%s'", self.name)
+ # create stack early to support cleanup, e.g. ctrl-c while waiting
+ stack = HeatStack(self.name)
+
heat = self._get_heat_client()
json_template = json.dumps(self._template)
start_time = time.time()
- self.uuid = heat.stacks.create(stack_name=self.name,
- template=json_template)['stack']['id']
-
- # create stack early to support cleanup, e.g. ctrl-c while waiting
- stack = HeatStack(self.uuid, self.name)
+ stack.uuid = self.uuid = heat.stacks.create(
+ stack_name=self.name, template=json_template)['stack']['id']
status = self.status()