summaryrefslogtreecommitdiffstats
path: root/yardstick/benchmark/scenarios/availability/serviceha.py
diff options
context:
space:
mode:
Diffstat (limited to 'yardstick/benchmark/scenarios/availability/serviceha.py')
-rwxr-xr-xyardstick/benchmark/scenarios/availability/serviceha.py159
1 files changed, 48 insertions, 111 deletions
diff --git a/yardstick/benchmark/scenarios/availability/serviceha.py b/yardstick/benchmark/scenarios/availability/serviceha.py
index 3e03e1da5..10134ea6d 100755
--- a/yardstick/benchmark/scenarios/availability/serviceha.py
+++ b/yardstick/benchmark/scenarios/availability/serviceha.py
@@ -6,13 +6,11 @@
# which accompanies this distribution, and is available at
# http://www.apache.org/licenses/LICENSE-2.0
##############################################################################
-import pkg_resources
import logging
import time
-import yaml
-import yardstick.ssh as ssh
from yardstick.benchmark.scenarios import base
from yardstick.benchmark.scenarios.availability import monitor
+from yardstick.benchmark.scenarios.availability.attacker import baseattacker
LOG = logging.getLogger(__name__)
@@ -22,87 +20,34 @@ class ServiceHA(base.Scenario):
"""
__scenario_type__ = "ServiceHA"
- HA_CONF = "ha_tools/ha_conf.yaml"
-
def __init__(self, scenario_cfg, context_cfg):
+ LOG.debug(
+ "scenario_cfg:%s context_cfg:%s" %
+ (scenario_cfg, context_cfg))
self.scenario_cfg = scenario_cfg
self.context_cfg = context_cfg
- self.service_name = scenario_cfg["options"]["component"]
- self.fault_type = scenario_cfg["options"]["fault_type"]
- self.fault_time = scenario_cfg["options"].get("fault_time", 0)
- self.fault_cfg = None
self.setup_done = False
- self.need_teardown = False
def setup(self):
'''scenario setup'''
- self.ha_conf_file = pkg_resources.resource_filename(
- "yardstick.benchmark.scenarios.availability",
- ServiceHA.HA_CONF)
- ha_cfg = []
- with open(self.ha_conf_file) as stream:
- ha_cfg = yaml.load(stream)
- LOG.debug("ha_cfg content:%s" % ha_cfg)
-
- # check the ha_conf contains the service defined in test cases yaml
- service_cfg = ha_cfg.get(self.service_name, None)
- if not service_cfg:
- LOG.error(
- "The component %s can not be supported!" % self.service_name)
- return
-
- for fault in service_cfg:
- if fault["type"] == self.fault_type:
- self.fault_cfg = fault
- break
- if not self.fault_cfg:
- LOG.error(
- "The fualt_type %s can not be supproted!" % self.fault_type)
- return
- LOG.debug("the fault_cfg :%s" % self.fault_cfg)
-
- self.fault_script = pkg_resources.resource_filename(
- "yardstick.benchmark.scenarios.availability",
- self.fault_cfg["inject_script"])
- self.recovery_script = pkg_resources.resource_filename(
- "yardstick.benchmark.scenarios.availability",
- self.fault_cfg["recovery_script"])
- self.check_script = pkg_resources.resource_filename(
- "yardstick.benchmark.scenarios.availability",
- self.fault_cfg["check_script"])
-
- host = self.context_cfg.get("host", None)
- ip = host.get("ip", None)
- user = host.get("user", "root")
- key_filename = host.get("key_filename", "~/.ssh/id_rsa")
- LOG.info("The host: %s the service: %s" % (ip, self.service_name))
- LOG.debug("The params, host:%s fault_cfg:%s" % (host, self.fault_cfg))
-
- LOG.debug(
- "ssh connection ip:%s, user:%s, key_file:%s",
- ip, user, key_filename)
- self.connection = ssh.SSH(user, ip, key_filename=key_filename)
- self.connection.wait(timeout=600)
- LOG.debug("ssh host success!")
-
- # check the host envrioment
- exit_status, stdout, stderr = self.connection.execute(
- "/bin/sh -s {0}".format(self.service_name),
- stdin=open(self.check_script, "r"))
- LOG.info(
- "the exit_status:%s stdout:%s stderr:%s" %
- (exit_status, stdout, stderr))
- if exit_status:
- raise RuntimeError(stderr)
-
- if stdout and "running" in stdout:
- LOG.info("check the envrioment success!")
- else:
- LOG.error(
- "the host envrioment is error, stdout:%s, stderr:%s" %
- (stdout, stderr))
+ nodes = self.context_cfg.get("nodes", None)
+ if nodes is None:
+ LOG.error("the nodes info is none")
return
-
+ self.attackers = []
+ attacker_cfgs = self.scenario_cfg["options"]["attackers"]
+ for attacker_cfg in attacker_cfgs:
+ attacker_cls = baseattacker.BaseAttacker.get_attacker_cls(
+ attacker_cfg)
+ attacker_ins = attacker_cls(attacker_cfg, nodes)
+ attacker_ins.setup()
+ self.attackers.append(attacker_ins)
+
+ monitor_cfgs = self.scenario_cfg["options"]["monitors"]
+
+ self.monitor_ins = monitor.Monitor()
+ self.monitor_ins.setup(monitor_cfgs[0])
+ self.monitor_ins.monitor_time = monitor_cfgs[0]["monitor_time"]
self.setup_done = True
def run(self, result):
@@ -111,27 +56,18 @@ class ServiceHA(base.Scenario):
LOG.error("The setup not finished!")
return
- monitorInstance = monitor.Monitor()
- monitorInstance.setup(self.fault_cfg)
- monitorInstance.start()
+ self.monitor_ins.start()
LOG.info("monitor start!")
- LOG.info("Inject fault!")
- exit_status, stdout, stderr = self.connection.execute(
- "/bin/sh -s {0}".format(self.service_name),
- stdin=open(self.fault_script, "r"))
-
- if exit_status != 0:
- monitorInstance.stop()
- raise RuntimeError(stderr)
+ for attacker in self.attackers:
+ attacker.inject_fault()
- self.need_teardown = True
- time.sleep(self.fault_time)
+ time.sleep(self.monitor_ins.monitor_time)
- monitorInstance.stop()
+ self.monitor_ins.stop()
LOG.info("monitor stop!")
- ret = monitorInstance.get_result()
+ ret = self.monitor_ins.get_result()
LOG.info("The monitor result:%s" % ret)
outage_time = ret.get("outage_time")
result["outage_time"] = outage_time
@@ -146,34 +82,36 @@ class ServiceHA(base.Scenario):
def teardown(self):
'''scenario teardown'''
- LOG.info("recory the everiment!")
-
- if self.need_teardown:
- exit_status, stdout, stderr = self.connection.execute(
- "/bin/sh -s {0} ".format(self.service_name),
- stdin=open(self.recovery_script, "r"))
+ for attacker in self.attackers:
+ if not attacker.check():
+ attacker.recover()
- if exit_status:
- raise RuntimeError(stderr)
- else:
- self.need_teardown = False
-"""
-def _test():
+def _test(): # pragma: no cover
'''internal test function'''
host = {
"ip": "10.20.0.5",
"user": "root",
"key_filename": "/root/.ssh/id_rsa"
}
- ctx = {"host": host}
-
- logger = logging.getLogger("yardstick")
- logger.setLevel(logging.DEBUG)
+ ctx = {"nodes": {"node1": host}}
+ attacker_cfg = {
+ "fault_type": "kill-process",
+ "process_name": "nova-api",
+ "host": "node1"
+ }
+ attacker_cfgs = []
+ attacker_cfgs.append(attacker_cfg)
+ monitor_cfg = {
+ "monitor_cmd": "nova image-list",
+ "monitor_tme": 10
+ }
+ monitor_cfgs = []
+ monitor_cfgs.append(monitor_cfg)
options = {
- "component": "nova-api",
- "fault_type": "stop-service"
+ "attackers": attacker_cfgs,
+ "monitors": monitor_cfgs
}
sla = {"outage_time": 5}
args = {"options": options, "sla": sla}
@@ -188,6 +126,5 @@ def _test():
terstInstance.teardown()
-if __name__ == '__main__':
+if __name__ == '__main__': # pragma: no cover
_test()
-"""