aboutsummaryrefslogtreecommitdiffstats
path: root/func/args_handler.py
blob: 597128002d6a8f90a9d3f4c91806bbb4583c7a7f (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
##############################################################################
# Copyright (c) 2016 ZTE Corp and others.
#
# All rights reserved. This program and the accompanying materials
# are made available under the terms of the Apache License, Version 2.0
# which accompanies this distribution, and is available at
# http://www.apache.org/licenses/LICENSE-2.0
##############################################################################
import os
from operator import add
import simplejson as json
from func.env_setup import Env_setup
from func.spawn_vm import SpawnVM
from func.driver import Driver


def get_files_in_test_list(suite_name, case_type='all'):
    benchmark_list = json.load(file('test_list/{0}'.format(suite_name)))
    return reduce(add, benchmark_list.values()) \
        if case_type == 'all' else benchmark_list[case_type]


def get_files_in_test_case(lab, suite_name, case_type='all'):
    test_case_all = os.listdir('./test_cases/{0}/{1}'.format(lab, suite_name))
    return test_case_all if case_type == 'all' else \
        filter(lambda x: case_type in x, test_case_all)


def get_benchmark_path(lab, suit, benchmark):
    return './test_cases/{0}/{1}/{2}'.format(lab, suit, benchmark)


def check_suite_in_test_list(suite_name):
    return True if os.path.isfile('test_list/' + suite_name) else False


def check_lab_name(lab_name):
    return True if os.path.isdir('test_cases/' + lab_name) else False


def check_benchmark_name(lab, file, benchmark):
    return os.path.isfile('test_cases/' + lab + '/' + file + '/' + benchmark)


def _get_f_name(test_case_path):
    return test_case_path.split('/')[-1]


def prepare_ansible_env(benchmark_test_case):
    env_setup = Env_setup()
    [benchmark, vm_info, benchmark_details, proxy_info] = env_setup.parse(benchmark_test_case)
    SpawnVM(vm_info) if len(vm_info) else None
    env_setup.call_ping_test()
    env_setup.call_ssh_test()
    env_setup.update_ansible()
    return benchmark, benchmark_details, proxy_info, env_setup


def run_benchmark(installer_type, pwd, benchmark, benchmark_details,
                  proxy_info, env_setup, benchmark_test_case):
    driver = Driver()
    return driver.drive_bench(installer_type, pwd, benchmark,
                              env_setup.roles_dict.items(),
                              _get_f_name(benchmark_test_case),
                              benchmark_details, env_setup.ip_pw_dict.items(), proxy_info)


def prepare_and_run_benchmark(installer_type, pwd, benchmark_test_case):
    benchmark, benchmark_details, proxy_info, env_setup = prepare_ansible_env(benchmark_test_case)
    return run_benchmark(installer_type, pwd, benchmark, benchmark_details,
                         proxy_info, env_setup, benchmark_test_case)