diff options
author | asteroide <thomas.duval@orange.com> | 2017-07-06 15:37:03 +0200 |
---|---|---|
committer | asteroide <thomas.duval@orange.com> | 2017-07-06 15:37:03 +0200 |
commit | 5cc39c4b3a2de6d670f357ed3329a4d6587c7fc7 (patch) | |
tree | 76f4220f458f073f24e3d3a2fc11d66d397f697e /moonv4/moon_interface/tests/apitests/set_authz.py | |
parent | 19dd4e461d1239754bd1a82e30f2a770c99a7f90 (diff) |
Update test and visualization scripts
Change-Id: I62d50d8d3e839a1d6c8ff948cc84ecc45bb0e92f
Diffstat (limited to 'moonv4/moon_interface/tests/apitests/set_authz.py')
-rw-r--r-- | moonv4/moon_interface/tests/apitests/set_authz.py | 355 |
1 files changed, 214 insertions, 141 deletions
diff --git a/moonv4/moon_interface/tests/apitests/set_authz.py b/moonv4/moon_interface/tests/apitests/set_authz.py index 9ac13142..270c9de2 100644 --- a/moonv4/moon_interface/tests/apitests/set_authz.py +++ b/moonv4/moon_interface/tests/apitests/set_authz.py @@ -8,6 +8,7 @@ import itertools import requests import time import json +import random import plotly from plotly.graph_objs import Scatter, Layout import plotly.figure_factory as ff @@ -18,29 +19,44 @@ from utils.pdp import check_pdp logger = None HOST = None PORT = None +HOST_AUTHZ = None +PORT_AUTHZ = None lock = threading.Lock() def init(): - global logger, HOST, PORT + global logger, HOST, PORT, HOST_AUTHZ, PORT_AUTHZ parser = argparse.ArgumentParser() parser.add_argument('filename', help='scenario filename', nargs=1) parser.add_argument("--verbose", "-v", action='store_true', help="verbose mode") + parser.add_argument("--debug", action='store_true', help="debug mode") + parser.add_argument("--dry-run", "-n", action='store_true', help="Dry run", dest="dry_run") parser.add_argument("--host", help="Set the name of the host to test (default: 172.18.0.11).", default="172.18.0.11") + parser.add_argument("--host-authz", + help="Set the name of the host to test authorization (default: 172.18.0.11).", + default="172.18.0.11", + dest="host_authz") parser.add_argument("--port", "-p", help="Set the port of the host to test (default: 38001).", default="38001") + parser.add_argument("--port-authz", + help="Set the port of the host to test authorization (default: 38001).", + default="38001", + dest="port_authz") parser.add_argument("--test-only", "-t", action='store_true', dest='testonly', help="Do not generate graphs") + parser.add_argument("--stress-test", "-s", action='store_true', dest='stress_test', + help="Execute stressing tests (warning delta measures will be false, implies -t)") parser.add_argument("--write", "-w", help="Write test data to a JSON file", default="/tmp/data.json") + parser.add_argument("--pdp", help="Test on pdp PDP") parser.add_argument("--input", "-i", help="Get data from a JSON input file") parser.add_argument("--legend", "-l", help="Set the legend (default: 'rbac,rbac+session')", default='rbac,rbac+session') - parser.add_argument("--distgraph", "-d", - help="Show a distribution graph instead of a linear graph", - action='store_true') + # parser.add_argument("--distgraph", "-d", + # help="Show a distribution graph instead of a linear graph", + # action='store_true') parser.add_argument("--request-per-second", help="Number of requests per seconds", type=int, dest="request_second", default=1) parser.add_argument("--limit", help="Limit request to LIMIT", type=int) @@ -49,17 +65,35 @@ def init(): args = parser.parse_args() FORMAT = '%(asctime)-15s %(levelname)s %(message)s' - logging.basicConfig( - format=FORMAT, - level=logging.INFO) + if args.debug: + logging.basicConfig( + format=FORMAT, + level=logging.DEBUG) + elif args.verbose: + logging.basicConfig( + format=FORMAT, + level=logging.INFO) + else: + logging.basicConfig( + format=FORMAT, + level=logging.WARNING) logger = logging.getLogger(__name__) + requests_log = logging.getLogger("requests.packages.urllib3") + requests_log.setLevel(logging.WARNING) + requests_log.propagate = True + + if args.stress_test: + args.testonly = True + if args.filename: logger.info("Loading: {}".format(args.filename[0])) HOST = args.host PORT = args.port + HOST_AUTHZ = args.host_authz + PORT_AUTHZ = args.port_authz return args @@ -68,12 +102,16 @@ def get_scenario(args): return m.load_module() -def get_keystone_id(): +def get_keystone_id(pdp_name): keystone_project_id = None + logger.error("get_keystone_id url={}".format("http://{}:{}".format(HOST, PORT))) for pdp_key, pdp_value in check_pdp(moon_url="http://{}:{}".format(HOST, PORT))["pdps"].items(): - print(pdp_value) + logger.debug(pdp_value) + if pdp_name: + if pdp_name != pdp_value["name"]: + continue if pdp_value['security_pipeline'] and pdp_value["keystone_project_id"]: - print("Found pdp with keystone_project_id={}".format(pdp_value["keystone_project_id"])) + logger.debug("Found pdp with keystone_project_id={}".format(pdp_value["keystone_project_id"])) keystone_project_id = pdp_value["keystone_project_id"] if not keystone_project_id: @@ -82,7 +120,50 @@ def get_keystone_id(): return keystone_project_id +def _send(url, stress_test=False): + current_request = dict() + current_request['url'] = url + try: + if stress_test: + current_request['start'] = time.time() + # with lock: + res = requests.get(url) + current_request['end'] = time.time() + current_request['delta'] = current_request["end"] - current_request["start"] + else: + with lock: + current_request['start'] = time.time() + res = requests.get(url) + current_request['end'] = time.time() + current_request['delta'] = current_request["end"] - current_request["start"] + except requests.exceptions.ConnectionError: + logger.warning("Unable to connect to server") + return {} + if res and not stress_test: + logger.debug(res.status_code) + logger.debug(res.text) + if res.status_code == 200: + # logger.warning("error code 200 for {}".format(self.url)) + logger.info("\033[1m{}\033[m {}".format(url, res.status_code)) + try: + j = res.json() + except Exception as e: + logger.debug(e) + logger.error(res.text) + else: + if j.get("authz"): + logger.warning("{} \033[32m{}\033[m".format(url, j.get("authz"))) + logger.debug("{}".format(j.get("error", ""))) + current_request['result'] = "Grant" + else: + logger.warning("{} \033[31m{}\033[m".format(url, j.get("authz"))) + logger.debug("{}".format(j.get("error", ""))) + current_request['result'] = "Deny" + return current_request + + class AsyncGet(threading.Thread): + def __init__(self, url, semaphore=None, *args, **kwargs): threading.Thread.__init__(self) self.url = url @@ -90,67 +171,53 @@ class AsyncGet(threading.Thread): self.sema = semaphore self.result = dict() self.uuid = uuid4().hex + self.index = kwargs.get("index", 0) def run(self): - - # self.sema.acquire() - current_request = dict() - current_request['url'] = self.url - try: - with lock: - current_request['start'] = time.time() - r = requests.get(self.url, **self.kwargs) - current_request['end'] = time.time() - current_request['delta'] = current_request["end"] - current_request["start"] - except requests.exceptions.ConnectionError: - logger.warning("Unable to connect to server") - return {} - if r: - logger.debug(r.status_code) - logger.debug(r.text) - if r.status_code == 200: - # logger.warning("error code 200 for {}".format(self.url)) - logger.info("\033[1m{}\033[m {}".format(self.url, r.status_code)) - try: - j = r.json() - except Exception as e: - logger.error(r.text) - else: - if j.get("authz"): - logger.info("\t\033[32m{}\033[m {}".format(j.get("authz"), j.get("error", ""))) - current_request['result'] = "Grant" - else: - logger.info("\t\033[31m{}\033[m {}".format(j.get("authz"), j.get("error", ""))) - current_request['result'] = "Deny" - self.result = current_request - # self.sema.release() + self.result = _send(self.url, self.kwargs.get("stress_test", False)) + self.result['index'] = self.index -def send_requests(scenario, keystone_project_id, request_second=1, limit=None): +def send_requests(scenario, keystone_project_id, request_second=1, limit=500, + dry_run=None, stress_test=False): # sema = threading.BoundedSemaphore(value=request_second) backgrounds = [] - time_data = dict() + time_data = list() start_timing = time.time() request_cpt = 0 - rules = itertools.product(scenario.subjects.keys(), scenario.objects.keys(), scenario.actions.keys()) - for rule in rules: - url = "http://{}:{}/authz/{}/{}".format(HOST, PORT, keystone_project_id, "/".join(rule)) + indexes = [] + # rules = itertools.product(scenario.subjects.keys(), scenario.objects.keys(), scenario.actions.keys()) + SUBJECTS = tuple(scenario.subjects.keys()) + OBJECTS = tuple(scenario.objects.keys()) + ACTIONS = tuple(scenario.actions.keys()) + # for rule in rules: + while request_cpt <= limit: + rule = (random.choice(SUBJECTS), random.choice(OBJECTS), random.choice(ACTIONS)) + url = "http://{}:{}/authz/{}/{}".format(HOST_AUTHZ, PORT_AUTHZ, keystone_project_id, "/".join(rule)) + indexes.append(url) + if dry_run: + logger.info(url) + continue request_cpt += 1 - background = AsyncGet(url) - backgrounds.append(background) - background.start() - if limit and limit < request_cpt: - break + if stress_test: + time_data.append(copy.deepcopy(_send(url, stress_test=stress_test))) + else: + background = AsyncGet(url, stress_test=stress_test, index=request_cpt) + backgrounds.append(background) + background.start() + # if limit and limit < request_cpt: + # break if request_cpt % request_second == 0: if time.time()-start_timing < 1: while True: if time.time()-start_timing > 1: break start_timing = time.time() - for background in backgrounds: - background.join() - if background.result: - time_data[background.url] = copy.deepcopy(background.result) + if not stress_test: + for background in backgrounds: + background.join() + if background.result: + time_data.append(copy.deepcopy(background.result)) return time_data @@ -161,10 +228,10 @@ def save_data(filename, time_data): def get_delta(time_data): time_delta = list() time_delta_sum1 = 0 - for key in time_data: - time_delta.append(time_data[key]['delta']) - time_delta_sum1 += time_data[key]['delta'] - time_delta_average1 = time_delta_sum1 / len(time_data.keys()) + for item in time_data: + time_delta.append(item['delta']) + time_delta_sum1 += item['delta'] + time_delta_average1 = time_delta_sum1 / len(time_data) return time_delta, time_delta_average1 @@ -174,51 +241,49 @@ def write_graph(time_data, legend=None, input=None, image_file=None, html_file=N result_data = [] time_delta, time_delta_average1 = get_delta(time_data) time_delta_average2 = None - if input: - for _input in input.split(","): - current_legend = legends.pop(0) - time_data2 = json.load(open(_input)) - time_delta2, time_delta_average2 = get_delta(time_data2) - for key in time_data.keys(): - if key in time_data2: - time_delta2.append(time_data2[key]['delta']) - else: - time_delta2.append(None) - data2 = Scatter( - x=list(range(len(time_data.keys()))), - y=time_delta2, - name=current_legend, - line=dict( - color='rgb(255, 192, 118)', - shape='spline') - ) - result_data.append(data2) - data2_a = Scatter( - x=list(range(len(time_data.keys()))), - y=[time_delta_average2 for x in range(len(time_data.keys()))], - name=current_legend + " average", - line=dict( - color='rgb(255, 152, 33)', - shape='spline') - ) - result_data.append(data2_a) + # if input: + # for _input in input.split(","): + # current_legend = legends.pop(0) + # time_data2 = json.load(open(_input)) + # time_delta2, time_delta_average2 = get_delta(time_data2) + # for item in time_data: + # if key in time_data2: + # time_delta2.append(time_data2[key]['delta']) + # else: + # time_delta2.append(None) + # data2 = Scatter( + # x=list(range(len(time_data))), + # y=time_delta2, + # name=current_legend, + # line=dict( + # color='rgb(255, 192, 118)', + # shape='spline') + # ) + # result_data.append(data2) + # data2_a = Scatter( + # x=list(range(len(time_data))), + # y=[time_delta_average2 for x in range(len(time_data))], + # name=current_legend + " average", + # line=dict( + # color='rgb(255, 152, 33)', + # shape='spline') + # ) + # result_data.append(data2_a) current_legend = legends.pop(0) data1 = Scatter( - x=list(range(len(time_data.keys()))), + x=list(range(len(time_data))), y=time_delta, name=current_legend, line=dict( - color='rgb(123, 118, 255)', - shape='spline') + color='rgb(123, 118, 255)') ) result_data.append(data1) data1_a = Scatter( - x=list(range(len(time_data.keys()))), - y=[time_delta_average1 for x in range(len(time_data.keys()))], + x=list(range(len(time_data))), + y=[time_delta_average1 for x in range(len(time_data))], name=current_legend + " average", line=dict( - color='rgb(28, 20, 255)', - shape='spline') + color='rgb(28, 20, 255)') ) result_data.append(data1_a) @@ -256,59 +321,67 @@ def write_graph(time_data, legend=None, input=None, image_file=None, html_file=N return 0 -def write_distgraph(time_data, legend=None, input=None, image_file=None, html_file=None): - - logger.info("Writing graph") - legends = legend.split(",") - result_data = [] - - time_delta_average2 = None - - if input: - for _input in input.split(","): - logger.info("Analysing input {}".format(_input)) - time_data2 = json.load(open(_input)) - time_delta2, time_delta_average2 = get_delta(time_data2) - result_data.append(time_delta2) - - time_delta, time_delta_average1 = get_delta(time_data) - result_data.append(time_delta) - - # Create distplot with custom bin_size - if len(legends) < len(result_data): - for _cpt in range(len(result_data)-len(legends)): - legends.append("NC") - fig = ff.create_distplot(result_data, legends, bin_size=.2) - - # Plot! - plotly.offline.plot( - fig, - image="svg", - image_filename=image_file, - image_height=1000, - image_width=1200, - filename=html_file - ) - if time_delta_average2: - logger.info("Average: {} and {}".format(time_delta_average1, time_delta_average2)) - return 1-time_delta_average2/time_delta_average1 - return 0 +# def write_distgraph(time_data, legend=None, input=None, image_file=None, html_file=None): +# +# logger.info("Writing graph") +# legends = legend.split(",") +# result_data = [] +# +# time_delta_average2 = None +# +# if input: +# for _input in input.split(","): +# logger.info("Analysing input {}".format(_input)) +# time_data2 = json.load(open(_input)) +# time_delta2, time_delta_average2 = get_delta(time_data2) +# result_data.append(time_delta2) +# +# time_delta, time_delta_average1 = get_delta(time_data) +# result_data.append(time_delta) +# +# # Create distplot with custom bin_size +# if len(legends) < len(result_data): +# for _cpt in range(len(result_data)-len(legends)): +# legends.append("NC") +# fig = ff.create_distplot(result_data, legends, bin_size=.2) +# +# # Plot! +# plotly.offline.plot( +# fig, +# image="svg", +# image_filename=image_file, +# image_height=1000, +# image_width=1200, +# filename=html_file +# ) +# if time_delta_average2: +# logger.info("Average: {} and {}".format(time_delta_average1, time_delta_average2)) +# return 1-time_delta_average2/time_delta_average1 +# return 0 def main(): args = init() scenario = get_scenario(args) - keystone_project_id = get_keystone_id() - time_data = send_requests(scenario, keystone_project_id, request_second=args.request_second, limit=args.limit) - save_data(args.write, time_data) - if not args.testonly: - if args.distgraph: - overhead = write_distgraph(time_data, legend=args.legend, input=args.input, image_file=args.write_image, - html_file=args.write_html) - else: + keystone_project_id = get_keystone_id(args.pdp) + time_data = send_requests( + scenario, + keystone_project_id, + request_second=args.request_second, + limit=args.limit, + dry_run=args.dry_run, + stress_test=args.stress_test + ) + if not args.dry_run: + save_data(args.write, time_data) + if not args.testonly: + # if args.distgraph: + # overhead = write_distgraph(time_data, legend=args.legend, input=args.input, image_file=args.write_image, + # html_file=args.write_html) + # else: overhead = write_graph(time_data, legend=args.legend, input=args.input, image_file=args.write_image, html_file=args.write_html) - logger.info("Overhead: {:.2%}".format(overhead)) + logger.info("Overhead: {:.2%}".format(overhead)) if __name__ == "__main__": |