import requests import docker import concurrent.futures import time import sys import os from graph import generate_req_graph, generate_resource_graph from math import floor if len(sys.argv) != 3 or sys.argv[1] == '-h' or sys.argv[1] == '--help': print("Usage: python testes.py ") sys.exit(1) THREADS = 10 FRAMEWORK_NAME = sys.argv[1] CONTAINER_NAME = sys.argv[2] URL_BASE = 'http://172.26.48.1:9080' API_REQUESTS = [ ('/status/ok', range(0, 30_000, 5000)), ('/image/load-image', range(0, 30_000, 5000)), ('/static/simpleimage.png', range(0, 30_000, 5000)), ('/image/load-big-image', range(0, 1_000, 200)), ('/static/bigimage.png', range(0, 1_000, 200)), ('/static/video.mp4', range(0, 10_000, 1_000)), ] def send_request(url): success = False responses = { 2: 0, # OK 4: 0, # Bad Request 5: 0, # Server Error } while not success: try: response = requests.get(url) except: continue success = response.status_code == 200 responses[floor(response.status_code/100)] += 1 return responses def getFileNames(endpoint): endpoint = endpoint.replace('/', '') files = [ f"req_{FRAMEWORK_NAME}_{endpoint}.csv", f"resource_{FRAMEWORK_NAME}_{endpoint}.csv", ] return files def record(filename, requests, reqpersec): with open(filename, "a") as file: file.write(f"{requests},{reqpersec}\n") def record_resource(filename, requests, cpu, ram): with open(filename, "a") as file: file.write(f"{requests},{cpu},{ram}\n") def run_tests(endpoint, num_requests): files = getFileNames(endpoint) for filename in files: if os.path.exists(filename): os.remove(filename) for num_request in num_requests: if num_request <= 0: continue ok_responses = 0 bad_responses = 0 server_errors = 0 cpu, ram = 0, 0 with concurrent.futures.ThreadPoolExecutor(max_workers=THREADS) as executor: url = f'{URL_BASE}{endpoint}' start_time = time.time() futures = [] #with requests.Session() as session: # futures = [executor.submit(send_request, session, url) for _ in range(num_request)] futures = [executor.submit(send_request, url) for _ in range(num_request)] concurrent.futures.wait(futures) cpu, ram = get_resource_usage() elapsed_time = time.time() - start_time for future in futures: responses = future.result() ok_responses += responses[2] bad_responses += responses[4] server_errors += responses[5] print(f"{num_request}: {elapsed_time:.2f} seconds. {elapsed_time/num_request:.4f} seconds per request. {num_request/elapsed_time:.2f} requests per second. [OK: {ok_responses}, Bad Request: {bad_responses}, Server Error: {server_errors}]]") record(files[0], num_request, f"{num_request/elapsed_time:.2f}") record_resource(files[1], num_request, cpu, ram) generate_req_graph(files[0], FRAMEWORK_NAME, endpoint) generate_resource_graph(files[1], FRAMEWORK_NAME, endpoint) time.sleep(3) def get_resource_usage(): try: client = docker.from_env() stats = client.containers.get(CONTAINER_NAME).stats(stream=False) except: return 0, 0 # unable to get stats return get_cpu_usage(stats), get_ram_usage(stats) def get_cpu_usage(stats): UsageDelta = stats['cpu_stats']['cpu_usage']['total_usage'] - stats['precpu_stats']['cpu_usage']['total_usage'] SystemDelta = stats['cpu_stats']['system_cpu_usage'] - stats['precpu_stats']['system_cpu_usage'] len_cpu = stats['cpu_stats']['online_cpus'] percentage = (UsageDelta / SystemDelta) * len_cpu * 100 return f"{percentage:.2f}" def get_ram_usage(stats): usage = stats['memory_stats']['usage'] limit = stats['memory_stats']['limit'] percentage = (usage / limit) * 100 # percent = round(percentage, 2) return f"{percentage:.2f}" for endpoint, num_requests in API_REQUESTS: print(f"# {endpoint}") run_tests(endpoint, num_requests)