mirror of
https://github.com/ivanch/tcc.git
synced 2025-08-25 07:11:50 +00:00
melhorando testes e adicionando actix
This commit is contained in:
27
scripts/common.py
Normal file
27
scripts/common.py
Normal file
@@ -0,0 +1,27 @@
|
||||
FRAMEWORKS = [
|
||||
('Actix', 'tcc-actix'),
|
||||
('ASP.NET', 'tcc-aspnet'),
|
||||
('Flask', 'tcc-flask'),
|
||||
('Express', 'tcc-express'),
|
||||
]
|
||||
|
||||
ENDPOINTS = {
|
||||
'Actix': 'http://localhost:9083',
|
||||
'ASP.NET': 'http://localhost:9081',
|
||||
'Flask': 'http://localhost:9082',
|
||||
'Express': 'http://localhost:9084',
|
||||
}
|
||||
|
||||
BLUR_RADIUS = 5
|
||||
AVG_RUNS = 3
|
||||
|
||||
API_REQUESTS = [
|
||||
('/status/ok', 'GET', range(0, 30_000, 5000), None),
|
||||
('/image/save-big-image', 'POST', range(0, 10_000, 1_000), open('big-image.png', 'rb').read()),
|
||||
(f'/image/blur?radius={BLUR_RADIUS}', 'POST', range(0, 1_000, 50), open('small-image.png', 'rb').read()),
|
||||
('/image/load-small-image', 'GET', range(0, 30_000, 5000), None),
|
||||
('/static/small-image.png', 'GET', range(0, 30_000, 5000), None),
|
||||
('/image/load-big-image', 'GET', range(0, 500, 50), None),
|
||||
('/static/big-image.png', 'GET', range(0, 500, 50), None),
|
||||
('/static/video.mp4', 'GET', range(0, 10_000, 1_000), None),
|
||||
]
|
@@ -1,11 +1,17 @@
|
||||
import numpy as np
|
||||
import matplotlib.pyplot as plt
|
||||
from common import API_REQUESTS, FRAMEWORKS
|
||||
|
||||
FRAMEWORKS = [f for f, _ in FRAMEWORKS]
|
||||
|
||||
def plot_graph(x_data, y_data, title, x_label, y_label, filename):
|
||||
for i, framework in enumerate(FRAMEWORKS):
|
||||
plt.plot(x_data, y_data[i], markersize=1, linewidth=1, linestyle='solid', label=framework)
|
||||
|
||||
def plot_graph(x, y, title, x_label, y_label, filename):
|
||||
plt.plot(x, y, 'ro', markersize=1, linewidth=0.5, linestyle='solid')
|
||||
plt.title(title)
|
||||
plt.xlabel(x_label)
|
||||
plt.ylabel(y_label)
|
||||
plt.legend()
|
||||
plt.savefig(f'{filename}.png')
|
||||
|
||||
plt.clf()
|
||||
@@ -13,18 +19,22 @@ def plot_graph(x, y, title, x_label, y_label, filename):
|
||||
|
||||
def plot_resource_graph(x_data, y_data, title, x_label, y_label, filename):
|
||||
requests = x_data
|
||||
resource = {
|
||||
'CPU': [p[0] for p in y_data],
|
||||
'RAM': [p[1] for p in y_data],
|
||||
}
|
||||
frameworks = {}
|
||||
print(y_data)
|
||||
for i, framework in enumerate(FRAMEWORKS):
|
||||
frameworks[framework] = y_data[i]
|
||||
|
||||
x = np.arange(len(requests))
|
||||
width = 0.25
|
||||
width = 0.2
|
||||
multiplier = 0
|
||||
|
||||
fig, ax = plt.subplots(layout='constrained')
|
||||
fig, ax = plt.subplots(layout='constrained', figsize=(7.5, 5))
|
||||
|
||||
for attribute, measurement in resource.items():
|
||||
print(x)
|
||||
for framework, measurements in frameworks.items():
|
||||
print(framework, measurements)
|
||||
|
||||
for attribute, measurement in frameworks.items():
|
||||
offset = width * multiplier
|
||||
|
||||
rects = ax.bar(x + offset, measurement, width, label=attribute)
|
||||
@@ -35,9 +45,9 @@ def plot_resource_graph(x_data, y_data, title, x_label, y_label, filename):
|
||||
ax.set_xlabel(x_label)
|
||||
ax.set_ylabel(y_label)
|
||||
ax.set_title(title)
|
||||
ax.set_xticks(x + (width/2), requests)
|
||||
ax.legend(loc='upper left', ncols=len(resource.items()))
|
||||
ax.set_ylim(0, 100)
|
||||
ax.set_xticks(x + (width*1.5), requests)
|
||||
ax.legend(loc='upper left', ncols=len(frameworks.items()))
|
||||
ax.set_ylim(0, 115)
|
||||
|
||||
plt.savefig(f'{filename}.png')
|
||||
|
||||
@@ -69,21 +79,54 @@ def get_resource_data(filename):
|
||||
for line in lines:
|
||||
line = line.strip().split(',')
|
||||
if line:
|
||||
r = [round(float(line[1])*100), round(float(line[2])*100)]
|
||||
if r[0] > 100:
|
||||
r[0] = 100
|
||||
|
||||
if r[1] > 100:
|
||||
r[1] = 100
|
||||
|
||||
x.append(int(line[0])) # requests
|
||||
y.append([float(v)*100 for v in line[1:]]) # cpu, ram
|
||||
y.append(r) # cpu, ram
|
||||
|
||||
return x, y
|
||||
|
||||
def generate_req_graph(filename, framework_name, endpoint_name):
|
||||
x, y = get_data(filename)
|
||||
x, _ = get_data(filename)
|
||||
y = []
|
||||
|
||||
filename = filename.split('/')[-1]
|
||||
new_filename = filename.replace('.csv', '')
|
||||
plot_graph(x, y, f'{framework_name} - {endpoint_name}', 'Número de requisições', 'Requisições por segundo', new_filename)
|
||||
for f in FRAMEWORKS:
|
||||
newfile = filename.replace(framework_name, f)
|
||||
_, y_data = get_data(newfile)
|
||||
|
||||
y.append(y_data)
|
||||
|
||||
graph_file = f'req_{endpoint_name.replace("/", "").replace("?", "")}'
|
||||
plot_graph(x, y, f'Requisições atendidas por segundo - {endpoint_name}', 'Número de requisições', 'Requisições/segundo', graph_file)
|
||||
|
||||
def generate_resource_graph(filename, framework_name, endpoint_name):
|
||||
x, y = get_resource_data(filename)
|
||||
x, _ = get_resource_data(filename)
|
||||
|
||||
filename = filename.split('/')[-1]
|
||||
new_filename = filename.replace('.csv', '')
|
||||
plot_resource_graph(x, y, f'{framework_name} - {endpoint_name}', 'Uso de recursos', 'Uso (%)', new_filename)
|
||||
for resource_index, resource in enumerate(['cpu', 'ram']):
|
||||
y = []
|
||||
|
||||
for f in FRAMEWORKS:
|
||||
newfile = filename.replace(framework_name, f)
|
||||
_, y_data = get_resource_data(newfile)
|
||||
y.append([data[resource_index] for data in y_data])
|
||||
|
||||
graph_file = f'{resource}_{endpoint_name.replace("/", "").replace("?", "")}'
|
||||
plot_resource_graph(x, y, f'Uso de {resource.upper()} - {endpoint_name}', 'Número de requisições', f'Uso de {resource.upper()} (%)', graph_file)
|
||||
|
||||
if __name__ == '__main__':
|
||||
endpoints = [config[0] for config in API_REQUESTS]
|
||||
|
||||
for endpoint_name in endpoints:
|
||||
framework_name = 'ASP.NET'
|
||||
endpoint_file = endpoint_name.replace('/', '')
|
||||
|
||||
filename = f'data/resource_ASP.NET_{endpoint_file}.csv'
|
||||
generate_resource_graph(filename, framework_name, endpoint_name)
|
||||
|
||||
filename = f'data/req_ASP.NET_{endpoint_file}.csv'
|
||||
generate_req_graph(filename, framework_name, endpoint_name)
|
||||
|
@@ -2,33 +2,17 @@ import requests
|
||||
import docker
|
||||
import concurrent.futures
|
||||
import time
|
||||
import sys
|
||||
import os
|
||||
from graph import generate_req_graph, generate_resource_graph
|
||||
from math import floor
|
||||
from init import init
|
||||
|
||||
if len(sys.argv) < 2 or len(sys.argv) > 3 or sys.argv[1] == '-h' or sys.argv[1] == '--help':
|
||||
print("Usage: python testes.py <framework name> [container name]")
|
||||
sys.exit(1)
|
||||
from common import FRAMEWORKS, ENDPOINTS, API_REQUESTS, AVG_RUNS
|
||||
|
||||
init()
|
||||
|
||||
THREADS = 10
|
||||
FRAMEWORK_NAME = sys.argv[1]
|
||||
CONTAINER_NAME = sys.argv[2] if len(sys.argv) > 2 else ""
|
||||
FRAMEWORK_NAME = ""
|
||||
CONTAINER_NAME = ""
|
||||
URL_BASE = 'http://localhost:9090'
|
||||
BLUR_RADIUS = 5
|
||||
API_REQUESTS = [
|
||||
('/image/save-big-image', 'POST', range(0, 10_000, 1_000), open('big-image.png', 'rb').read()),
|
||||
(f'/image/blur?radius={BLUR_RADIUS}', 'POST', range(0, 1_000, 50), open('small-image.png', 'rb').read()),
|
||||
('/status/ok', 'GET', range(0, 30_000, 5000), None),
|
||||
('/image/load-small-image', 'GET', range(0, 30_000, 5000), None),
|
||||
('/static/small-image.png', 'GET', range(0, 30_000, 5000), None),
|
||||
('/image/load-big-image', 'GET', range(0, 500, 50), None),
|
||||
('/static/big-image.png', 'GET', range(0, 500, 50), None),
|
||||
('/static/video.mp4', 'GET', range(0, 10_000, 1_000), None),
|
||||
]
|
||||
|
||||
def send_request(url, method = 'GET', payload = None):
|
||||
success = False
|
||||
@@ -77,46 +61,57 @@ def run_tests(endpoint, method, num_requests, metadata):
|
||||
for num_request in num_requests:
|
||||
if num_request <= 0: continue
|
||||
|
||||
ok_responses = 0
|
||||
bad_responses = 0
|
||||
server_errors = 0
|
||||
cpu, ram = 0, 0
|
||||
total_cpu, total_ram = 0, 0
|
||||
total_time = 0
|
||||
|
||||
with concurrent.futures.ThreadPoolExecutor(max_workers=THREADS) as executor:
|
||||
url = f'{URL_BASE}{endpoint}'
|
||||
for run in range(AVG_RUNS):
|
||||
ok_responses = 0
|
||||
bad_responses = 0
|
||||
server_errors = 0
|
||||
|
||||
start_time = time.time()
|
||||
with concurrent.futures.ThreadPoolExecutor(max_workers=THREADS) as executor:
|
||||
url = f'{URL_BASE}{endpoint}'
|
||||
|
||||
futures = []
|
||||
#with requests.Session() as session:
|
||||
# futures = [executor.submit(send_request, session, url) for _ in range(num_request)]
|
||||
start_time = time.time()
|
||||
|
||||
half = floor(num_request/2)
|
||||
for i in range(num_request):
|
||||
futures.append(executor.submit(send_request, url, method, metadata))
|
||||
futures = []
|
||||
#with requests.Session() as session:
|
||||
# futures = [executor.submit(send_request, session, url) for _ in range(num_request)]
|
||||
|
||||
if i == half:
|
||||
cpu, ram = get_resource_usage()
|
||||
half = floor(num_request/2)
|
||||
for i in range(num_request):
|
||||
futures.append(executor.submit(send_request, url, method, metadata))
|
||||
|
||||
concurrent.futures.wait(futures)
|
||||
if i == half:
|
||||
cpu, ram = get_resource_usage()
|
||||
total_cpu += float(cpu)
|
||||
total_ram += float(ram)
|
||||
|
||||
elapsed_time = time.time() - start_time
|
||||
concurrent.futures.wait(futures)
|
||||
|
||||
for future in futures:
|
||||
responses = future.result()
|
||||
ok_responses += responses[2]
|
||||
bad_responses += responses[4]
|
||||
server_errors += responses[5]
|
||||
elapsed_time = time.time() - start_time
|
||||
total_time += elapsed_time
|
||||
|
||||
for future in futures:
|
||||
responses = future.result()
|
||||
ok_responses += responses[2]
|
||||
bad_responses += responses[4]
|
||||
server_errors += responses[5]
|
||||
|
||||
print(f"[#{run}] {num_request}: {elapsed_time:.2f} seconds. {elapsed_time/num_request:.4f} seconds per request. {num_request/elapsed_time:.2f} requests per second. [OK: {ok_responses}, Bad Request: {bad_responses}, Server Error: {server_errors}]]")
|
||||
|
||||
client = docker.from_env()
|
||||
client.containers.get(CONTAINER_NAME).restart()
|
||||
|
||||
time.sleep(3)
|
||||
|
||||
cpu = total_cpu / AVG_RUNS
|
||||
ram = total_ram / AVG_RUNS
|
||||
elapsed_time = total_time / AVG_RUNS
|
||||
|
||||
print(f"{num_request}: {elapsed_time:.2f} seconds. {elapsed_time/num_request:.4f} seconds per request. {num_request/elapsed_time:.2f} requests per second. [OK: {ok_responses}, Bad Request: {bad_responses}, Server Error: {server_errors}]]")
|
||||
record(files[0], num_request, f"{num_request/elapsed_time:.2f}")
|
||||
record_resource(files[1], num_request, cpu, ram)
|
||||
|
||||
generate_req_graph(files[0], FRAMEWORK_NAME, endpoint)
|
||||
generate_resource_graph(files[1], FRAMEWORK_NAME, endpoint)
|
||||
|
||||
time.sleep(3)
|
||||
|
||||
def get_resource_usage():
|
||||
if CONTAINER_NAME == "": return 0, 0
|
||||
|
||||
@@ -147,9 +142,12 @@ def get_ram_usage(stats):
|
||||
if __name__ == "__main__":
|
||||
if not os.path.exists("data"):
|
||||
os.mkdir("data")
|
||||
else:
|
||||
os.system("rm -rf data/*")
|
||||
|
||||
for endpoint, method, num_requests, metadata in API_REQUESTS:
|
||||
print(f"# {endpoint}")
|
||||
run_tests(endpoint, method, num_requests, metadata)
|
||||
for i in range(len(FRAMEWORKS)):
|
||||
FRAMEWORK_NAME = FRAMEWORKS[i][0]
|
||||
CONTAINER_NAME = FRAMEWORKS[i][1]
|
||||
URL_BASE = ENDPOINTS[FRAMEWORK_NAME]
|
||||
|
||||
for endpoint, method, num_requests, metadata in API_REQUESTS:
|
||||
print(f"# {FRAMEWORK_NAME} - {endpoint}")
|
||||
run_tests(endpoint, method, num_requests, metadata)
|
||||
|
Reference in New Issue
Block a user