forked from haiwan/sledge
parent
abbce7df68
commit
197e1c5a5a
@ -0,0 +1 @@
|
||||
res
|
After Width: | Height: | Size: 1.3 KiB |
File diff suppressed because one or more lines are too long
@ -0,0 +1,6 @@
|
||||
#!/bin/bash
|
||||
|
||||
for i in {1..10000}; do
|
||||
echo "$i"
|
||||
curl -H 'Expect:' -H "Content-Type: text/plain" --data-binary "@handwrt1.pnm" localhost:10000
|
||||
done
|
@ -0,0 +1,19 @@
|
||||
#!/bin/bash
|
||||
# Executes the runtime in GDB
|
||||
# Substitutes the absolute path from the container with a path relatively derived from the location of this script
|
||||
# This allows debugging outside of the Docker container
|
||||
# Also disables pagination and stopping on SIGUSR1
|
||||
|
||||
experiment_directory=$(pwd)
|
||||
project_directory=$(cd ../.. && pwd)
|
||||
binary_directory=$(cd "$project_directory"/bin && pwd)
|
||||
|
||||
export LD_LIBRARY_PATH="$binary_directory:$LD_LIBRARY_PATH"
|
||||
export PATH="$binary_directory:$PATH"
|
||||
|
||||
gdb --eval-command="handle SIGUSR1 nostop" \
|
||||
--eval-command="handle SIGPIPE nostop" \
|
||||
--eval-command="set pagination off" \
|
||||
--eval-command="set substitute-path /sledge/runtime $project_directory" \
|
||||
--eval-command="run $experiment_directory/spec.json" \
|
||||
sledgert
|
After Width: | Height: | Size: 42 KiB |
File diff suppressed because one or more lines are too long
After Width: | Height: | Size: 823 KiB |
File diff suppressed because one or more lines are too long
@ -0,0 +1,20 @@
|
||||
reset
|
||||
|
||||
set term jpeg
|
||||
set output "latency.jpg"
|
||||
|
||||
set xlabel "Connections"
|
||||
set xrange [-5:105]
|
||||
|
||||
set ylabel "Latency (ms)"
|
||||
set yrange [0:]
|
||||
|
||||
set key left top
|
||||
|
||||
|
||||
set style histogram columnstacked
|
||||
|
||||
plot 'latency.dat' using 1:2 title 'p50', \
|
||||
'latency.dat' using 1:3 title 'p90', \
|
||||
'latency.dat' using 1:4 title 'p99', \
|
||||
'latency.dat' using 1:5 title 'p100', \
|
@ -0,0 +1,139 @@
|
||||
#!/bin/bash
|
||||
|
||||
# This experiment is intended to document how the level of concurrent requests influence the latency, throughput, and success/failure rate
|
||||
# Use -d flag if running under gdb
|
||||
|
||||
timestamp=$(date +%s)
|
||||
experiment_directory=$(pwd)
|
||||
binary_directory=$(cd ../../bin && pwd)
|
||||
results_directory="$experiment_directory/res/$timestamp"
|
||||
log=log.txt
|
||||
|
||||
mkdir -p "$results_directory"
|
||||
|
||||
{
|
||||
echo "*******"
|
||||
echo "* Git *"
|
||||
echo "*******"
|
||||
git log | head -n 1 | cut -d' ' -f2
|
||||
git status
|
||||
echo ""
|
||||
|
||||
echo "************"
|
||||
echo "* Makefile *"
|
||||
echo "************"
|
||||
cat ../../Makefile
|
||||
echo ""
|
||||
|
||||
echo "**********"
|
||||
echo "* Run.sh *"
|
||||
echo "**********"
|
||||
cat run.sh
|
||||
echo ""
|
||||
|
||||
echo "************"
|
||||
echo "* Hardware *"
|
||||
echo "************"
|
||||
lscpu
|
||||
echo ""
|
||||
|
||||
echo "*************"
|
||||
echo "* Execution *"
|
||||
echo "*************"
|
||||
} >>"$results_directory/$log"
|
||||
|
||||
# Start the runtime
|
||||
if [ "$1" != "-d" ]; then
|
||||
PATH="$binary_directory:$PATH" LD_LIBRARY_PATH="$binary_directory:$LD_LIBRARY_PATH" sledgert "$experiment_directory/spec.json" >>"$results_directory/$log" 2>>"$results_directory/$log" &
|
||||
sleep 1
|
||||
else
|
||||
echo "Running under gdb"
|
||||
echo "Running under gdb" >>"$results_directory/$log"
|
||||
fi
|
||||
|
||||
# Execute workloads long enough for runtime to learn excepted execution time
|
||||
echo -n "Running Samples: "
|
||||
hey -n 10000 -c 3 -q 200 -o csv -m GET http://localhost:10000
|
||||
sleep 5
|
||||
echo "[DONE]"
|
||||
|
||||
# Execute the experiments
|
||||
concurrency=(1 20 40 60 80 100)
|
||||
echo "Running Experiments"
|
||||
for conn in ${concurrency[*]}; do
|
||||
printf "\t%d Concurrency: " "$conn"
|
||||
hey -n 10000 -c "$conn" -cpus 2 -o csv -m GET http://localhost:10000 >"$results_directory/con$conn.csv"
|
||||
echo "[DONE]"
|
||||
done
|
||||
|
||||
# Stop the runtime
|
||||
|
||||
if [ "$1" != "-d" ]; then
|
||||
sleep 5
|
||||
echo -n "Running Cleanup: "
|
||||
pkill sledgert >/dev/null 2>/dev/null
|
||||
pkill wrk >/dev/null 2>/dev/null
|
||||
echo "[DONE]"
|
||||
fi
|
||||
|
||||
# Generate *.csv and *.dat results
|
||||
echo -n "Parsing Results: "
|
||||
|
||||
printf "Concurrency,Success_Rate\n" >>"$results_directory/success.csv"
|
||||
printf "Concurrency,Throughput\n" >>"$results_directory/throughput.csv"
|
||||
printf "Con,p50,p90,p99,p100\n" >>"$results_directory/latency.csv"
|
||||
|
||||
for conn in ${concurrency[*]}; do
|
||||
# Calculate Success Rate for csv
|
||||
awk -F, '
|
||||
$7 == 200 {ok++}
|
||||
END{printf "'"$conn"',%3.5f\n", (ok / (NR - 1) * 100)}
|
||||
' <"$results_directory/con$conn.csv" >>"$results_directory/success.csv"
|
||||
|
||||
# Filter on 200s, convery from s to ms, and sort
|
||||
awk -F, '$7 == 200 {print ($1 * 1000)}' <"$results_directory/con$conn.csv" |
|
||||
sort -g >"$results_directory/con$conn-response.csv"
|
||||
|
||||
# Get Number of 200s
|
||||
oks=$(wc -l <"$results_directory/con$conn-response.csv")
|
||||
|
||||
# Get Latest Timestamp
|
||||
duration=$(tail -n1 "$results_directory/con$conn.csv" | cut -d, -f8)
|
||||
throughput=$(echo "$oks/$duration" | bc)
|
||||
printf "%d,%f\n" "$conn" "$throughput" >>"$results_directory/throughput.csv"
|
||||
|
||||
# Generate Latency Data for csv
|
||||
awk '
|
||||
BEGIN {
|
||||
sum = 0
|
||||
p50 = int('"$oks"' * 0.5)
|
||||
p90 = int('"$oks"' * 0.9)
|
||||
p99 = int('"$oks"' * 0.99)
|
||||
p100 = '"$oks"'
|
||||
printf "'"$conn"',"
|
||||
}
|
||||
NR==p50 {printf "%1.4f,", $0}
|
||||
NR==p90 {printf "%1.4f,", $0}
|
||||
NR==p99 {printf "%1.4f,", $0}
|
||||
NR==p100 {printf "%1.4f\n", $0}
|
||||
' <"$results_directory/con$conn-response.csv" >>"$results_directory/latency.csv"
|
||||
|
||||
# Delete scratch file used for sorting/counting
|
||||
rm -rf "$results_directory/con$conn-response.csv"
|
||||
done
|
||||
|
||||
# Transform csvs to dat files for gnuplot
|
||||
for file in success latency throughput; do
|
||||
echo -n "#" >"$results_directory/$file.dat"
|
||||
tr ',' ' ' <"$results_directory/$file.csv" | column -t >>"$results_directory/$file.dat"
|
||||
done
|
||||
|
||||
# Generate gnuplots
|
||||
cd "$results_directory" || exit
|
||||
gnuplot ../../latency.gnuplot
|
||||
gnuplot ../../success.gnuplot
|
||||
gnuplot ../../throughput.gnuplot
|
||||
cd "$experiment_directory" || exit
|
||||
|
||||
# Cleanup, if requires
|
||||
echo "[DONE]"
|
@ -0,0 +1,14 @@
|
||||
{
|
||||
"active": "yes",
|
||||
"name": "gocr",
|
||||
"path": "gocr.aso",
|
||||
"port": 10000,
|
||||
"relative-deadline-us": 50000000000,
|
||||
"argsize": 1,
|
||||
"http-req-headers": [],
|
||||
"http-req-content-type": "text/plain",
|
||||
"http-req-size": 1024000,
|
||||
"http-resp-headers": [],
|
||||
"http-resp-size": 1024000,
|
||||
"http-resp-content-type": "text/plain"
|
||||
}
|
@ -0,0 +1,12 @@
|
||||
reset
|
||||
|
||||
set term jpeg
|
||||
set output "success.jpg"
|
||||
|
||||
set xlabel "Connections"
|
||||
set xrange [-5:105]
|
||||
|
||||
set ylabel "% 2XX"
|
||||
set yrange [0:110]
|
||||
|
||||
plot 'success.dat' using 1:2 title '2XX'
|
@ -0,0 +1,13 @@
|
||||
reset
|
||||
|
||||
set term jpeg
|
||||
set output "throughput.jpg"
|
||||
|
||||
# TODO: Axis shouldn't be linear
|
||||
set xlabel "Connections"
|
||||
set xrange [-5:105]
|
||||
|
||||
set ylabel "Requests/sec"
|
||||
set yrange [0:]
|
||||
|
||||
plot 'throughput.dat' using 1:2 title 'Reqs/sec'
|
Loading…
Reference in new issue