mirror of
https://github.com/kata-containers/kata-containers.git
synced 2025-08-14 14:14:15 +00:00
metrics: Adds blogbench and webtool metrics tests
This PR adds blogbench and webtooling metrics checks to this repo. The function running the test intentionally returns zero, so the test will be enabled in another PR once the workflow is green. Fixes: #7069 Signed-off-by: David Esparza <david.esparza.borquez@intel.com>
This commit is contained in:
parent
a25d5b9807
commit
35d096b607
2
.github/workflows/run-metrics.yaml
vendored
2
.github/workflows/run-metrics.yaml
vendored
@ -40,3 +40,5 @@ jobs:
|
||||
- name: run memory usage inside container test
|
||||
run: bash tests/metrics/gha-run.sh run-test-memory-usage-inside-container
|
||||
|
||||
- name: run blogbench test
|
||||
run: bash tests/metrics/gha-run.sh run-test-blogbench
|
||||
|
@ -131,6 +131,15 @@ function run_test_memory_usage_inside_container() {
|
||||
bash tests/metrics/density/memory_usage_inside_container.sh 5
|
||||
}
|
||||
|
||||
function run_test_blogbench() {
|
||||
info "Running Blogbench test using ${KATA_HYPERVISOR} hypervisor"
|
||||
|
||||
# ToDo: remove the exit once the metrics workflow is stable
|
||||
exit 0
|
||||
create_symbolic_links
|
||||
bash tests/metrics/storage/blogbench.sh
|
||||
}
|
||||
|
||||
function main() {
|
||||
action="${1:-}"
|
||||
case "${action}" in
|
||||
@ -138,6 +147,7 @@ function main() {
|
||||
run-test-launchtimes) run_test_launchtimes ;;
|
||||
run-test-memory-usage) run_test_memory_usage ;;
|
||||
run-test-memory-usage-inside-container) run_test_memory_usage_inside_container ;;
|
||||
run-test-blogbench) run_test_blogbench ;;
|
||||
*) >&2 die "Invalid argument" ;;
|
||||
esac
|
||||
}
|
||||
|
124
tests/metrics/storage/blogbench.sh
Executable file
124
tests/metrics/storage/blogbench.sh
Executable file
@ -0,0 +1,124 @@
|
||||
#!/bin/bash
|
||||
#
|
||||
# Copyright (c) 2018-2023 Intel Corporation
|
||||
#
|
||||
# SPDX-License-Identifier: Apache-2.0
|
||||
|
||||
# Description of the test:
|
||||
# This test runs the 'blogbench', and extracts the 'scores' for reads
|
||||
# and writes
|
||||
# Note - the scores are *not* normalised for the number of iterations run,
|
||||
# they are total scores for all iterations (this is the blogbench default output)
|
||||
|
||||
set -e
|
||||
|
||||
# General env
|
||||
SCRIPT_PATH=$(dirname "$(readlink -f "$0")")
|
||||
source "${SCRIPT_PATH}/../lib/common.bash"
|
||||
|
||||
TEST_NAME="blogbench"
|
||||
IMAGE="docker.io/library/local-blogbench:latest"
|
||||
DOCKERFILE="${SCRIPT_PATH}/blogbench_dockerfile/Dockerfile"
|
||||
|
||||
# Number of iterations for blogbench to run - note, results are not
|
||||
# scaled to iterations - more iterations results in bigger results
|
||||
ITERATIONS="${ITERATIONS:-30}"
|
||||
|
||||
# Directory to run the test on
|
||||
# This is run inside of the container
|
||||
TESTDIR="${TESTDIR:-/tmp}"
|
||||
CMD="blogbench -i ${ITERATIONS} -d ${TESTDIR}"
|
||||
|
||||
function main() {
|
||||
# Check tools/commands dependencies
|
||||
cmds=("awk" "docker")
|
||||
|
||||
init_env
|
||||
check_cmds "${cmds[@]}"
|
||||
check_ctr_images "${IMAGE}" "${DOCKERFILE}"
|
||||
metrics_json_init
|
||||
|
||||
local output=$(sudo -E ${CTR_EXE} run --rm --runtime=${CTR_RUNTIME} ${IMAGE} test ${CMD})
|
||||
|
||||
# Save configuration
|
||||
metrics_json_start_array
|
||||
|
||||
local frequency=$(echo "${output}" | grep "Frequency" | cut -d "=" -f2 | cut -d ' ' -f2)
|
||||
local iterations=$(echo "${output}" | grep -w "iterations" | cut -d ' ' -f3)
|
||||
local spawing_writers=$(echo "${output}" | grep -w "writers" | cut -d ' ' -f2)
|
||||
local spawing_rewriters=$(echo "${output}" | grep -w "rewriters" | cut -d ' ' -f2)
|
||||
local spawing_commenters=$(echo "${output}" | grep -w "commenters" | cut -d ' ' -f2)
|
||||
local spawing_readers=$(echo "${output}" | grep -w "readers" | cut -d ' ' -f2)
|
||||
|
||||
local json="$(cat << EOF
|
||||
{
|
||||
"Frequency" : ${frequency},
|
||||
"Iterations" : ${iterations},
|
||||
"Number of spawing writers" : ${spawing_writers},
|
||||
"Number of spawing rewriters" : ${spawing_rewriters},
|
||||
"Number of spawing commenters" : ${spawing_commenters},
|
||||
"Number of spawing readers" : ${spawing_readers}
|
||||
}
|
||||
EOF
|
||||
)"
|
||||
metrics_json_add_array_element "${json}"
|
||||
metrics_json_end_array "Config"
|
||||
|
||||
# Save results
|
||||
metrics_json_start_array
|
||||
|
||||
local writes=$(tail -2 <<< "${output}" | head -1 | awk '{print $5}')
|
||||
local reads=$(tail -1 <<< "${output}" | awk '{print $6}')
|
||||
|
||||
# Obtaining other Blogbench results
|
||||
local -r data=$(echo "${output}" | tail -n +12 | head -n -3)
|
||||
local nb_blogs=$(echo "${data}" | awk ' BEGIN {ORS="\t"} {print $1} ' | tr '\t' ',' | sed '$ s/.$//')
|
||||
local r_articles=$(echo "${data}" | awk ' BEGIN {ORS="\t"} {print $2} ' | tr '\t' ',' | sed '$ s/.$//')
|
||||
local w_articles=$(echo "${data}" | awk ' BEGIN {ORS="\t"} {print $3} ' | tr '\t' ',' | sed '$ s/.$//')
|
||||
local r_pictures=$(echo "${data}" | awk ' BEGIN {ORS="\t"} {print $4} ' | tr '\t' ',' | sed '$ s/.$//')
|
||||
local w_pictures=$(echo "${data}" | awk ' BEGIN {ORS="\t"} {print $5} ' | tr '\t' ',' | sed '$ s/.$//')
|
||||
local r_comments=$(echo "${data}" | awk ' BEGIN {ORS="\t"} {print $6} ' | tr '\t' ',' | sed '$ s/.$//')
|
||||
local w_comments=$(echo "${data}" | awk ' BEGIN {ORS="\t"} {print $7} ' | tr '\t' ',' | sed '$ s/.$//')
|
||||
|
||||
local json="$(cat << EOF
|
||||
{
|
||||
"write": {
|
||||
"Result" : "${writes}",
|
||||
"Units" : "items"
|
||||
},
|
||||
"read": {
|
||||
"Result" : "${reads}",
|
||||
"Units" : "items"
|
||||
},
|
||||
"Nb blogs": {
|
||||
"Result" : "${nb_blogs}"
|
||||
},
|
||||
"R articles": {
|
||||
"Result" : "${r_articles}"
|
||||
},
|
||||
"W articles": {
|
||||
"Result" : "${w_articles}"
|
||||
},
|
||||
"R pictures": {
|
||||
"Result" : "${r_pictures}"
|
||||
},
|
||||
"W pictures": {
|
||||
"Result" : "${w_pictures}"
|
||||
},
|
||||
"R comments": {
|
||||
"Result" : "${r_comments}"
|
||||
},
|
||||
"W comments": {
|
||||
"Result" : "${w_comments}"
|
||||
}
|
||||
}
|
||||
EOF
|
||||
)"
|
||||
|
||||
metrics_json_add_array_element "${json}"
|
||||
metrics_json_end_array "Results"
|
||||
metrics_json_save
|
||||
clean_env_ctr
|
||||
}
|
||||
|
||||
main "$@"
|
32
tests/metrics/storage/blogbench_dockerfile/Dockerfile
Normal file
32
tests/metrics/storage/blogbench_dockerfile/Dockerfile
Normal file
@ -0,0 +1,32 @@
|
||||
# Copyright (c) 2018-2023 Intel Corporation
|
||||
#
|
||||
# SPDX-License-Identifier: Apache-2.0
|
||||
|
||||
# Set up an Ubuntu image with 'blogbench' installed
|
||||
|
||||
# Usage: FROM [image name]
|
||||
# hadolint ignore=DL3007
|
||||
FROM docker.io/library/ubuntu:latest
|
||||
|
||||
# Version of the Dockerfile
|
||||
LABEL DOCKERFILE_VERSION="1.0"
|
||||
|
||||
# URL for blogbench test and blogbench version
|
||||
ENV BLOGBENCH_URL "https://download.pureftpd.org/pub/blogbench"
|
||||
ENV BLOGBENCH_VERSION 1.1
|
||||
|
||||
RUN apt-get update && \
|
||||
apt-get install -y --no-install-recommends build-essential curl && \
|
||||
apt-get remove -y unattended-upgrades && \
|
||||
apt-get clean && \
|
||||
rm -rf /var/lib/apt/lists/ && \
|
||||
curl -OkL "${BLOGBENCH_URL}/blogbench-${BLOGBENCH_VERSION}.tar.gz" && \
|
||||
tar xzf "blogbench-${BLOGBENCH_VERSION}.tar.gz" -C /
|
||||
WORKDIR "/blogbench-${BLOGBENCH_VERSION}"
|
||||
RUN arch="$(uname -m)" && \
|
||||
export arch && \
|
||||
./configure --build="${arch}" && \
|
||||
make && \
|
||||
make install-strip
|
||||
|
||||
CMD ["/bin/bash"]
|
28
tests/metrics/storage/web-tooling-dockerfile/Dockerfile.in
Executable file
28
tests/metrics/storage/web-tooling-dockerfile/Dockerfile.in
Executable file
@ -0,0 +1,28 @@
|
||||
# Copyright (c) 2020-2021 Intel Corporation
|
||||
#
|
||||
# SPDX-License-Identifier: Apache-2.0
|
||||
|
||||
# Set up an Ubuntu image with 'web tooling' installed
|
||||
|
||||
# Usage: FROM [image name]
|
||||
# hadolint ignore=DL3007
|
||||
FROM @UBUNTU_REGISTRY@/ubuntu:latest
|
||||
|
||||
# Version of the Dockerfile
|
||||
LABEL DOCKERFILE_VERSION="1.0"
|
||||
|
||||
# URL for web tooling test
|
||||
ENV WEB_TOOLING_URL "https://github.com/v8/web-tooling-benchmark"
|
||||
ENV NODEJS_VERSION "setup_14.x"
|
||||
|
||||
RUN apt-get update && \
|
||||
apt-get install -y --no-install-recommends build-essential git curl sudo && \
|
||||
apt-get remove -y unattended-upgrades && \
|
||||
curl -OkL https://deb.nodesource.com/${NODEJS_VERSION} && chmod +x ${NODEJS_VERSION} && ./${NODEJS_VERSION} && \
|
||||
apt-get install -y --no-install-recommends nodejs && \
|
||||
apt-get clean && rm -rf /var/lib/apt/lists && \
|
||||
git clone ${WEB_TOOLING_URL} /web-tooling-benchmark
|
||||
WORKDIR /web-tooling-benchmark/
|
||||
RUN npm install --unsafe-perm
|
||||
|
||||
CMD ["/bin/bash"]
|
283
tests/metrics/storage/webtooling.sh
Executable file
283
tests/metrics/storage/webtooling.sh
Executable file
@ -0,0 +1,283 @@
|
||||
#!/bin/bash
|
||||
#
|
||||
# Copyright (c) 2020-2023 Intel Corporation
|
||||
#
|
||||
# SPDX-License-Identifier: Apache-2.0
|
||||
|
||||
# Description of the test:
|
||||
# This test runs the 'web tooling benchmark'
|
||||
# https://github.com/v8/web-tooling-benchmark
|
||||
|
||||
set -o pipefail
|
||||
|
||||
# General env
|
||||
SCRIPT_PATH=$(dirname "$(readlink -f "$0")")
|
||||
source "${SCRIPT_PATH}/../lib/common.bash"
|
||||
NUM_CONTAINERS="${1:-}"
|
||||
|
||||
IMAGE="docker.io/library/local-web-tooling:latest"
|
||||
DOCKERFILE="${SCRIPT_PATH}/web-tooling-dockerfile/Dockerfile"
|
||||
|
||||
# Directory to run the test inside of the container
|
||||
TESTDIR="${TESTDIR:-/testdir}"
|
||||
file_path="/web-tooling-benchmark"
|
||||
file_name="output"
|
||||
|
||||
# Directory where the webtooling results are stored
|
||||
TMP_DIR=$(mktemp --tmpdir -d webtool.XXXXXXXXXX)
|
||||
|
||||
# Options to control the start of the workload using a trigger-file
|
||||
dst_dir="/host"
|
||||
src_dir=$(mktemp --tmpdir -d webtool.XXXXXXXXXX)
|
||||
trigger_file="$RANDOM.txt"
|
||||
guest_trigger_file="$dst_dir/$trigger_file"
|
||||
host_trigger_file="$src_dir/$trigger_file"
|
||||
start_script="webtooling_start.sh"
|
||||
|
||||
# CMD points to the script that starts the workload
|
||||
CMD="$dst_dir/$start_script"
|
||||
MOUNT_OPTIONS="type=bind,src=$src_dir,dst=$dst_dir,options=rbind:ro"
|
||||
PAYLOAD_ARGS="${PAYLOAD_ARGS:-tail -f /dev/null}"
|
||||
|
||||
# This timeout is related with the amount of time that
|
||||
# webtool benchmark needs to run inside the container
|
||||
timeout=600
|
||||
INITIAL_NUM_PIDS=1
|
||||
|
||||
cpu_period="100000"
|
||||
cpu_quota="200000"
|
||||
|
||||
function remove_tmp_dir() {
|
||||
rm -rf "$TMP_DIR"
|
||||
rm -rf "$src_dir"
|
||||
}
|
||||
|
||||
trap remove_tmp_dir EXIT
|
||||
|
||||
# Show help about this script
|
||||
function help(){
|
||||
cat << EOF
|
||||
Usage: $0 <count>
|
||||
Description:
|
||||
<count> : Number of containers to run.
|
||||
EOF
|
||||
}
|
||||
|
||||
# script used to launch the workload
|
||||
function create_start_script() {
|
||||
local script="${src_dir/$start_script}"
|
||||
rm -rf "${script}"
|
||||
|
||||
cat <<EOF >>"${script}"
|
||||
#!/bin/bash
|
||||
mkdir -p "${TESTDIR}"
|
||||
|
||||
until [ -f ${guest_trigger_file} ]; do
|
||||
sleep 1
|
||||
done
|
||||
pushd "${file_path}"
|
||||
node dist/cli.js > "${file_name}"
|
||||
EOF
|
||||
chmod +x "${script}"
|
||||
}
|
||||
|
||||
function verify_task_is_completed_on_all_containers() {
|
||||
local containers=( $(sudo -E "${CTR_EXE}" c list -q) )
|
||||
local sleep_secs=10
|
||||
local max=$(bc <<<"${timeout} / ${sleep_secs}")
|
||||
local wip_list=()
|
||||
local count=1
|
||||
local sum=0
|
||||
local i=""
|
||||
|
||||
while (( ${sum} < ${NUM_CONTAINERS} )); do
|
||||
|
||||
for i in "${containers[@]}"; do
|
||||
# Only check containers that have not completed the workload at this step
|
||||
num_pids=$(sudo -E "${CTR_EXE}" t metrics "${i}" | grep pids.current | grep pids.current | xargs | cut -d ' ' -f 2)
|
||||
|
||||
if [ "${num_pids}" -lt "${INITIAL_NUM_PIDS}" ]; then
|
||||
((sum++))
|
||||
else
|
||||
wip_list+=("${i}")
|
||||
fi
|
||||
done
|
||||
|
||||
# hold the list of containers that are still running the workload
|
||||
containers=(${wip_list[*]})
|
||||
wip_list=()
|
||||
|
||||
info "loop ${count} of ${max}: sleeping for ${sleep_secs} seconds"
|
||||
sleep "${sleep_secs}"
|
||||
((count++))
|
||||
done
|
||||
}
|
||||
|
||||
function check_containers_are_up() {
|
||||
info "Verify that the containers are running"
|
||||
local containers_launched=0
|
||||
while (( $containers_launched < ${NUM_CONTAINERS} )); do
|
||||
containers_launched="$(sudo -E ${CTR_EXE} t list | grep -c "RUNNING")"
|
||||
sleep 1
|
||||
done
|
||||
}
|
||||
|
||||
function save_config() {
|
||||
metrics_json_start_array
|
||||
|
||||
local json="$(cat << EOF
|
||||
{
|
||||
"containers": "${NUM_CONTAINERS}",
|
||||
"image": "${IMAGE}",
|
||||
"units": "runs/s"
|
||||
}
|
||||
EOF
|
||||
)"
|
||||
metrics_json_add_array_element "${json}"
|
||||
metrics_json_end_array "Config"
|
||||
}
|
||||
|
||||
function main() {
|
||||
# Verify enough arguments
|
||||
if [ $# != 1 ]; then
|
||||
echo >&2 "error: Not enough arguments [$@]"
|
||||
help
|
||||
exit 1
|
||||
fi
|
||||
|
||||
local i=0
|
||||
local containers=()
|
||||
local cmds=("docker")
|
||||
local not_started_count=$NUM_CONTAINERS
|
||||
|
||||
restart_containerd_service
|
||||
# Check tools/commands dependencies
|
||||
init_env
|
||||
check_cmds "${cmds[@]}"
|
||||
check_ctr_images "$IMAGE" "$DOCKERFILE"
|
||||
metrics_json_init
|
||||
save_config
|
||||
create_start_script
|
||||
rm -rf "${host_trigger_file}"
|
||||
|
||||
info "Creating ${NUM_CONTAINERS} containers"
|
||||
|
||||
for ((i=1; i<= "${NUM_CONTAINERS}"; i++)); do
|
||||
containers+=($(random_name))
|
||||
# Web tool benchmark needs 2 cpus to run completely in its cpu utilization
|
||||
sudo -E "${CTR_EXE}" run -d --runtime "${CTR_RUNTIME}" --cpu-quota "${cpu_quota}" --cpu-period "${cpu_period}" --mount="${MOUNT_OPTIONS}" "${IMAGE}" "${containers[-1]}" sh -c "${PAYLOAD_ARGS}"
|
||||
((not_started_count--))
|
||||
info "${not_started_count} remaining containers"
|
||||
done
|
||||
|
||||
# Check that the requested number of containers are running
|
||||
local timeout_launch="10"
|
||||
check_containers_are_up & pid=$!
|
||||
(sleep "${timeout_launch}" && kill -HUP ${pid}) 2>/dev/null & pid_tout=$!
|
||||
|
||||
if wait $pid 2>/dev/null; then
|
||||
pkill -HUP -P "${pid_tout}"
|
||||
wait "${pid_tout}"
|
||||
else
|
||||
warn "Time out exceeded"
|
||||
return 1
|
||||
fi
|
||||
|
||||
# Get the initial number of pids in a single container before the workload starts
|
||||
INITIAL_NUM_PIDS=$(sudo -E "${CTR_EXE}" t metrics "${containers[-1]}" | grep pids.current | grep pids.current | xargs | cut -d ' ' -f 2)
|
||||
((INITIAL_NUM_PIDS++))
|
||||
|
||||
# Launch webtooling benchmark
|
||||
local pids=()
|
||||
local j=0
|
||||
for i in "${containers[@]}"; do
|
||||
$(sudo -E "${CTR_EXE}" t exec -d --exec-id "$(random_name)" "${i}" sh -c "${CMD}") &
|
||||
pids[${j}]=$!
|
||||
((j++))
|
||||
done
|
||||
|
||||
# wait for all pids
|
||||
for pid in ${pids[*]}; do
|
||||
wait "${pid}"
|
||||
done
|
||||
|
||||
touch "${host_trigger_file}"
|
||||
info "All containers are running the workload..."
|
||||
|
||||
# Verify that all containers have completed the assigned task
|
||||
verify_task_is_completed_on_all_containers & pid=$!
|
||||
(sleep "$timeout" && kill -HUP $pid) 2>/dev/null & pid_tout=$!
|
||||
if wait ${pid} 2>/dev/null; then
|
||||
pkill -HUP -P "${pid_tout}"
|
||||
wait "${pid_tout}"
|
||||
else
|
||||
warn "Time out exceeded"
|
||||
return 1
|
||||
fi
|
||||
|
||||
RESULTS_CMD="cat ${file_path}/${file_name}"
|
||||
for i in "${containers[@]}"; do
|
||||
sudo -E "${CTR_EXE}" t exec --exec-id "${RANDOM}" "${i}" sh -c "${RESULTS_CMD}" >> "${TMP_DIR}/results"
|
||||
done
|
||||
|
||||
# Save configuration
|
||||
metrics_json_start_array
|
||||
|
||||
local output=$(cat "${TMP_DIR}/results")
|
||||
local cut_results="cut -d':' -f2 | sed -e 's/^[ \t]*//'| cut -d ' ' -f1 | tr '\n' ',' | sed 's/.$//'"
|
||||
|
||||
local acorn=$(echo "${output}" | grep -w "acorn" | eval "${cut_results}")
|
||||
local babel=$(echo "${output}" | grep -w "babel" | sed '/babel-minify/d' | eval "${cut_results}")
|
||||
local babel_minify=$(echo "${output}" | grep -w "babel-minify" | eval "${cut_results}")
|
||||
local babylon=$(echo "${output}" | grep -w "babylon" | eval "${cut_results}")
|
||||
local buble=$(echo "${output}" | grep -w "buble" | eval "${cut_results}")
|
||||
local chai=$(echo "${output}" | grep -w "chai" | eval "${cut_results}")
|
||||
local coffeescript=$(echo "${output}" | grep -w "coffeescript" | eval "${cut_results}")
|
||||
local espree=$(echo "${output}" | grep -w "espree" | eval "${cut_results}")
|
||||
local esprima=$(echo "${output}" | grep -w "esprima" | eval "${cut_results}")
|
||||
local jshint=$(echo "${output}" | grep -w "jshint" | eval "${cut_results}")
|
||||
local lebab=$(echo "${output}" | grep -w "lebab" | eval "${cut_results}")
|
||||
local postcss=$(echo "${output}" | grep -w "postcss" | eval "${cut_results}")
|
||||
local prepack=$(echo "${output}" | grep -w "prepack" | eval "${cut_results}")
|
||||
local prettier=$(echo "${output}" | grep -w "prettier" | eval "${cut_results}")
|
||||
local source_map=$(echo "${output}" | grep -w "source-map" | eval "${cut_results}")
|
||||
local terser=$(echo "${output}" | grep -w "terser" | eval "${cut_results}")
|
||||
local typescript=$(echo "${output}" | grep -w "typescript" | eval "${cut_results}")
|
||||
local uglify_js=$(echo "${output}" | grep -w "uglify-js" | eval "${cut_results}")
|
||||
local geometric_mean=$(echo "${output}" | grep -w "Geometric" | eval "${cut_results}")
|
||||
local average_tps=$(echo "${geometric_mean}" | sed "s/,/+/g;s/.*/(&)\/$NUM_CONTAINERS/g" | bc -l)
|
||||
local tps=$(echo "${average_tps}*${NUM_CONTAINERS}" | bc -l)
|
||||
|
||||
local json="$(cat << EOF
|
||||
{
|
||||
"Acorn" : "${acorn}",
|
||||
"Babel" : "${babel}",
|
||||
"Babel minify" : "${babel_minify}",
|
||||
"Babylon" : "${babylon}",
|
||||
"Buble" : "${buble}",
|
||||
"Chai" : "${chai}",
|
||||
"Coffeescript" : "${coffeescript}",
|
||||
"Espree" : "${espree}",
|
||||
"Esprima" : "${esprima}",
|
||||
"Jshint" : "${jshint}",
|
||||
"Lebab" : "${lebab}",
|
||||
"Postcss" : "${postcss}",
|
||||
"Prepack" : "${prepack}",
|
||||
"Prettier" : "${prettier}",
|
||||
"Source map" : "${source_map}",
|
||||
"Terser" : "${terser}",
|
||||
"Typescript" : "${typescript}",
|
||||
"Uglify js" : "${uglify_js}",
|
||||
"Geometric mean" : "${geometric_mean}",
|
||||
"Average TPS" : "${average_tps}",
|
||||
"TPS" : "${tps}"
|
||||
}
|
||||
EOF
|
||||
)"
|
||||
metrics_json_add_array_element "${json}"
|
||||
metrics_json_end_array "Results"
|
||||
metrics_json_save
|
||||
clean_env_ctr
|
||||
}
|
||||
|
||||
main "$@"
|
Loading…
Reference in New Issue
Block a user