Files
kata-containers/tests/stability/soak_parallel_rm.sh
Fabiano Fidêncio 47f66299c3 tests: Fix shellcheck issues in soak_parallel_rm.sh
Fix shellcheck warnings and notes identified by running
shellcheck --severity=style.

Signed-off-by: Fabiano Fidêncio <ffidencio@nvidia.com>
2026-04-24 08:14:08 +02:00

219 lines
6.0 KiB
Bash
Executable File

#!/bin/bash
#
# Copyright (c) 2017-2023 Intel Corporation
#
# SPDX-License-Identifier: Apache-2.0
#
# This test will run a number of parallel containers, and then try to
# 'rm -f' them all at the same time. It will check after each run and
# rm that we have the expected number of containers, shims,
# qemus and runtimes active
# The goals are two fold:
# - spot any stuck or non-started components
# - catch any hang ups
cidir=$(dirname "$0")
# shellcheck source=/dev/null
source "${cidir}/common_stability.bash"
# shellcheck source=/dev/null
source "/etc/os-release" || source "/usr/lib/os-release"
set -x
# How many times will we run the test loop...
ITERATIONS="${ITERATIONS:-5}"
# the system 'free available' level where we stop running the tests, as otherwise
# the system can crawl to a halt, and/or start refusing to launch new VMs anyway
# We choose 2G, as that is one of the default VM sizes for Kata
MEM_CUTOFF="${MEM_CUTOFF:-(2*1024*1024*1024)}"
# do we need a command argument for this payload?
COMMAND="${COMMAND:-tail -f /dev/null}"
# Runtime path
# shellcheck disable=SC2154
RUNTIME_PATH=$(command -v "${RUNTIME}")
# The place where virtcontainers keeps its active pod info
# This is ultimately what 'kata-runtime list' uses to get its info, but
# we can also check it for sanity directly
VC_POD_DIR="${VC_POD_DIR:-/run/vc/sbs}"
# let's cap the test. If you want to run until you hit the memory limit
# then just set this to a very large number
MAX_CONTAINERS="${MAX_CONTAINERS:-110}"
KATA_HYPERVISOR="${KATA_HYPERVISOR:-qemu}"
function check_vsock_active() {
vsock_configured=$("${RUNTIME_PATH}" kata-env | awk '/UseVSock/ {print $3}')
vsock_supported=$("${RUNTIME_PATH}" kata-env | awk '/SupportVSock/ {print $3}')
if [[ "${vsock_configured}" == true ]] && [[ "${vsock_supported}" == true ]]; then
return 0
else
return 1
fi
}
function count_containers() {
# shellcheck disable=SC2154
sudo "${CTR_EXE}" c list -q | wc -l
}
function check_all_running() {
local goterror=0
info "Checking ${how_many} containers have all relevant components"
# check what docker thinks
how_many_running=$(count_containers)
if (( how_many_running != how_many )); then
info "Wrong number of containers running (${how_many_running} != ${how_many}) - stopping"
((goterror++))
fi
# Only check for Kata components if we are using a Kata runtime
if (( check_kata_components )); then
# check we have the right number of shims
# shellcheck disable=SC2154
how_many_shims=$(pgrep -a -f "${SHIM_PATH}" | grep -c containerd.sock)
# one shim process per container...
if (( how_many_running != how_many_shims )); then
info "Wrong number of shims running (${how_many_running} != ${how_many_shims}) - stopping"
((goterror++))
fi
# check we have the right number of vm's
if [[ "${KATA_HYPERVISOR}" != "dragonball" ]]; then
# shellcheck disable=SC2154
how_many_vms=$(pgrep -a "$(basename "${HYPERVISOR_PATH}" | cut -d '-' -f1)" | wc -l)
if (( how_many_running != how_many_vms )); then
info "Wrong number of ${KATA_HYPERVISOR} running (${how_many_running} != ${how_many_vms}) - stopping"
((goterror++))
fi
fi
# if this is kata-runtime, check how many pods virtcontainers thinks we have
if [[ "${RUNTIME}" == "containerd-shim-kata-v2" ]]; then
if [[ -d "${VC_POD_DIR}" ]]; then
num_vc_pods=$(sudo ls -1 "${VC_POD_DIR}" | wc -l)
if (( how_many_running != num_vc_pods )); then
info "Wrong number of pods in ${VC_POD_DIR} (${how_many_running} != ${num_vc_pods}) - stopping)"
((goterror++))
fi
fi
fi
fi
if (( goterror != 0 )); then
show_system_ctr_state
die "Got ${goterror} errors, quitting"
fi
}
# reported system 'available' memory
function get_system_avail() {
free -b | head -2 | tail -1 | awk '{print $7}'
}
function go() {
info "Running..."
how_many=0
while true; do {
check_all_running
local i
for ((i=1; i<= MAX_CONTAINERS; i++)); do
local name
name=$(random_name)
containers+=("${name}")
# shellcheck disable=SC2154
sudo "${CTR_EXE}" run --runtime="${CTR_RUNTIME}" -d "${nginx_image}" "${containers[-1]}" sh -c "${COMMAND}"
((how_many++))
done
if (( how_many >= MAX_CONTAINERS )); then
info "And we have hit the max ${how_many} containers"
return
fi
how_much=$(get_system_avail)
if (( how_much < MEM_CUTOFF )); then
info "And we are out of memory on container ${how_many} (${how_much} < ${MEM_CUTOFF})"
return
fi
}
done
}
function count_mounts() {
mount | wc -l
}
function check_mounts() {
final_mount_count=$(count_mounts)
if [[ "${final_mount_count}" -lt "${initial_mount_count}" ]]; then
info "Final mount count does not match initial count (${final_mount_count} != ${initial_mount_count})"
fi
}
function init() {
restart_containerd_service
extract_kata_env
clean_env_ctr
# remember how many mount points we had before we do anything
# and then sanity check we end up with no new ones dangling at the end
initial_mount_count=$(count_mounts)
# Only check Kata items if we are using a Kata runtime
if [[ "${RUNTIME}" == "containerd-shim-kata-v2" ]]; then
info "Checking Kata runtime"
check_kata_components=1
else
info "Not a Kata runtime, not checking for Kata components"
check_kata_components=0
fi
versions_file="${cidir}/../../versions.yaml"
# shellcheck disable=SC2154
nginx_registry=$("${GOPATH}/bin/yq" ".docker_images.nginx.registry" "${versions_file}")
nginx_digest=$("${GOPATH}/bin/yq" ".docker_images.nginx.digest" "${versions_file}")
nginx_image="${nginx_registry}@${nginx_digest}"
# Pull nginx image
if ! sudo "${CTR_EXE}" image pull "${nginx_image}"; then
die "Unable to retry docker image ${nginx_image}"
fi
}
function spin() {
local i
for ((i=1; i<= ITERATIONS; i++)); do {
info "Start iteration ${i} of ${ITERATIONS}"
#spin them up
go
#check we are in a sane state
check_all_running
#shut them all down
clean_env_ctr
#Note there should be none running
how_many=0
#and check they all died
check_all_running
#and that we have no dangling mounts
check_mounts
}
done
}
init
spin