cb90136c33
We will be able to distinguish case and suite in the post process scripts based on if they have any internal tests. Change-Id: Iff2aa5caa251924d8a842085d87eb2a17aca45d4 Signed-off-by: Seth Howell <seth.howell@intel.com> Reviewed-on: https://review.gerrithub.io/c/spdk/spdk/+/478522 Tested-by: SPDK CI Jenkins <sys_sgci@intel.com> Reviewed-by: Karol Latecki <karol.latecki@intel.com> Reviewed-by: Jim Harris <james.r.harris@intel.com> Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com> Community-CI: SPDK CI Jenkins <sys_sgci@intel.com>
298 lines
9.2 KiB
Bash
Executable File
298 lines
9.2 KiB
Bash
Executable File
#!/usr/bin/env bash
|
|
|
|
testdir=$(readlink -f $(dirname $0))
|
|
rootdir=$(readlink -f $testdir/../..)
|
|
source $rootdir/test/common/autotest_common.sh
|
|
source $testdir/nbd_common.sh
|
|
|
|
rpc_py="$rootdir/scripts/rpc.py"
|
|
|
|
function bdev_bounds() {
|
|
$testdir/bdevio/bdevio -w -s $PRE_RESERVED_MEM -c $testdir/bdev.conf &
|
|
bdevio_pid=$!
|
|
trap 'killprocess $bdevio_pid; exit 1' SIGINT SIGTERM EXIT
|
|
echo "Process bdevio pid: $bdevio_pid"
|
|
waitforlisten $bdevio_pid
|
|
$testdir/bdevio/tests.py perform_tests
|
|
killprocess $bdevio_pid
|
|
trap - SIGINT SIGTERM EXIT
|
|
}
|
|
|
|
function nbd_function_test() {
|
|
if [ $(uname -s) = Linux ] && modprobe -n nbd; then
|
|
local rpc_server=/var/tmp/spdk-nbd.sock
|
|
local conf=$1
|
|
local nbd_num=6
|
|
local nbd_all=($(ls /dev/nbd* | grep -v p))
|
|
local bdev_all=($bdevs_name)
|
|
local nbd_list=(${nbd_all[@]:0:$nbd_num})
|
|
local bdev_list=(${bdev_all[@]:0:$nbd_num})
|
|
|
|
if [ ! -e $conf ]; then
|
|
return 1
|
|
fi
|
|
|
|
modprobe nbd
|
|
$rootdir/test/app/bdev_svc/bdev_svc -r $rpc_server -i 0 -c ${conf} &
|
|
nbd_pid=$!
|
|
trap 'killprocess $nbd_pid; exit 1' SIGINT SIGTERM EXIT
|
|
echo "Process nbd pid: $nbd_pid"
|
|
waitforlisten $nbd_pid $rpc_server
|
|
|
|
nbd_rpc_start_stop_verify $rpc_server "${bdev_list[*]}"
|
|
nbd_rpc_data_verify $rpc_server "${bdev_list[*]}" "${nbd_list[*]}"
|
|
|
|
$rpc_py -s $rpc_server bdev_passthru_delete TestPT
|
|
|
|
killprocess $nbd_pid
|
|
trap - SIGINT SIGTERM EXIT
|
|
fi
|
|
|
|
return 0
|
|
}
|
|
|
|
function fio_test_suite() {
|
|
# Generate the fio config file given the list of all unclaimed bdevs
|
|
fio_config_gen $testdir/bdev.fio verify AIO
|
|
for b in $(echo $bdevs | jq -r '.name'); do
|
|
fio_config_add_job $testdir/bdev.fio $b
|
|
done
|
|
|
|
local fio_params="--ioengine=spdk_bdev --iodepth=8 --bs=4k --runtime=10 $testdir/bdev.fio --spdk_conf=./test/bdev/bdev.conf"
|
|
local fio_ext_params="--ioengine=spdk_bdev --iodepth=128 --bs=192k --runtime=100 $testdir/bdev.fio --spdk_conf=./test/bdev/bdev.conf"
|
|
|
|
if [ $RUN_NIGHTLY -eq 0 ]; then
|
|
run_test "bdev_fio_rw_verify" fio_bdev $fio_params --spdk_mem=$PRE_RESERVED_MEM \
|
|
--output=$output_dir/blockdev_fio_verify.txt
|
|
elif [ $RUN_NIGHTLY_FAILING -eq 1 ]; then
|
|
# Use size 192KB which both exceeds typical 128KB max NVMe I/O
|
|
# size and will cross 128KB Intel DC P3700 stripe boundaries.
|
|
run_test "bdev_fio_rw_verify_ext" fio_bdev $fio_ext_params --spdk_mem=$PRE_RESERVED_MEM \
|
|
--output=$output_dir/blockdev_fio_verify.txt
|
|
fi
|
|
rm -f ./*.state
|
|
rm -f $testdir/bdev.fio
|
|
|
|
# Generate the fio config file given the list of all unclaimed bdevs that support unmap
|
|
fio_config_gen $testdir/bdev.fio trim
|
|
for b in $(echo $bdevs | jq -r 'select(.supported_io_types.unmap == true) | .name'); do
|
|
fio_config_add_job $testdir/bdev.fio $b
|
|
done
|
|
|
|
if [ $RUN_NIGHTLY -eq 0 ]; then
|
|
run_test "bdev_fio_trim" fio_bdev $fio_params --output=$output_dir/blockdev_trim.txt
|
|
elif [ $RUN_NIGHTLY_FAILING -eq 1 ]; then
|
|
run_test "bdev_fio_trim_ext" fio_bdev $fio_ext_params --output=$output_dir/blockdev_trim.txt
|
|
fi
|
|
|
|
rm -f ./*.state
|
|
rm -f $testdir/bdev.fio
|
|
report_test_completion "bdev_fio"
|
|
}
|
|
|
|
function get_io_result() {
|
|
local limit_type=$1
|
|
local qos_dev=$2
|
|
|
|
io_result=$($rpc_py bdev_get_iostat -b $qos_dev)
|
|
if [ $limit_type = IOPS ]; then
|
|
io_result_before=$(echo $io_result | jq -r '.bdevs[0].num_read_ops')
|
|
else
|
|
io_result_before=$(echo $io_result | jq -r '.bdevs[0].bytes_read')
|
|
fi
|
|
tick_rate=$(echo $io_result | jq -r '.tick_rate')
|
|
ticks_before=$(echo $io_result | jq -r '.ticks')
|
|
|
|
sleep $QOS_RUN_TIME
|
|
|
|
io_result=$($rpc_py bdev_get_iostat -b $qos_dev)
|
|
if [ $limit_type = IOPS ]; then
|
|
io_result_after=$(echo $io_result | jq -r '.bdevs[0].num_read_ops')
|
|
else
|
|
io_result_after=$(echo $io_result | jq -r '.bdevs[0].bytes_read')
|
|
fi
|
|
ticks_after=$(echo $io_result | jq -r '.ticks')
|
|
|
|
if [ $limit_type = IOPS ]; then
|
|
io_result_diff=$((io_result_after-io_result_before))
|
|
else
|
|
# To avoid potential overflow as throughput is in byte
|
|
# Return throughput in kilobyte
|
|
io_result_diff=$(((io_result_after-io_result_before)/1024))
|
|
fi
|
|
echo $(((io_result_diff*tick_rate)/(ticks_after-ticks_before)))
|
|
}
|
|
|
|
function run_qos_test() {
|
|
local qos_limit=$1
|
|
local qos_result=0
|
|
|
|
qos_result=$(get_io_result $2 $3)
|
|
if [ $2 = BANDWIDTH ]; then
|
|
qos_limit=$((qos_limit*1024))
|
|
fi
|
|
lower_limit=$((qos_limit*9/10))
|
|
upper_limit=$((qos_limit*11/10))
|
|
|
|
# QoS realization is related with bytes transfered. It currently has some variation.
|
|
if [ $qos_result -lt $lower_limit ] || [ $qos_result -gt $upper_limit ]; then
|
|
echo "Failed to limit the io read rate of NULL bdev by qos"
|
|
$rpc_py bdev_null_delete $QOS_DEV_1
|
|
$rpc_py bdev_null_delete $QOS_DEV_2
|
|
killprocess $QOS_PID
|
|
exit 1
|
|
fi
|
|
}
|
|
|
|
function qos_function_test() {
|
|
local qos_lower_iops_limit=1000
|
|
local qos_lower_bw_limit=2
|
|
local io_result=0
|
|
local iops_limit=0
|
|
local bw_limit=0
|
|
|
|
io_result=$(get_io_result IOPS $QOS_DEV_1)
|
|
# Set the IOPS limit as one quarter of the measured performance without QoS
|
|
iops_limit=$(((io_result/4)/qos_lower_iops_limit*qos_lower_iops_limit))
|
|
if [ $iops_limit -gt $qos_lower_iops_limit ]; then
|
|
|
|
# Run bdevperf with IOPS rate limit on bdev 1
|
|
$rpc_py bdev_set_qos_limit --rw_ios_per_sec $iops_limit $QOS_DEV_1
|
|
run_test "bdev_qos_iops" run_qos_test $iops_limit IOPS $QOS_DEV_1
|
|
|
|
# Run bdevperf with bandwidth rate limit on bdev 2
|
|
# Set the bandwidth limit as 1/10 of the measure performance without QoS
|
|
bw_limit=$(get_io_result BANDWIDTH $QOS_DEV_2)
|
|
bw_limit=$((bw_limit/1024/10))
|
|
if [ $bw_limit -lt $qos_lower_bw_limit ]; then
|
|
bw_limit=$qos_lower_bw_limit
|
|
fi
|
|
$rpc_py bdev_set_qos_limit --rw_mbytes_per_sec $bw_limit $QOS_DEV_2
|
|
run_test "bdev_qos_bw" run_qos_test $bw_limit BANDWIDTH $QOS_DEV_2
|
|
|
|
# Run bdevperf with additional read only bandwidth rate limit on bdev 1
|
|
$rpc_py bdev_set_qos_limit --r_mbytes_per_sec $qos_lower_bw_limit $QOS_DEV_1
|
|
run_test "bdev_qos_ro_bw" run_qos_test $qos_lower_bw_limit BANDWIDTH $QOS_DEV_1
|
|
else
|
|
echo "Actual IOPS without limiting is too low - exit testing"
|
|
fi
|
|
}
|
|
|
|
function qos_test_suite() {
|
|
# Run bdevperf with QoS disabled first
|
|
$testdir/bdevperf/bdevperf -z -m 0x2 -q 256 -o 4096 -w randread -t 60 &
|
|
QOS_PID=$!
|
|
echo "Process qos testing pid: $QOS_PID"
|
|
trap 'killprocess $QOS_PID; exit 1' SIGINT SIGTERM EXIT
|
|
waitforlisten $QOS_PID
|
|
|
|
$rpc_py bdev_null_create $QOS_DEV_1 128 512
|
|
waitforbdev $QOS_DEV_1
|
|
$rpc_py bdev_null_create $QOS_DEV_2 128 512
|
|
waitforbdev $QOS_DEV_2
|
|
|
|
$rootdir/test/bdev/bdevperf/bdevperf.py perform_tests &
|
|
qos_function_test
|
|
|
|
$rpc_py bdev_null_delete $QOS_DEV_1
|
|
$rpc_py bdev_null_delete $QOS_DEV_2
|
|
killprocess $QOS_PID
|
|
trap - SIGINT SIGTERM EXIT
|
|
}
|
|
|
|
# Inital bdev creation and configuration
|
|
#-----------------------------------------------------
|
|
QOS_DEV_1="Null_0"
|
|
QOS_DEV_2="Null_1"
|
|
QOS_RUN_TIME=5
|
|
|
|
if [ $(uname -s) = Linux ]; then
|
|
# Test dynamic memory management. All hugepages will be reserved at runtime
|
|
PRE_RESERVED_MEM=0
|
|
else
|
|
# Dynamic memory management is not supported on BSD
|
|
PRE_RESERVED_MEM=2048
|
|
fi
|
|
|
|
# Create a file to be used as an AIO backend
|
|
dd if=/dev/zero of=/tmp/aiofile bs=2048 count=5000
|
|
|
|
cp $testdir/bdev.conf.in $testdir/bdev.conf
|
|
$rootdir/scripts/gen_nvme.sh >> $testdir/bdev.conf
|
|
|
|
if [ $SPDK_TEST_RBD -eq 1 ]; then
|
|
timing_enter rbd_setup
|
|
rbd_setup 127.0.0.1
|
|
timing_exit rbd_setup
|
|
|
|
$rootdir/scripts/gen_rbd.sh >> $testdir/bdev.conf
|
|
fi
|
|
|
|
if [ $SPDK_TEST_CRYPTO -eq 1 ]; then
|
|
$testdir/gen_crypto.sh Malloc6 Malloc7 >> $testdir/bdev.conf
|
|
fi
|
|
|
|
if hash pmempool; then
|
|
rm -f /tmp/spdk-pmem-pool
|
|
pmempool create blk --size=32M 512 /tmp/spdk-pmem-pool
|
|
echo "[Pmem]" >> $testdir/bdev.conf
|
|
echo " Blk /tmp/spdk-pmem-pool Pmem0" >> $testdir/bdev.conf
|
|
fi
|
|
|
|
timing_enter nbd_gpt
|
|
if grep -q Nvme0 $testdir/bdev.conf; then
|
|
part_dev_by_gpt $testdir/bdev.conf Nvme0n1 $rootdir
|
|
fi
|
|
timing_exit nbd_gpt
|
|
|
|
timing_enter bdev_svc
|
|
bdevs=$(discover_bdevs $rootdir $testdir/bdev.conf | jq -r '.[] | select(.claimed == false)')
|
|
timing_exit bdev_svc
|
|
|
|
bdevs_name=$(echo $bdevs | jq -r '.name')
|
|
|
|
# Create conf file for bdevperf with gpt
|
|
cat > $testdir/bdev_gpt.conf << EOL
|
|
[Gpt]
|
|
Disable No
|
|
EOL
|
|
|
|
# Get Nvme info through filtering gen_nvme.sh's result
|
|
$rootdir/scripts/gen_nvme.sh >> $testdir/bdev_gpt.conf
|
|
# End bdev configuration
|
|
#-----------------------------------------------------
|
|
|
|
run_test "bdev_hello_world" $rootdir/examples/bdev/hello_world/hello_bdev -c $testdir/bdev.conf -b Malloc0
|
|
run_test "bdev_bounds" bdev_bounds
|
|
run_test "bdev_nbd" nbd_function_test $testdir/bdev.conf "$bdevs_name"
|
|
if [ -d /usr/src/fio ]; then
|
|
run_test "bdev_fio" fio_test_suite
|
|
else
|
|
echo "FIO not available"
|
|
exit 1
|
|
fi
|
|
|
|
# Run bdevperf with gpt
|
|
run_test "bdev_gpt_verify" $testdir/bdevperf/bdevperf -c $testdir/bdev_gpt.conf -q 128 -o 4096 -w verify -t 5
|
|
run_test "bdev_gpt_write_zeroes" $testdir/bdevperf/bdevperf -c $testdir/bdev_gpt.conf -q 128 -o 4096 -w write_zeroes -t 1
|
|
run_test "bdev_qos" qos_test_suite
|
|
|
|
# Temporarily disabled - infinite loop
|
|
# if [ $RUN_NIGHTLY -eq 1 ]; then
|
|
# run_test "bdev_gpt_reset" $testdir/bdevperf/bdevperf -c $testdir/bdev.conf -q 16 -w reset -o 4096 -t 60
|
|
# report_test_completion "nightly_bdev_reset"
|
|
# fi
|
|
|
|
# Bdev and configuration cleanup below this line
|
|
#-----------------------------------------------------
|
|
if grep -q Nvme0 $testdir/bdev.conf; then
|
|
part_dev_by_gpt $testdir/bdev.conf Nvme0n1 $rootdir reset
|
|
fi
|
|
|
|
rm -f $testdir/bdev_gpt.conf
|
|
rm -f /tmp/aiofile
|
|
rm -f /tmp/spdk-pmem-pool
|
|
rm -f $testdir/bdev.conf
|
|
rbd_cleanup
|
|
report_test_completion "bdev"
|