numam-spdk/test/ocf/common.sh
Michal Berger 1dfac8ba69 test/ocf: Don't depend on gen_nvme.sh while detecting nvme ctrls
gen_nvme.sh requires that all the nvme devices are not bound to the
nvme driver which doesn't have to be the case when clear_nvme() is
called. To not depend on which driver is in use, simply iterate over
pci bus to find a proper device. This avoids additional setup.sh
ping-pong which is required to make use of gen_nvme.sh.

Additionally, upon exiting, in case mounts for given nvme device were
detected, keep the test environment reset to leave nvmes bound to the
nvme driver.

Change-Id: I428c5e2b0517553d267411a30316bde5d1ce8eef
Signed-off-by: Michal Berger <michalx.berger@intel.com>
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/1843
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Reviewed-by: Tomasz Zawadzki <tomasz.zawadzki@intel.com>
Reviewed-by: Shuhei Matsumoto <shuhei.matsumoto.xt@hitachi.com>
2020-04-20 14:28:56 +00:00

30 lines
648 B
Bash

source $rootdir/scripts/common.sh
source $rootdir/test/common/autotest_common.sh
rpc_py=$rootdir/scripts/rpc.py
function nvme_cfg() {
if [ -z "$ocf_nvme_cfg" ]; then
ocf_nvme_cfg=$($rootdir/scripts/gen_nvme.sh)
fi
echo "$ocf_nvme_cfg"
}
function clear_nvme()
{
mapfile -t bdf < <(iter_all_pci_class_code 01 08 02)
# Clear metadata on NVMe device
$rootdir/scripts/setup.sh reset
sleep 5
name=$(get_nvme_name_from_bdf "${bdf[0]}")
mountpoints=$(lsblk /dev/$name --output MOUNTPOINT -n | wc -w)
if [ "$mountpoints" != "0" ]; then
exit 1
fi
dd if=/dev/zero of=/dev/$name bs=1M count=1000 oflag=direct
$rootdir/scripts/setup.sh
}