mirror of
https://git.proxmox.com/git/mirror_zfs.git
synced 2024-12-26 03:09:34 +03:00
Retire zconfig.sh
All of the test coverage provided by this script is now handled as part of the ZFS Test Suite. Remove it. Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov> Closes #6128
This commit is contained in:
parent
3f03fc8df3
commit
4358afa0f2
4
TEST
4
TEST
@ -13,10 +13,6 @@
|
||||
#TEST_ZTEST_DIR="/var/tmp/"
|
||||
#TEST_ZTEST_OPTIONS="-V"
|
||||
|
||||
### zconfig
|
||||
#TEST_ZCONFIG_SKIP="yes"
|
||||
TEST_ZCONFIG_OPTIONS="-c -s10"
|
||||
|
||||
### zimport
|
||||
#TEST_ZIMPORT_SKIP="yes"
|
||||
#TEST_ZIMPORT_DIR="/var/tmp/zimport"
|
||||
|
@ -5,7 +5,6 @@ EXTRA_DIST = dkms.mkconf dkms.postbuild kmodtool zfs2zol-patch.sed cstyle.pl
|
||||
pkgdatadir = $(datadir)/@PACKAGE@
|
||||
dist_pkgdata_SCRIPTS = \
|
||||
$(top_builddir)/scripts/common.sh \
|
||||
$(top_srcdir)/scripts/zconfig.sh \
|
||||
$(top_srcdir)/scripts/zimport.sh \
|
||||
$(top_srcdir)/scripts/zfs.sh \
|
||||
$(top_srcdir)/scripts/zfs-tests.sh \
|
||||
|
@ -1,696 +0,0 @@
|
||||
#!/bin/bash
|
||||
#
|
||||
# ZFS/ZPOOL configuration test script.
|
||||
|
||||
basedir="$(dirname $0)"
|
||||
|
||||
SCRIPT_COMMON=common.sh
|
||||
if [ -f "${basedir}/${SCRIPT_COMMON}" ]; then
|
||||
. "${basedir}/${SCRIPT_COMMON}"
|
||||
else
|
||||
echo "Missing helper script ${SCRIPT_COMMON}" && exit 1
|
||||
fi
|
||||
|
||||
PROG=zconfig.sh
|
||||
|
||||
usage() {
|
||||
cat << EOF
|
||||
USAGE:
|
||||
$0 [hvcts]
|
||||
|
||||
DESCRIPTION:
|
||||
ZFS/ZPOOL configuration tests
|
||||
|
||||
OPTIONS:
|
||||
-h Show this message
|
||||
-v Verbose
|
||||
-c Cleanup lo+file devices at start
|
||||
-t <#> Run listed tests
|
||||
-s <#> Skip listed tests
|
||||
|
||||
EOF
|
||||
}
|
||||
|
||||
while getopts 'hvct:s:?' OPTION; do
|
||||
case $OPTION in
|
||||
h)
|
||||
usage
|
||||
exit 1
|
||||
;;
|
||||
v)
|
||||
VERBOSE=1
|
||||
;;
|
||||
c)
|
||||
CLEANUP=1
|
||||
;;
|
||||
t)
|
||||
TESTS_RUN=($OPTARG)
|
||||
;;
|
||||
s)
|
||||
TESTS_SKIP=($OPTARG)
|
||||
;;
|
||||
?)
|
||||
usage
|
||||
exit
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
if [ $(id -u) != 0 ]; then
|
||||
die "Must run as root"
|
||||
fi
|
||||
|
||||
# Initialize the test suite
|
||||
init
|
||||
|
||||
# Disable the udev rule 90-zfs.rules to prevent the zfs module
|
||||
# stack from being loaded due to the detection of a zfs device.
|
||||
# This is important because this test scripts require full control
|
||||
# over when and how the modules are loaded/unloaded. A trap is
|
||||
# set to ensure the udev rule is correctly replaced on exit.
|
||||
RULE=${udevruledir}/90-zfs.rules
|
||||
if test -e ${RULE}; then
|
||||
trap "mv ${RULE}.disabled ${RULE}" INT TERM EXIT
|
||||
mv ${RULE} ${RULE}.disabled
|
||||
fi
|
||||
|
||||
# Perform pre-cleanup is requested
|
||||
if [ ${CLEANUP} ]; then
|
||||
${ZFS_SH} -u
|
||||
cleanup_md_devices
|
||||
cleanup_loop_devices
|
||||
rm -f /tmp/zpool.cache.*
|
||||
fi
|
||||
|
||||
# Check if we need to skip the tests that require scsi_debug and lsscsi.
|
||||
SCSI_DEBUG=0
|
||||
${INFOMOD} scsi_debug &>/dev/null && SCSI_DEBUG=1
|
||||
HAVE_LSSCSI=0
|
||||
test -f ${LSSCSI} && HAVE_LSSCSI=1
|
||||
if [ ${SCSI_DEBUG} -eq 0 ] || [ ${HAVE_LSSCSI} -eq 0 ]; then
|
||||
echo "Skipping test 10 which requires the scsi_debug " \
|
||||
"module and the ${LSSCSI} utility"
|
||||
fi
|
||||
|
||||
# Validate persistent zpool.cache configuration.
|
||||
test_1() {
|
||||
local POOL_NAME=test1
|
||||
local TMP_FILE1=`mktemp`
|
||||
local TMP_FILE2=`mktemp`
|
||||
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
|
||||
# Create a pool save its status for comparison.
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
|
||||
|
||||
# Unload/load the module stack and verify the pool persists.
|
||||
${ZFS_SH} -u || fail 4
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 5
|
||||
${ZPOOL} import -c ${TMP_CACHE} ${POOL_NAME} || fail 5
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 6
|
||||
cmp ${TMP_FILE1} ${TMP_FILE2} || fail 7
|
||||
|
||||
# Cleanup the test pool and temporary files
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 8
|
||||
rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_CACHE} || fail 9
|
||||
${ZFS_SH} -u || fail 10
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 1 "persistent zpool.cache"
|
||||
|
||||
# Validate ZFS disk scanning and import w/out zpool.cache configuration.
|
||||
test_2() {
|
||||
local POOL_NAME=test2
|
||||
local TMP_FILE1=`mktemp`
|
||||
local TMP_FILE2=`mktemp`
|
||||
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
|
||||
# Create a pool save its status for comparison.
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
|
||||
|
||||
# Unload the module stack, remove the cache file, load the module
|
||||
# stack and attempt to probe the disks to import the pool. As
|
||||
# a cross check verify the old pool state against the imported.
|
||||
${ZFS_SH} -u || fail 4
|
||||
rm -f ${TMP_CACHE} || fail 5
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 6
|
||||
${ZPOOL} import -d /dev ${POOL_NAME} || fail 8
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 9
|
||||
cmp ${TMP_FILE1} ${TMP_FILE2} || fail 10
|
||||
|
||||
# Cleanup the test pool and temporary files
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 11
|
||||
rm -f ${TMP_FILE1} ${TMP_FILE2} || fail 12
|
||||
${ZFS_SH} -u || fail 13
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 2 "scan disks for pools to import"
|
||||
|
||||
zconfig_zvol_device_stat() {
|
||||
local EXPECT=$1
|
||||
local POOL_NAME=/dev/zvol/$2
|
||||
local ZVOL_NAME=/dev/zvol/$3
|
||||
local SNAP_NAME=/dev/zvol/$4
|
||||
local CLONE_NAME=/dev/zvol/$5
|
||||
local COUNT=0
|
||||
|
||||
# Briefly delay for udev
|
||||
udev_trigger
|
||||
|
||||
# Pool exists
|
||||
stat ${POOL_NAME} &>/dev/null && let COUNT=$COUNT+1
|
||||
|
||||
# Volume and partitions
|
||||
stat ${ZVOL_NAME} &>/dev/null && let COUNT=$COUNT+1
|
||||
stat ${ZVOL_NAME}-part1 &>/dev/null && let COUNT=$COUNT+1
|
||||
stat ${ZVOL_NAME}-part2 &>/dev/null && let COUNT=$COUNT+1
|
||||
|
||||
# Snapshot with partitions
|
||||
stat ${SNAP_NAME} &>/dev/null && let COUNT=$COUNT+1
|
||||
stat ${SNAP_NAME}-part1 &>/dev/null && let COUNT=$COUNT+1
|
||||
stat ${SNAP_NAME}-part2 &>/dev/null && let COUNT=$COUNT+1
|
||||
|
||||
# Clone with partitions
|
||||
stat ${CLONE_NAME} &>/dev/null && let COUNT=$COUNT+1
|
||||
stat ${CLONE_NAME}-part1 &>/dev/null && let COUNT=$COUNT+1
|
||||
stat ${CLONE_NAME}-part2 &>/dev/null && let COUNT=$COUNT+1
|
||||
|
||||
if [ $EXPECT -ne $COUNT ]; then
|
||||
return 1
|
||||
fi
|
||||
|
||||
return 0
|
||||
}
|
||||
|
||||
# zpool import/export device check
|
||||
# (1 volume, 2 partitions, 1 snapshot, 1 clone)
|
||||
test_3() {
|
||||
local POOL_NAME=tank
|
||||
local ZVOL_NAME=volume
|
||||
local SNAP_NAME=snap
|
||||
local CLONE_NAME=clone
|
||||
local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
||||
local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
|
||||
local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
|
||||
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
|
||||
# Create a pool, volume, partition, snapshot, and clone.
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
||||
${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3
|
||||
${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
|
||||
label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
|
||||
partition /dev/zvol/${FULL_ZVOL_NAME} primary 1% 50% || fail 4
|
||||
partition /dev/zvol/${FULL_ZVOL_NAME} primary 51% -1 || fail 4
|
||||
${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5
|
||||
${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6
|
||||
|
||||
# Verify the devices were created
|
||||
zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7
|
||||
|
||||
# Export the pool
|
||||
${ZPOOL} export ${POOL_NAME} || fail 8
|
||||
|
||||
# verify the devices were removed
|
||||
zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9
|
||||
|
||||
# Import the pool, wait 1 second for udev
|
||||
${ZPOOL} import ${POOL_NAME} || fail 10
|
||||
|
||||
# Verify the devices were created
|
||||
zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11
|
||||
|
||||
# Toggle the snapdev and observe snapshot device links toggled
|
||||
${ZFS} set snapdev=hidden ${FULL_ZVOL_NAME} || fail 12
|
||||
|
||||
zconfig_zvol_device_stat 7 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
"invalid" ${FULL_CLONE_NAME} || fail 13
|
||||
|
||||
${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 14
|
||||
|
||||
zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 15
|
||||
|
||||
# Destroy the pool and consequently the devices
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 16
|
||||
|
||||
# verify the devices were removed
|
||||
zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 17
|
||||
|
||||
${ZFS_SH} -u || fail 18
|
||||
rm -f ${TMP_CACHE} || fail 19
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 3 "zpool import/export device"
|
||||
|
||||
# zpool insmod/rmmod device check (1 volume, 1 snapshot, 1 clone)
|
||||
test_4() {
|
||||
POOL_NAME=tank
|
||||
ZVOL_NAME=volume
|
||||
SNAP_NAME=snap
|
||||
CLONE_NAME=clone
|
||||
FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
||||
FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
|
||||
FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
|
||||
TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
|
||||
# Create a pool, volume, snapshot, and clone
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
||||
${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3
|
||||
${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
|
||||
label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
|
||||
partition /dev/zvol/${FULL_ZVOL_NAME} primary 1% 50% || fail 4
|
||||
partition /dev/zvol/${FULL_ZVOL_NAME} primary 51% -1 || fail 4
|
||||
${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5
|
||||
${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6
|
||||
|
||||
# Verify the devices were created
|
||||
zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7
|
||||
|
||||
# Unload the modules
|
||||
${ZFS_SH} -u || fail 8
|
||||
|
||||
# Verify the devices were removed
|
||||
zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9
|
||||
|
||||
# Load the modules, list the pools to ensure they are opened
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 10
|
||||
${ZPOOL} import -c ${TMP_CACHE} ${POOL_NAME} || fail 10
|
||||
${ZPOOL} list &>/dev/null
|
||||
|
||||
# Verify the devices were created
|
||||
zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11
|
||||
|
||||
# Destroy the pool and consequently the devices
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 12
|
||||
|
||||
# Verify the devices were removed
|
||||
zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
||||
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 13
|
||||
|
||||
${ZFS_SH} -u || fail 14
|
||||
rm -f ${TMP_CACHE} || fail 15
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 4 "zpool insmod/rmmod device"
|
||||
|
||||
# ZVOL volume sanity check
|
||||
test_5() {
|
||||
local POOL_NAME=tank
|
||||
local ZVOL_NAME=fish
|
||||
local FULL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
||||
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
|
||||
# Create a pool and volume.
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raid0 || fail 2
|
||||
${ZFS} create -V 800M ${FULL_NAME} || fail 3
|
||||
label /dev/zvol/${FULL_NAME} msdos || fail 4
|
||||
partition /dev/zvol/${FULL_NAME} primary 1 -1 || fail 4
|
||||
format /dev/zvol/${FULL_NAME}-part1 ext2 || fail 5
|
||||
|
||||
# Mount the ext2 filesystem and copy some data to it.
|
||||
mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 6
|
||||
mount /dev/zvol/${FULL_NAME}-part1 /tmp/${ZVOL_NAME}-part1 || fail 7
|
||||
cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 8
|
||||
sync
|
||||
|
||||
# Verify the copied files match the original files.
|
||||
diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \
|
||||
&>/dev/null || fail 9
|
||||
|
||||
# Remove the files, umount, destroy the volume and pool.
|
||||
rm -Rf /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} || fail 10
|
||||
umount /tmp/${ZVOL_NAME}-part1 || fail 11
|
||||
rmdir /tmp/${ZVOL_NAME}-part1 || fail 12
|
||||
|
||||
${ZFS} destroy ${FULL_NAME} || fail 13
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 14
|
||||
${ZFS_SH} -u || fail 15
|
||||
rm -f ${TMP_CACHE} || fail 16
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 5 "zvol+ext2 volume"
|
||||
|
||||
# ZVOL snapshot sanity check
|
||||
test_6() {
|
||||
local POOL_NAME=tank
|
||||
local ZVOL_NAME=fish
|
||||
local SNAP_NAME=pristine
|
||||
local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
||||
local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
|
||||
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
|
||||
# Create a pool and volume.
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raid0 || fail 2
|
||||
${ZFS} create -s -V 800M ${FULL_ZVOL_NAME} || fail 3
|
||||
${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
|
||||
label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
|
||||
partition /dev/zvol/${FULL_ZVOL_NAME} primary 1 -1 || fail 4
|
||||
format /dev/zvol/${FULL_ZVOL_NAME}-part1 ext2 || fail 5
|
||||
|
||||
# Mount the ext2 filesystem and copy some data to it.
|
||||
mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 6
|
||||
mount /dev/zvol/${FULL_ZVOL_NAME}-part1 /tmp/${ZVOL_NAME}-part1 \
|
||||
|| fail 7
|
||||
|
||||
# Snapshot the pristine ext2 filesystem and mount it read-only.
|
||||
${ZFS} snapshot ${FULL_SNAP_NAME} || fail 8
|
||||
wait_udev /dev/zvol/${FULL_SNAP_NAME}-part1 30 || fail 8
|
||||
mkdir -p /tmp/${SNAP_NAME}-part1 || fail 9
|
||||
mount /dev/zvol/${FULL_SNAP_NAME}-part1 /tmp/${SNAP_NAME}-part1 \
|
||||
&>/dev/null || fail 10
|
||||
|
||||
# Copy to original volume
|
||||
cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 11
|
||||
sync
|
||||
|
||||
# Verify the copied files match the original files,
|
||||
# and the copied files do NOT appear in the snapshot.
|
||||
diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \
|
||||
&>/dev/null || fail 12
|
||||
diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}-part1/${SRC_DIR##*/} \
|
||||
&>/dev/null && fail 13
|
||||
|
||||
# umount, destroy the snapshot, volume, and pool.
|
||||
umount /tmp/${SNAP_NAME}-part1 || fail 14
|
||||
rmdir /tmp/${SNAP_NAME}-part1 || fail 15
|
||||
${ZFS} destroy ${FULL_SNAP_NAME} || fail 16
|
||||
|
||||
umount /tmp/${ZVOL_NAME}-part1 || fail 17
|
||||
rmdir /tmp/${ZVOL_NAME}-part1 || fail 18
|
||||
${ZFS} destroy ${FULL_ZVOL_NAME} || fail 19
|
||||
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 20
|
||||
${ZFS_SH} -u || fail 21
|
||||
rm -f ${TMP_CACHE} || fail 22
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 6 "zvol+ext2 snapshot"
|
||||
|
||||
# ZVOL clone sanity check
|
||||
test_7() {
|
||||
local POOL_NAME=tank
|
||||
local ZVOL_NAME=fish
|
||||
local SNAP_NAME=pristine
|
||||
local CLONE_NAME=clone
|
||||
local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
||||
local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
|
||||
local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
|
||||
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
|
||||
# Create a pool and volume.
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
||||
${ZFS} create -V 300M ${FULL_ZVOL_NAME} || fail 3
|
||||
${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
|
||||
label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
|
||||
partition /dev/zvol/${FULL_ZVOL_NAME} primary 1 -1 || fail 4
|
||||
format /dev/zvol/${FULL_ZVOL_NAME}-part1 ext2 || fail 5
|
||||
|
||||
# Snapshot the pristine ext2 filesystem.
|
||||
${ZFS} snapshot ${FULL_SNAP_NAME} || fail 6
|
||||
wait_udev /dev/zvol/${FULL_SNAP_NAME}-part1 30 || fail 7
|
||||
|
||||
# Mount the ext2 filesystem so some data can be copied to it.
|
||||
mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 7
|
||||
mount /dev/zvol/${FULL_ZVOL_NAME}-part1 \
|
||||
/tmp/${ZVOL_NAME}-part1 || fail 8
|
||||
|
||||
# Mount the pristine ext2 snapshot.
|
||||
mkdir -p /tmp/${SNAP_NAME}-part1 || fail 9
|
||||
mount /dev/zvol/${FULL_SNAP_NAME}-part1 \
|
||||
/tmp/${SNAP_NAME}-part1 &>/dev/null || fail 10
|
||||
|
||||
# Copy to original volume.
|
||||
cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 11
|
||||
sync
|
||||
|
||||
# Verify the copied files match the original files,
|
||||
# and the copied files do NOT appear in the snapshot.
|
||||
diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \
|
||||
&>/dev/null || fail 12
|
||||
diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}-part1/${SRC_DIR##*/} \
|
||||
&>/dev/null && fail 13
|
||||
|
||||
# Clone from the original pristine snapshot
|
||||
${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 14
|
||||
wait_udev /dev/zvol/${FULL_CLONE_NAME}-part1 30 || fail 14
|
||||
mkdir -p /tmp/${CLONE_NAME}-part1 || fail 15
|
||||
mount /dev/zvol/${FULL_CLONE_NAME}-part1 \
|
||||
/tmp/${CLONE_NAME}-part1 || fail 16
|
||||
|
||||
# Verify the clone matches the pristine snapshot,
|
||||
# and the files copied to the original volume are NOT there.
|
||||
diff -ur /tmp/${SNAP_NAME}-part1 /tmp/${CLONE_NAME}-part1 \
|
||||
&>/dev/null || fail 17
|
||||
diff -ur /tmp/${ZVOL_NAME}-part1 /tmp/${CLONE_NAME}-part1 \
|
||||
&>/dev/null && fail 18
|
||||
|
||||
# Copy to cloned volume.
|
||||
cp -RL ${SRC_DIR} /tmp/${CLONE_NAME}-part1 || fail 19
|
||||
sync
|
||||
|
||||
# Verify the clone matches the modified original volume.
|
||||
diff -ur /tmp/${ZVOL_NAME}-part1 /tmp/${CLONE_NAME}-part1 \
|
||||
&>/dev/null || fail 20
|
||||
|
||||
# umount, destroy the snapshot, volume, and pool.
|
||||
umount /tmp/${CLONE_NAME}-part1 || fail 21
|
||||
rmdir /tmp/${CLONE_NAME}-part1 || fail 22
|
||||
${ZFS} destroy ${FULL_CLONE_NAME} || fail 23
|
||||
|
||||
umount /tmp/${SNAP_NAME}-part1 || fail 24
|
||||
rmdir /tmp/${SNAP_NAME}-part1 || fail 25
|
||||
${ZFS} destroy ${FULL_SNAP_NAME} || fail 26
|
||||
|
||||
umount /tmp/${ZVOL_NAME}-part1 || fail 27
|
||||
rmdir /tmp/${ZVOL_NAME}-part1 || fail 28
|
||||
${ZFS} destroy ${FULL_ZVOL_NAME} || fail 29
|
||||
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 30
|
||||
${ZFS_SH} -u || fail 31
|
||||
rm -f ${TMP_CACHE} || fail 32
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 7 "zvol+ext2 clone"
|
||||
|
||||
# Send/Receive sanity check
|
||||
test_8() {
|
||||
local POOL_NAME1=tank1
|
||||
local POOL_NAME2=tank2
|
||||
local ZVOL_NAME=fish
|
||||
local SNAP_NAME=snap
|
||||
local FULL_ZVOL_NAME1=${POOL_NAME1}/${ZVOL_NAME}
|
||||
local FULL_ZVOL_NAME2=${POOL_NAME2}/${ZVOL_NAME}
|
||||
local FULL_SNAP_NAME1=${POOL_NAME1}/${ZVOL_NAME}@${SNAP_NAME}
|
||||
local FULL_SNAP_NAME2=${POOL_NAME2}/${ZVOL_NAME}@${SNAP_NAME}
|
||||
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
|
||||
# Create two pools and a volume
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 || fail 2
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 || fail 2
|
||||
${ZFS} create -V 300M ${FULL_ZVOL_NAME1} || fail 3
|
||||
${ZFS} set snapdev=visible ${FULL_ZVOL_NAME1} || fail 3
|
||||
label /dev/zvol/${FULL_ZVOL_NAME1} msdos || fail 4
|
||||
partition /dev/zvol/${FULL_ZVOL_NAME1} primary 1 -1 || fail 4
|
||||
format /dev/zvol/${FULL_ZVOL_NAME1}-part1 ext2 || fail 5
|
||||
|
||||
# Mount the ext2 filesystem and copy some data to it.
|
||||
mkdir -p /tmp/${FULL_ZVOL_NAME1}-part1 || fail 6
|
||||
mount /dev/zvol/${FULL_ZVOL_NAME1}-part1 \
|
||||
/tmp/${FULL_ZVOL_NAME1}-part1 || fail 7
|
||||
cp -RL ${SRC_DIR} /tmp/${FULL_ZVOL_NAME1}-part1 || fail 8
|
||||
|
||||
# Unmount, snapshot, mount the ext2 filesystem so it may be sent.
|
||||
# We only unmount to ensure the ext2 filesystem is clean.
|
||||
umount /tmp/${FULL_ZVOL_NAME1}-part1 || fail 9
|
||||
${ZFS} snapshot ${FULL_SNAP_NAME1} || fail 10
|
||||
wait_udev /dev/zvol/${FULL_SNAP_NAME1} 30 || fail 10
|
||||
mount /dev/zvol/${FULL_ZVOL_NAME1}-part1 \
|
||||
/tmp/${FULL_ZVOL_NAME1}-part1 || 11
|
||||
|
||||
# Send/receive the snapshot from POOL_NAME1 to POOL_NAME2
|
||||
(${ZFS} send ${FULL_SNAP_NAME1} | \
|
||||
${ZFS} receive ${FULL_ZVOL_NAME2}) || fail 12
|
||||
wait_udev /dev/zvol/${FULL_ZVOL_NAME2}-part1 30 || fail 12
|
||||
|
||||
# Mount the sent ext2 filesystem.
|
||||
mkdir -p /tmp/${FULL_ZVOL_NAME2}-part1 || fail 13
|
||||
mount /dev/zvol/${FULL_ZVOL_NAME2}-part1 \
|
||||
/tmp/${FULL_ZVOL_NAME2}-part1 || fail 14
|
||||
|
||||
# Verify the contents of the volumes match
|
||||
diff -ur /tmp/${FULL_ZVOL_NAME1}-part1 /tmp/${FULL_ZVOL_NAME2}-part1 \
|
||||
&>/dev/null || fail 15
|
||||
|
||||
# Umount, destroy the volume and pool.
|
||||
umount /tmp/${FULL_ZVOL_NAME1}-part1 || fail 16
|
||||
umount /tmp/${FULL_ZVOL_NAME2}-part1 || fail 17
|
||||
rmdir /tmp/${FULL_ZVOL_NAME1}-part1 || fail 18
|
||||
rmdir /tmp/${FULL_ZVOL_NAME2}-part1 || fail 19
|
||||
rmdir /tmp/${POOL_NAME1} || fail 20
|
||||
rmdir /tmp/${POOL_NAME2} || fail 21
|
||||
|
||||
${ZFS} destroy ${FULL_SNAP_NAME1} || fail 22
|
||||
${ZFS} destroy ${FULL_SNAP_NAME2} || fail 23
|
||||
${ZFS} destroy ${FULL_ZVOL_NAME1} || fail 24
|
||||
${ZFS} destroy ${FULL_ZVOL_NAME2} || fail 25
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 -d || fail 26
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 -d || fail 27
|
||||
${ZFS_SH} -u || fail 28
|
||||
rm -f ${TMP_CACHE} || fail 29
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 8 "zfs send/receive"
|
||||
|
||||
# zpool event sanity check
|
||||
test_9() {
|
||||
local POOL_NAME=tank
|
||||
local ZVOL_NAME=fish
|
||||
local FULL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
||||
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
local TMP_EVENTS=`mktemp -p /tmp zpool.events.XXXXXXXX`
|
||||
|
||||
# Create a pool and volume.
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
||||
${ZFS} create -V 300M ${FULL_NAME} || fail 3
|
||||
udev_trigger
|
||||
|
||||
# Dump the events, there should be a pool create event
|
||||
${ZPOOL} events >${TMP_EVENTS} || fail 4
|
||||
MATCHES=`grep -c sysevent\.fs\.zfs\.pool_create ${TMP_EVENTS}`
|
||||
[ $MATCHES -eq 1 ] || fail 5
|
||||
|
||||
# Clear the events and ensure there are none.
|
||||
${ZPOOL} events -c >/dev/null || fail 6
|
||||
${ZPOOL} events >${TMP_EVENTS} || fail 7
|
||||
EVENTS=`wc -l ${TMP_EVENTS} | cut -f1 -d' '`
|
||||
[ $EVENTS -gt 1 ] && fail 8
|
||||
|
||||
${ZFS} destroy ${FULL_NAME} || fail 9
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 10
|
||||
${ZFS_SH} -u || fail 11
|
||||
rm -f ${TMP_CACHE} || fail 12
|
||||
rm -f ${TMP_EVENTS} || fail 13
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 9 "zpool events"
|
||||
|
||||
zconfig_add_vdev() {
|
||||
local POOL_NAME=$1
|
||||
local TYPE=$2
|
||||
local DEVICE=$3
|
||||
local TMP_FILE1=`mktemp`
|
||||
local TMP_FILE2=`mktemp`
|
||||
local TMP_FILE3=`mktemp`
|
||||
|
||||
BASE_DEVICE=`basename ${DEVICE}`
|
||||
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE1}
|
||||
${ZPOOL} add -f ${POOL_NAME} ${TYPE} ${DEVICE} 2>/dev/null || return 1
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE2}
|
||||
diff ${TMP_FILE1} ${TMP_FILE2} > ${TMP_FILE3}
|
||||
|
||||
[ `wc -l ${TMP_FILE3}|${AWK} '{print $1}'` -eq 3 ] || return 1
|
||||
|
||||
PARENT_VDEV=`tail -2 ${TMP_FILE3} | head -1 | ${AWK} '{print $NF}'`
|
||||
case $TYPE in
|
||||
cache)
|
||||
[ "${PARENT_VDEV}" = "${TYPE}" ] || return 1
|
||||
;;
|
||||
log)
|
||||
[ "${PARENT_VDEV}" = "logs" ] || return 1
|
||||
;;
|
||||
esac
|
||||
|
||||
if ! tail -1 ${TMP_FILE3} |
|
||||
egrep -q "^>[[:space:]]+${BASE_DEVICE}[[:space:]]+ONLINE" ; then
|
||||
return 1
|
||||
fi
|
||||
rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_FILE3}
|
||||
|
||||
return 0
|
||||
}
|
||||
|
||||
# zpool add and remove sanity check
|
||||
test_10() {
|
||||
local POOL_NAME=tank
|
||||
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
||||
local TMP_FILE1=`mktemp`
|
||||
local TMP_FILE2=`mktemp`
|
||||
|
||||
if [ ${SCSI_DEBUG} -eq 0 ] || [ ${HAVE_LSSCSI} -eq 0 ] ; then
|
||||
skip
|
||||
return
|
||||
fi
|
||||
|
||||
test `${LSMOD} | grep -c scsi_debug` -gt 0 && \
|
||||
(${RMMOD} scsi_debug || exit 1)
|
||||
|
||||
/sbin/modprobe scsi_debug dev_size_mb=128 ||
|
||||
die "Error $? creating scsi_debug device"
|
||||
udev_trigger
|
||||
|
||||
SDDEVICE=`${LSSCSI}|${AWK} '/scsi_debug/ { print $6; exit }'`
|
||||
BASE_SDDEVICE=`basename $SDDEVICE`
|
||||
|
||||
# Create a pool
|
||||
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
|
||||
|
||||
# Add and remove a cache vdev by full path
|
||||
zconfig_add_vdev ${POOL_NAME} cache ${SDDEVICE} || fail 4
|
||||
${ZPOOL} remove ${POOL_NAME} ${SDDEVICE} || fail 5
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 6
|
||||
cmp ${TMP_FILE1} ${TMP_FILE2} || fail 7
|
||||
sleep 1
|
||||
|
||||
# Add and remove a cache vdev by shorthand path
|
||||
zconfig_add_vdev ${POOL_NAME} cache ${BASE_SDDEVICE} || fail 8
|
||||
${ZPOOL} remove ${POOL_NAME} ${BASE_SDDEVICE} || fail 9
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 10
|
||||
cmp ${TMP_FILE1} ${TMP_FILE2} || fail 11
|
||||
sleep 1
|
||||
|
||||
# Add and remove a log vdev
|
||||
zconfig_add_vdev ${POOL_NAME} log ${BASE_SDDEVICE} || fail 12
|
||||
${ZPOOL} remove ${POOL_NAME} ${BASE_SDDEVICE} || fail 13
|
||||
${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 14
|
||||
cmp ${TMP_FILE1} ${TMP_FILE2} || fail 15
|
||||
|
||||
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 16
|
||||
${ZFS_SH} -u || fail 17
|
||||
${RMMOD} scsi_debug || fail 18
|
||||
|
||||
rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_CACHE} || fail 19
|
||||
|
||||
pass
|
||||
}
|
||||
run_test 10 "zpool add/remove vdev"
|
||||
|
||||
exit 0
|
Loading…
Reference in New Issue
Block a user