2010-08-26 22:22:58 +04:00
|
|
|
#!/bin/bash
|
|
|
|
#
|
|
|
|
# ZFS/ZPOOL configuration test script.
|
|
|
|
|
|
|
|
basedir="$(dirname $0)"
|
|
|
|
|
|
|
|
SCRIPT_COMMON=common.sh
|
|
|
|
if [ -f "${basedir}/${SCRIPT_COMMON}" ]; then
|
|
|
|
. "${basedir}/${SCRIPT_COMMON}"
|
|
|
|
else
|
|
|
|
echo "Missing helper script ${SCRIPT_COMMON}" && exit 1
|
|
|
|
fi
|
|
|
|
|
|
|
|
PROG=zconfig.sh
|
|
|
|
|
|
|
|
usage() {
|
|
|
|
cat << EOF
|
|
|
|
USAGE:
|
|
|
|
$0 [hvc]
|
|
|
|
|
|
|
|
DESCRIPTION:
|
|
|
|
ZFS/ZPOOL configuration tests
|
|
|
|
|
|
|
|
OPTIONS:
|
|
|
|
-h Show this message
|
|
|
|
-v Verbose
|
|
|
|
-c Cleanup lo+file devices at start
|
|
|
|
|
|
|
|
EOF
|
|
|
|
}
|
|
|
|
|
2010-08-26 22:44:39 +04:00
|
|
|
while getopts 'hvct:s:?' OPTION; do
|
2010-08-26 22:22:58 +04:00
|
|
|
case $OPTION in
|
|
|
|
h)
|
|
|
|
usage
|
|
|
|
exit 1
|
|
|
|
;;
|
|
|
|
v)
|
|
|
|
VERBOSE=1
|
|
|
|
;;
|
|
|
|
c)
|
|
|
|
CLEANUP=1
|
|
|
|
;;
|
2010-08-26 22:44:39 +04:00
|
|
|
t)
|
|
|
|
TESTS_RUN=($OPTARG)
|
|
|
|
;;
|
|
|
|
s)
|
|
|
|
TESTS_SKIP=($OPTARG)
|
|
|
|
;;
|
2010-08-26 22:22:58 +04:00
|
|
|
?)
|
|
|
|
usage
|
|
|
|
exit
|
|
|
|
;;
|
|
|
|
esac
|
|
|
|
done
|
|
|
|
|
|
|
|
if [ $(id -u) != 0 ]; then
|
|
|
|
die "Must run as root"
|
|
|
|
fi
|
|
|
|
|
|
|
|
# Perform pre-cleanup is requested
|
|
|
|
if [ ${CLEANUP} ]; then
|
|
|
|
cleanup_loop_devices
|
|
|
|
rm -f /tmp/zpool.cache.*
|
|
|
|
fi
|
|
|
|
|
|
|
|
zconfig_partition() {
|
|
|
|
local DEVICE=$1
|
|
|
|
local START=$2
|
|
|
|
local END=$3
|
|
|
|
local TMP_FILE=`mktemp`
|
|
|
|
|
|
|
|
/sbin/sfdisk -q ${DEVICE} << EOF &>${TMP_FILE} || fail 4
|
|
|
|
${START},${END}
|
|
|
|
;
|
|
|
|
;
|
|
|
|
;
|
|
|
|
EOF
|
|
|
|
|
|
|
|
rm ${TMP_FILE}
|
|
|
|
}
|
|
|
|
|
|
|
|
# Validate persistent zpool.cache configuration.
|
2010-08-26 22:44:39 +04:00
|
|
|
test_1() {
|
2010-08-26 22:22:58 +04:00
|
|
|
local POOL_NAME=test1
|
|
|
|
local TMP_FILE1=`mktemp`
|
|
|
|
local TMP_FILE2=`mktemp`
|
|
|
|
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
|
|
|
|
|
|
|
# Create a pool save its status for comparison.
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
|
|
|
${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
|
|
|
|
|
|
|
|
# Unload/load the module stack and verify the pool persists.
|
|
|
|
${ZFS_SH} -u || fail 4
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 5
|
|
|
|
${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 6
|
|
|
|
cmp ${TMP_FILE1} ${TMP_FILE2} || fail 7
|
|
|
|
|
|
|
|
# Cleanup the test pool and temporary files
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 8
|
|
|
|
rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_CACHE} || fail 9
|
|
|
|
${ZFS_SH} -u || fail 10
|
|
|
|
|
|
|
|
pass
|
|
|
|
}
|
2010-08-26 22:44:39 +04:00
|
|
|
run_test 1 "persistent zpool.cache"
|
2010-08-26 22:22:58 +04:00
|
|
|
|
|
|
|
# Validate ZFS disk scanning and import w/out zpool.cache configuration.
|
2010-08-26 22:44:39 +04:00
|
|
|
test_2() {
|
2010-08-26 22:22:58 +04:00
|
|
|
local POOL_NAME=test2
|
|
|
|
local TMP_FILE1=`mktemp`
|
|
|
|
local TMP_FILE2=`mktemp`
|
|
|
|
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
|
|
|
|
|
|
|
# Create a pool save its status for comparison.
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
|
|
|
${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
|
|
|
|
|
|
|
|
# Unload the module stack, remove the cache file, load the module
|
|
|
|
# stack and attempt to probe the disks to import the pool. As
|
|
|
|
# a cross check verify the old pool state against the imported.
|
|
|
|
${ZFS_SH} -u || fail 4
|
|
|
|
rm -f ${TMP_CACHE} || fail 5
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 6
|
|
|
|
${ZPOOL} import | grep ${POOL_NAME} >/dev/null || fail 7
|
|
|
|
${ZPOOL} import ${POOL_NAME} || fail 8
|
|
|
|
${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 9
|
|
|
|
cmp ${TMP_FILE1} ${TMP_FILE2} || fail 10
|
|
|
|
|
|
|
|
# Cleanup the test pool and temporary files
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 11
|
|
|
|
rm -f ${TMP_FILE1} ${TMP_FILE2} || fail 12
|
|
|
|
${ZFS_SH} -u || fail 13
|
|
|
|
|
|
|
|
pass
|
|
|
|
}
|
2010-08-26 22:44:39 +04:00
|
|
|
run_test 2 "scan disks for pools to import"
|
2010-08-26 22:22:58 +04:00
|
|
|
|
|
|
|
zconfig_zvol_device_stat() {
|
|
|
|
local EXPECT=$1
|
|
|
|
local POOL_NAME=/dev/$2
|
|
|
|
local ZVOL_NAME=/dev/$3
|
|
|
|
local SNAP_NAME=/dev/$4
|
|
|
|
local CLONE_NAME=/dev/$5
|
|
|
|
local COUNT=0
|
|
|
|
|
|
|
|
# Briefly delay for udev
|
|
|
|
sleep 1
|
|
|
|
|
|
|
|
# Pool exists
|
|
|
|
stat ${POOL_NAME} &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
|
|
|
|
# Volume and partitions
|
|
|
|
stat ${ZVOL_NAME} &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
stat ${ZVOL_NAME}1 &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
stat ${ZVOL_NAME}2 &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
|
|
|
|
# Snapshot with partitions
|
|
|
|
stat ${SNAP_NAME} &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
stat ${SNAP_NAME}1 &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
stat ${SNAP_NAME}2 &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
|
|
|
|
# Clone with partitions
|
|
|
|
stat ${CLONE_NAME} &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
stat ${CLONE_NAME}1 &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
stat ${CLONE_NAME}2 &>/dev/null && let COUNT=$COUNT+1
|
|
|
|
|
|
|
|
if [ $EXPECT -ne $COUNT ]; then
|
|
|
|
return 1
|
|
|
|
fi
|
|
|
|
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
|
|
|
|
# zpool import/export device check
|
|
|
|
# (1 volume, 2 partitions, 1 snapshot, 1 clone)
|
2010-08-26 22:44:39 +04:00
|
|
|
test_3() {
|
2010-08-26 22:22:58 +04:00
|
|
|
local POOL_NAME=tank
|
|
|
|
local ZVOL_NAME=volume
|
|
|
|
local SNAP_NAME=snap
|
|
|
|
local CLONE_NAME=clone
|
|
|
|
local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
|
|
|
local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
|
|
|
|
local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
|
|
|
|
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
|
|
|
|
|
|
|
# Create a pool, volume, partition, snapshot, and clone.
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
|
|
|
${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3
|
|
|
|
zconfig_partition /dev/${FULL_ZVOL_NAME} 0 64 || fail 4
|
|
|
|
${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5
|
|
|
|
${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6
|
|
|
|
|
|
|
|
# Verify the devices were created
|
|
|
|
zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
|
|
|
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7
|
|
|
|
|
|
|
|
# Export the pool
|
|
|
|
${ZPOOL} export ${POOL_NAME} || fail 8
|
|
|
|
|
|
|
|
# verify the devices were removed
|
|
|
|
zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
|
|
|
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9
|
|
|
|
|
|
|
|
# Import the pool, wait 1 second for udev
|
|
|
|
${ZPOOL} import ${POOL_NAME} || fail 10
|
|
|
|
|
|
|
|
# Verify the devices were created
|
|
|
|
zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
|
|
|
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11
|
|
|
|
|
|
|
|
# Destroy the pool and consequently the devices
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 12
|
|
|
|
|
|
|
|
# verify the devices were removed
|
|
|
|
zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
|
|
|
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 13
|
|
|
|
|
|
|
|
${ZFS_SH} -u || fail 14
|
|
|
|
rm -f ${TMP_CACHE} || fail 15
|
|
|
|
|
|
|
|
pass
|
|
|
|
}
|
2010-08-26 22:44:39 +04:00
|
|
|
run_test 3 "zpool import/export device"
|
2010-08-26 22:22:58 +04:00
|
|
|
|
|
|
|
# zpool insmod/rmmod device check (1 volume, 1 snapshot, 1 clone)
|
2010-08-26 22:44:39 +04:00
|
|
|
test_4() {
|
2010-08-26 22:22:58 +04:00
|
|
|
POOL_NAME=tank
|
|
|
|
ZVOL_NAME=volume
|
|
|
|
SNAP_NAME=snap
|
|
|
|
CLONE_NAME=clone
|
|
|
|
FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
|
|
|
FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
|
|
|
|
FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
|
|
|
|
TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
|
|
|
|
|
|
|
# Create a pool, volume, snapshot, and clone
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
|
|
|
${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3
|
|
|
|
zconfig_partition /dev/${FULL_ZVOL_NAME} 0 64 || fail 4
|
|
|
|
${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5
|
|
|
|
${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6
|
|
|
|
|
|
|
|
# Verify the devices were created
|
|
|
|
zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
|
|
|
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7
|
|
|
|
|
|
|
|
# Unload the modules
|
|
|
|
${ZFS_SH} -u || fail 8
|
|
|
|
|
|
|
|
# Verify the devices were removed
|
|
|
|
zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
|
|
|
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9
|
|
|
|
|
|
|
|
# Load the modules, wait 1 second for udev
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 10
|
|
|
|
|
|
|
|
# Verify the devices were created
|
|
|
|
zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
|
|
|
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11
|
|
|
|
|
|
|
|
# Destroy the pool and consequently the devices
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 12
|
|
|
|
|
|
|
|
# Verify the devices were removed
|
|
|
|
zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
|
|
|
|
${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 13
|
|
|
|
|
|
|
|
${ZFS_SH} -u || fail 14
|
|
|
|
rm -f ${TMP_CACHE} || fail 15
|
|
|
|
|
|
|
|
pass
|
|
|
|
}
|
2010-08-26 22:44:39 +04:00
|
|
|
run_test 4 "zpool insmod/rmmod device"
|
2010-08-26 22:22:58 +04:00
|
|
|
|
|
|
|
# ZVOL volume sanity check
|
2010-08-26 22:44:39 +04:00
|
|
|
test_5() {
|
2010-08-26 22:22:58 +04:00
|
|
|
local POOL_NAME=tank
|
|
|
|
local ZVOL_NAME=fish
|
|
|
|
local FULL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
|
|
|
local SRC_DIR=/bin/
|
|
|
|
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
|
|
|
|
|
|
|
# Create a pool and volume.
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
|
|
|
${ZFS} create -V 400M ${FULL_NAME} || fail 3
|
|
|
|
|
|
|
|
# Partition the volume, for a 400M volume there will be
|
|
|
|
# 812 cylinders, 16 heads, and 63 sectors per track.
|
|
|
|
zconfig_partition /dev/${FULL_NAME} 0 812
|
|
|
|
|
|
|
|
# Format the partition with ext3.
|
|
|
|
/sbin/mkfs.ext3 -q /dev/${FULL_NAME}1 || fail 5
|
|
|
|
|
|
|
|
# Mount the ext3 filesystem and copy some data to it.
|
|
|
|
mkdir -p /tmp/${ZVOL_NAME}1 || fail 6
|
|
|
|
mount /dev/${FULL_NAME}1 /tmp/${ZVOL_NAME}1 || fail 7
|
|
|
|
cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}1 || fail 8
|
|
|
|
sync
|
|
|
|
|
|
|
|
# Verify the copied files match the original files.
|
|
|
|
diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}1${SRC_DIR} &>/dev/null || fail 9
|
|
|
|
|
|
|
|
# Remove the files, umount, destroy the volume and pool.
|
|
|
|
rm -Rf /tmp/${ZVOL_NAME}1${SRC_DIR}* || fail 10
|
|
|
|
umount /tmp/${ZVOL_NAME}1 || fail 11
|
|
|
|
rmdir /tmp/${ZVOL_NAME}1 || fail 12
|
|
|
|
|
|
|
|
${ZFS} destroy ${FULL_NAME} || fail 13
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 14
|
|
|
|
${ZFS_SH} -u || fail 15
|
|
|
|
rm -f ${TMP_CACHE} || fail 16
|
|
|
|
|
|
|
|
pass
|
|
|
|
}
|
2010-08-26 22:44:39 +04:00
|
|
|
run_test 5 "zvol+ext3 volume"
|
2010-08-26 22:22:58 +04:00
|
|
|
|
|
|
|
# ZVOL snapshot sanity check
|
2010-08-26 22:44:39 +04:00
|
|
|
test_6() {
|
2010-08-26 22:22:58 +04:00
|
|
|
local POOL_NAME=tank
|
|
|
|
local ZVOL_NAME=fish
|
|
|
|
local SNAP_NAME=pristine
|
|
|
|
local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
|
|
|
local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
|
|
|
|
local SRC_DIR=/bin/
|
|
|
|
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
|
|
|
|
|
|
|
# Create a pool and volume.
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
|
|
|
${ZFS} create -V 400M ${FULL_ZVOL_NAME} || fail 3
|
|
|
|
|
|
|
|
# Partition the volume, for a 400M volume there will be
|
|
|
|
# 812 cylinders, 16 heads, and 63 sectors per track.
|
|
|
|
zconfig_partition /dev/${FULL_ZVOL_NAME} 0 812
|
|
|
|
|
|
|
|
# Format the partition with ext2 (no journal).
|
|
|
|
/sbin/mkfs.ext2 -q /dev/${FULL_ZVOL_NAME}1 || fail 5
|
|
|
|
|
|
|
|
# Mount the ext3 filesystem and copy some data to it.
|
|
|
|
mkdir -p /tmp/${ZVOL_NAME}1 || fail 6
|
|
|
|
mount /dev/${FULL_ZVOL_NAME}1 /tmp/${ZVOL_NAME}1 || fail 7
|
|
|
|
|
|
|
|
# Snapshot the pristine ext2 filesystem and mount it read-only.
|
|
|
|
${ZFS} snapshot ${FULL_SNAP_NAME} && sleep 1 || fail 8
|
|
|
|
mkdir -p /tmp/${SNAP_NAME}1 || fail 9
|
|
|
|
mount /dev/${FULL_SNAP_NAME}1 /tmp/${SNAP_NAME}1 &>/dev/null || fail 10
|
|
|
|
|
|
|
|
# Copy to original volume
|
|
|
|
cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}1 || fail 11
|
|
|
|
sync
|
|
|
|
|
|
|
|
# Verify the copied files match the original files,
|
|
|
|
# and the copied files do NOT appear in the snapshot.
|
|
|
|
diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}1${SRC_DIR} &>/dev/null || fail 12
|
|
|
|
diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}1${SRC_DIR} &>/dev/null && fail 13
|
|
|
|
|
|
|
|
# umount, destroy the snapshot, volume, and pool.
|
|
|
|
umount /tmp/${SNAP_NAME}1 || fail 14
|
|
|
|
rmdir /tmp/${SNAP_NAME}1 || fail 15
|
|
|
|
${ZFS} destroy ${FULL_SNAP_NAME} || fail 16
|
|
|
|
|
|
|
|
umount /tmp/${ZVOL_NAME}1 || fail 17
|
|
|
|
rmdir /tmp/${ZVOL_NAME}1 || fail 18
|
|
|
|
${ZFS} destroy ${FULL_ZVOL_NAME} || fail 19
|
|
|
|
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 20
|
|
|
|
${ZFS_SH} -u || fail 21
|
|
|
|
rm -f ${TMP_CACHE} || fail 22
|
|
|
|
|
|
|
|
pass
|
|
|
|
}
|
2010-08-26 22:44:39 +04:00
|
|
|
run_test 6 "zvol+ext2 snapshot"
|
2010-08-26 22:22:58 +04:00
|
|
|
|
|
|
|
# ZVOL clone sanity check
|
2010-08-26 22:44:39 +04:00
|
|
|
test_7() {
|
2010-08-26 22:22:58 +04:00
|
|
|
local POOL_NAME=tank
|
|
|
|
local ZVOL_NAME=fish
|
|
|
|
local SNAP_NAME=pristine
|
|
|
|
local CLONE_NAME=clone
|
|
|
|
local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
|
|
|
local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
|
|
|
|
local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
|
|
|
|
local SRC_DIR=/bin/
|
|
|
|
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
|
|
|
|
|
|
|
# Create a pool and volume.
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
|
|
|
${ZFS} create -V 400M ${FULL_ZVOL_NAME} || fail 3
|
|
|
|
|
|
|
|
# Partition the volume, for a 400M volume there will be
|
|
|
|
# 812 cylinders, 16 heads, and 63 sectors per track.
|
|
|
|
zconfig_partition /dev/${FULL_ZVOL_NAME} 0 812
|
|
|
|
|
|
|
|
# Format the partition with ext2 (no journal).
|
|
|
|
/sbin/mkfs.ext2 -q /dev/${FULL_ZVOL_NAME}1 || fail 5
|
|
|
|
|
|
|
|
# Mount the ext3 filesystem and copy some data to it.
|
|
|
|
mkdir -p /tmp/${ZVOL_NAME}1 || fail 6
|
|
|
|
mount /dev/${FULL_ZVOL_NAME}1 /tmp/${ZVOL_NAME}1 || fail 7
|
|
|
|
|
|
|
|
# Snapshot the pristine ext2 filesystem and mount it read-only.
|
|
|
|
${ZFS} snapshot ${FULL_SNAP_NAME} && sleep 1 || fail 8
|
|
|
|
mkdir -p /tmp/${SNAP_NAME}1 || fail 9
|
|
|
|
mount /dev/${FULL_SNAP_NAME}1 /tmp/${SNAP_NAME}1 &>/dev/null || fail 10
|
|
|
|
|
|
|
|
# Copy to original volume.
|
|
|
|
cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}1 || fail 11
|
|
|
|
sync
|
|
|
|
|
|
|
|
# Verify the copied files match the original files,
|
|
|
|
# and the copied files do NOT appear in the snapshot.
|
|
|
|
diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}1${SRC_DIR} &>/dev/null || fail 12
|
|
|
|
diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}1${SRC_DIR} &>/dev/null && fail 13
|
|
|
|
|
|
|
|
# Clone from the original pristine snapshot
|
|
|
|
${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} && sleep 1 || fail 14
|
|
|
|
mkdir -p /tmp/${CLONE_NAME}1 || fail 15
|
|
|
|
mount /dev/${FULL_CLONE_NAME}1 /tmp/${CLONE_NAME}1 || fail 16
|
|
|
|
|
|
|
|
# Verify the clone matches the pristine snapshot,
|
|
|
|
# and the files copied to the original volume are NOT there.
|
|
|
|
diff -ur /tmp/${SNAP_NAME}1 /tmp/${CLONE_NAME}1 &>/dev/null || fail 17
|
|
|
|
diff -ur /tmp/${ZVOL_NAME}1 /tmp/${CLONE_NAME}1 &>/dev/null && fail 18
|
|
|
|
|
|
|
|
# Copy to cloned volume.
|
|
|
|
cp -RL ${SRC_DIR} /tmp/${CLONE_NAME}1 || fail 19
|
|
|
|
sync
|
|
|
|
|
|
|
|
# Verify the clone matches the modified original volume.
|
|
|
|
diff -ur /tmp/${ZVOL_NAME}1 /tmp/${CLONE_NAME}1 &>/dev/null || fail 20
|
|
|
|
|
|
|
|
# umount, destroy the snapshot, volume, and pool.
|
|
|
|
umount /tmp/${CLONE_NAME}1 || fail 21
|
|
|
|
rmdir /tmp/${CLONE_NAME}1 || fail 22
|
|
|
|
${ZFS} destroy ${FULL_CLONE_NAME} || fail 23
|
|
|
|
|
|
|
|
umount /tmp/${SNAP_NAME}1 || fail 24
|
|
|
|
rmdir /tmp/${SNAP_NAME}1 || fail 25
|
|
|
|
${ZFS} destroy ${FULL_SNAP_NAME} || fail 26
|
|
|
|
|
|
|
|
umount /tmp/${ZVOL_NAME}1 || fail 27
|
|
|
|
rmdir /tmp/${ZVOL_NAME}1 || fail 28
|
|
|
|
${ZFS} destroy ${FULL_ZVOL_NAME} || fail 29
|
|
|
|
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 30
|
|
|
|
${ZFS_SH} -u || fail 31
|
|
|
|
rm -f ${TMP_CACHE} || fail 32
|
|
|
|
|
|
|
|
pass
|
|
|
|
}
|
2010-08-26 22:44:39 +04:00
|
|
|
run_test 7 "zvol+ext2 clone"
|
2010-08-26 22:22:58 +04:00
|
|
|
|
|
|
|
# Send/Receive sanity check
|
|
|
|
test_8() {
|
|
|
|
local POOL_NAME1=tank1
|
|
|
|
local POOL_NAME2=tank2
|
|
|
|
local ZVOL_NAME=fish
|
|
|
|
local SNAP_NAME=snap
|
|
|
|
local FULL_ZVOL_NAME1=${POOL_NAME1}/${ZVOL_NAME}
|
|
|
|
local FULL_ZVOL_NAME2=${POOL_NAME2}/${ZVOL_NAME}
|
|
|
|
local FULL_SNAP_NAME1=${POOL_NAME1}/${ZVOL_NAME}@${SNAP_NAME}
|
|
|
|
local FULL_SNAP_NAME2=${POOL_NAME2}/${ZVOL_NAME}@${SNAP_NAME}
|
|
|
|
local SRC_DIR=/bin/
|
|
|
|
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
|
|
|
|
|
|
|
# Create two pools and a volume
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 || fail 2
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 || fail 3
|
|
|
|
${ZFS} create -V 400M ${FULL_ZVOL_NAME1} || fail 4
|
|
|
|
|
|
|
|
# Partition the volume, for a 400M volume there will be
|
|
|
|
# 812 cylinders, 16 heads, and 63 sectors per track.
|
|
|
|
zconfig_partition /dev/${FULL_ZVOL_NAME1} 0 812
|
|
|
|
|
|
|
|
# Format the partition with ext2.
|
|
|
|
/sbin/mkfs.ext2 -q /dev/${FULL_ZVOL_NAME1}1 || fail 5
|
|
|
|
|
|
|
|
# Mount the ext3 filesystem and copy some data to it.
|
|
|
|
mkdir -p /tmp/${FULL_ZVOL_NAME1}1 || fail 6
|
|
|
|
mount /dev/${FULL_ZVOL_NAME1}1 /tmp/${FULL_ZVOL_NAME1}1 || fail 7
|
|
|
|
cp -RL ${SRC_DIR} /tmp/${FULL_ZVOL_NAME1}1 || fail 8
|
|
|
|
sync || fail 9
|
|
|
|
|
|
|
|
# Snapshot the ext3 filesystem so it may be sent.
|
|
|
|
${ZFS} snapshot ${FULL_SNAP_NAME1} && sleep 1 || fail 11
|
|
|
|
|
|
|
|
# Send/receive the snapshot from POOL_NAME1 to POOL_NAME2
|
|
|
|
(${ZFS} send ${FULL_SNAP_NAME1} | \
|
|
|
|
${ZFS} receive ${FULL_ZVOL_NAME2}) && sleep 1 || fail 12
|
|
|
|
|
|
|
|
# Mount the sent ext3 filesystem.
|
|
|
|
mkdir -p /tmp/${FULL_ZVOL_NAME2}1 || fail 13
|
|
|
|
mount /dev/${FULL_ZVOL_NAME2}1 /tmp/${FULL_ZVOL_NAME2}1 || fail 14
|
|
|
|
|
|
|
|
# Verify the contents of the volumes match
|
|
|
|
diff -ur /tmp/${FULL_ZVOL_NAME1}1 /tmp/${FULL_ZVOL_NAME2}1 \
|
|
|
|
&>/dev/null || fail 15
|
|
|
|
|
|
|
|
# Umount, destroy the volume and pool.
|
|
|
|
umount /tmp/${FULL_ZVOL_NAME1}1 || fail 16
|
|
|
|
umount /tmp/${FULL_ZVOL_NAME2}1 || fail 17
|
|
|
|
rmdir /tmp/${FULL_ZVOL_NAME1}1 || fail 18
|
|
|
|
rmdir /tmp/${FULL_ZVOL_NAME2}1 || fail 19
|
|
|
|
rmdir /tmp/${POOL_NAME1} || fail 20
|
|
|
|
rmdir /tmp/${POOL_NAME2} || fail 21
|
|
|
|
|
|
|
|
${ZFS} destroy ${FULL_SNAP_NAME1} || fail 22
|
|
|
|
${ZFS} destroy ${FULL_SNAP_NAME2} || fail 23
|
|
|
|
${ZFS} destroy ${FULL_ZVOL_NAME1} || fail 24
|
|
|
|
${ZFS} destroy ${FULL_ZVOL_NAME2} || fail 25
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 -d || fail 26
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 -d || fail 27
|
|
|
|
${ZFS_SH} -u || fail 28
|
|
|
|
rm -f ${TMP_CACHE} || fail 29
|
|
|
|
|
|
|
|
pass
|
|
|
|
}
|
|
|
|
run_test 8 "zfs send/receive"
|
|
|
|
|
|
|
|
# zpool event sanity check
|
|
|
|
test_9() {
|
|
|
|
local POOL_NAME=tank
|
|
|
|
local ZVOL_NAME=fish
|
|
|
|
local FULL_NAME=${POOL_NAME}/${ZVOL_NAME}
|
|
|
|
local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
|
|
|
|
local TMP_EVENTS=`mktemp -p /tmp zpool.events.XXXXXXXX`
|
|
|
|
|
|
|
|
# Create a pool and volume.
|
|
|
|
${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
|
|
|
|
${ZFS} create -V 400M ${FULL_NAME} || fail 3
|
|
|
|
|
|
|
|
# Dump the events, there should be at least 5 lines.
|
|
|
|
${ZPOOL} events >${TMP_EVENTS} || fail 4
|
|
|
|
EVENTS=`wc -l ${TMP_EVENTS} | cut -f1 -d' '`
|
|
|
|
[ $EVENTS -lt 5 ] && fail 5
|
|
|
|
|
|
|
|
# Clear the events and ensure there are none.
|
|
|
|
${ZPOOL} events -c >/dev/null || fail 6
|
|
|
|
${ZPOOL} events >${TMP_EVENTS} || fail 7
|
|
|
|
EVENTS=`wc -l ${TMP_EVENTS} | cut -f1 -d' '`
|
|
|
|
[ $EVENTS -gt 1 ] && fail 8
|
|
|
|
|
|
|
|
${ZFS} destroy ${FULL_NAME} || fail 9
|
|
|
|
${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 10
|
|
|
|
${ZFS_SH} -u || fail 11
|
|
|
|
rm -f ${TMP_CACHE} || fail 12
|
|
|
|
rm -f ${TMP_EVENTS} || fail 13
|
|
|
|
|
|
|
|
pass
|
|
|
|
}
|
|
|
|
run_test 9 "zpool events"
|
|
|
|
|
|
|
|
exit 0
|
|
|
|
|