]> granicus.if.org Git - zfs/commitdiff
Retire zconfig.sh
authorBrian Behlendorf <behlendorf1@llnl.gov>
Fri, 19 May 2017 17:08:23 +0000 (13:08 -0400)
committerBrian Behlendorf <behlendorf1@llnl.gov>
Mon, 22 May 2017 16:34:42 +0000 (12:34 -0400)
All of the test coverage provided by this script is now handled
as part of the ZFS Test Suite.  Remove it.

Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #6128

TEST
scripts/Makefile.am
scripts/zconfig.sh [deleted file]

diff --git a/TEST b/TEST
index 7a909a098bbfe0d011ef4a87f8609f5d34ebbc57..fbc47a4e286912223bfea9169e6a135841784b60 100644 (file)
--- a/TEST
+++ b/TEST
 #TEST_ZTEST_DIR="/var/tmp/"
 #TEST_ZTEST_OPTIONS="-V"
 
-### zconfig
-#TEST_ZCONFIG_SKIP="yes"
-TEST_ZCONFIG_OPTIONS="-c -s10"
-
 ### zimport
 #TEST_ZIMPORT_SKIP="yes"
 #TEST_ZIMPORT_DIR="/var/tmp/zimport"
index c8218db114869d5a01cb88070b4d7c540bf47e8c..74b8b31a50d9de1c9673b2de625d4c7689bc6978 100644 (file)
@@ -5,7 +5,6 @@ EXTRA_DIST = dkms.mkconf dkms.postbuild kmodtool zfs2zol-patch.sed cstyle.pl
 pkgdatadir = $(datadir)/@PACKAGE@
 dist_pkgdata_SCRIPTS = \
        $(top_builddir)/scripts/common.sh \
-       $(top_srcdir)/scripts/zconfig.sh \
        $(top_srcdir)/scripts/zimport.sh \
        $(top_srcdir)/scripts/zfs.sh \
        $(top_srcdir)/scripts/zfs-tests.sh \
diff --git a/scripts/zconfig.sh b/scripts/zconfig.sh
deleted file mode 100755 (executable)
index e78c80d..0000000
+++ /dev/null
@@ -1,696 +0,0 @@
-#!/bin/bash
-#
-# ZFS/ZPOOL configuration test script.
-
-basedir="$(dirname $0)"
-
-SCRIPT_COMMON=common.sh
-if [ -f "${basedir}/${SCRIPT_COMMON}" ]; then
-. "${basedir}/${SCRIPT_COMMON}"
-else
-echo "Missing helper script ${SCRIPT_COMMON}" && exit 1
-fi
-
-PROG=zconfig.sh
-
-usage() {
-cat << EOF
-USAGE:
-$0 [hvcts]
-
-DESCRIPTION:
-       ZFS/ZPOOL configuration tests
-
-OPTIONS:
-       -h      Show this message
-       -v      Verbose
-       -c      Cleanup lo+file devices at start
-       -t <#>  Run listed tests
-       -s <#>  Skip listed tests
-
-EOF
-}
-
-while getopts 'hvct:s:?' OPTION; do
-       case $OPTION in
-       h)
-               usage
-               exit 1
-               ;;
-       v)
-               VERBOSE=1
-               ;;
-       c)
-               CLEANUP=1
-               ;;
-       t)
-               TESTS_RUN=($OPTARG)
-               ;;
-       s)
-               TESTS_SKIP=($OPTARG)
-               ;;
-       ?)
-               usage
-               exit
-               ;;
-       esac
-done
-
-if [ $(id -u) != 0 ]; then
-       die "Must run as root"
-fi
-
-# Initialize the test suite
-init
-
-# Disable the udev rule 90-zfs.rules to prevent the zfs module
-# stack from being loaded due to the detection of a zfs device.
-# This is important because this test scripts require full control
-# over when and how the modules are loaded/unloaded.  A trap is
-# set to ensure the udev rule is correctly replaced on exit.
-RULE=${udevruledir}/90-zfs.rules
-if test -e  ${RULE}; then
-       trap "mv ${RULE}.disabled ${RULE}" INT TERM EXIT
-       mv ${RULE} ${RULE}.disabled
-fi
-
-# Perform pre-cleanup is requested
-if [ ${CLEANUP} ]; then
-       ${ZFS_SH} -u
-       cleanup_md_devices
-       cleanup_loop_devices
-       rm -f /tmp/zpool.cache.*
-fi
-
-# Check if we need to skip the tests that require scsi_debug and lsscsi.
-SCSI_DEBUG=0
-${INFOMOD} scsi_debug &>/dev/null && SCSI_DEBUG=1
-HAVE_LSSCSI=0
-test -f ${LSSCSI} && HAVE_LSSCSI=1
-if [ ${SCSI_DEBUG} -eq 0 ] || [ ${HAVE_LSSCSI} -eq 0 ]; then
-       echo "Skipping test 10 which requires the scsi_debug " \
-               "module and the ${LSSCSI} utility"
-fi
-
-# Validate persistent zpool.cache configuration.
-test_1() {
-       local POOL_NAME=test1
-       local TMP_FILE1=`mktemp`
-       local TMP_FILE2=`mktemp`
-       local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-
-       # Create a pool save its status for comparison.
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
-
-       # Unload/load the module stack and verify the pool persists.
-       ${ZFS_SH} -u || fail 4
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 5
-       ${ZPOOL} import -c ${TMP_CACHE} ${POOL_NAME} || fail 5
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 6
-       cmp ${TMP_FILE1} ${TMP_FILE2} || fail 7
-
-       # Cleanup the test pool and temporary files
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 8
-       rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_CACHE} || fail 9
-       ${ZFS_SH} -u || fail 10
-
-       pass
-}
-run_test 1 "persistent zpool.cache"
-
-# Validate ZFS disk scanning and import w/out zpool.cache configuration.
-test_2() {
-       local POOL_NAME=test2
-       local TMP_FILE1=`mktemp`
-       local TMP_FILE2=`mktemp`
-       local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-
-       # Create a pool save its status for comparison.
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
-
-       # Unload the module stack, remove the cache file, load the module
-       # stack and attempt to probe the disks to import the pool.  As
-       # a cross check verify the old pool state against the imported.
-       ${ZFS_SH} -u || fail 4
-       rm -f ${TMP_CACHE} || fail 5
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 6
-       ${ZPOOL} import -d /dev ${POOL_NAME} || fail 8
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 9
-       cmp ${TMP_FILE1} ${TMP_FILE2} || fail 10
-
-       # Cleanup the test pool and temporary files
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 11
-       rm -f ${TMP_FILE1} ${TMP_FILE2} || fail 12
-       ${ZFS_SH} -u || fail 13
-
-       pass
-}
-run_test 2 "scan disks for pools to import"
-
-zconfig_zvol_device_stat() {
-       local EXPECT=$1
-       local POOL_NAME=/dev/zvol/$2
-       local ZVOL_NAME=/dev/zvol/$3
-       local SNAP_NAME=/dev/zvol/$4
-       local CLONE_NAME=/dev/zvol/$5
-       local COUNT=0
-
-       # Briefly delay for udev
-       udev_trigger
-
-       # Pool exists
-       stat ${POOL_NAME} &>/dev/null   && let COUNT=$COUNT+1
-
-       # Volume and partitions
-       stat ${ZVOL_NAME}  &>/dev/null  && let COUNT=$COUNT+1
-       stat ${ZVOL_NAME}-part1 &>/dev/null  && let COUNT=$COUNT+1
-       stat ${ZVOL_NAME}-part2 &>/dev/null  && let COUNT=$COUNT+1
-
-       # Snapshot with partitions
-       stat ${SNAP_NAME}  &>/dev/null  && let COUNT=$COUNT+1
-       stat ${SNAP_NAME}-part1 &>/dev/null  && let COUNT=$COUNT+1
-       stat ${SNAP_NAME}-part2 &>/dev/null  && let COUNT=$COUNT+1
-
-       # Clone with partitions
-       stat ${CLONE_NAME}  &>/dev/null && let COUNT=$COUNT+1
-       stat ${CLONE_NAME}-part1 &>/dev/null && let COUNT=$COUNT+1
-       stat ${CLONE_NAME}-part2 &>/dev/null && let COUNT=$COUNT+1
-
-       if [ $EXPECT -ne $COUNT ]; then
-               return 1
-       fi
-
-       return 0
-}
-
-# zpool import/export device check
-# (1 volume, 2 partitions, 1 snapshot, 1 clone)
-test_3() {
-       local POOL_NAME=tank
-       local ZVOL_NAME=volume
-       local SNAP_NAME=snap
-       local CLONE_NAME=clone
-       local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
-       local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
-       local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
-       local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-
-       # Create a pool, volume, partition, snapshot, and clone.
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
-       ${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3
-       ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
-       label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
-       partition /dev/zvol/${FULL_ZVOL_NAME} primary 1% 50% || fail 4
-       partition /dev/zvol/${FULL_ZVOL_NAME} primary 51% -1 || fail 4
-       ${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5
-       ${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6
-
-       # Verify the devices were created
-       zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7
-
-       # Export the pool
-       ${ZPOOL} export ${POOL_NAME} || fail 8
-
-       # verify the devices were removed
-       zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9
-
-       # Import the pool, wait 1 second for udev
-       ${ZPOOL} import ${POOL_NAME} || fail 10
-
-       # Verify the devices were created
-       zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11
-
-       # Toggle the snapdev and observe snapshot device links toggled
-       ${ZFS} set snapdev=hidden ${FULL_ZVOL_NAME} || fail 12
-       
-       zconfig_zvol_device_stat 7 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           "invalid" ${FULL_CLONE_NAME} || fail 13
-
-       ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 14
-
-       zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 15
-
-       # Destroy the pool and consequently the devices
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 16
-
-       # verify the devices were removed
-       zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 17
-
-       ${ZFS_SH} -u || fail 18
-       rm -f ${TMP_CACHE} || fail 19
-
-       pass
-}
-run_test 3 "zpool import/export device"
-
-# zpool insmod/rmmod device check (1 volume, 1 snapshot, 1 clone)
-test_4() {
-       POOL_NAME=tank
-       ZVOL_NAME=volume
-       SNAP_NAME=snap
-       CLONE_NAME=clone
-       FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
-       FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
-       FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
-       TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-
-       # Create a pool, volume, snapshot, and clone
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
-       ${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3
-       ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
-       label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
-       partition /dev/zvol/${FULL_ZVOL_NAME} primary 1% 50% || fail 4
-       partition /dev/zvol/${FULL_ZVOL_NAME} primary 51% -1 || fail 4
-       ${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5
-       ${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6
-
-       # Verify the devices were created
-       zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7
-
-       # Unload the modules
-       ${ZFS_SH} -u || fail 8
-
-       # Verify the devices were removed
-       zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9
-
-       # Load the modules, list the pools to ensure they are opened
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 10
-       ${ZPOOL} import -c ${TMP_CACHE} ${POOL_NAME} || fail 10
-       ${ZPOOL} list &>/dev/null
-
-       # Verify the devices were created
-       zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11
-
-       # Destroy the pool and consequently the devices
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 12
-
-       # Verify the devices were removed
-       zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
-           ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 13
-
-       ${ZFS_SH} -u || fail 14
-       rm -f ${TMP_CACHE} || fail 15
-
-       pass
-}
-run_test 4 "zpool insmod/rmmod device"
-
-# ZVOL volume sanity check
-test_5() {
-       local POOL_NAME=tank
-       local ZVOL_NAME=fish
-       local FULL_NAME=${POOL_NAME}/${ZVOL_NAME}
-       local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-
-       # Create a pool and volume.
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raid0 || fail 2
-       ${ZFS} create -V 800M ${FULL_NAME} || fail 3
-       label /dev/zvol/${FULL_NAME} msdos || fail 4
-       partition /dev/zvol/${FULL_NAME} primary 1 -1 || fail 4
-       format /dev/zvol/${FULL_NAME}-part1 ext2 || fail 5
-
-       # Mount the ext2 filesystem and copy some data to it.
-       mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 6
-       mount /dev/zvol/${FULL_NAME}-part1 /tmp/${ZVOL_NAME}-part1 || fail 7
-       cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 8
-       sync
-
-       # Verify the copied files match the original files.
-       diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \
-               &>/dev/null || fail 9
-
-       # Remove the files, umount, destroy the volume and pool.
-       rm -Rf /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} || fail 10
-       umount /tmp/${ZVOL_NAME}-part1 || fail 11
-       rmdir /tmp/${ZVOL_NAME}-part1 || fail 12
-
-       ${ZFS} destroy ${FULL_NAME} || fail 13
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 14
-       ${ZFS_SH} -u || fail 15
-       rm -f ${TMP_CACHE} || fail 16
-
-       pass
-}
-run_test 5 "zvol+ext2 volume"
-
-# ZVOL snapshot sanity check
-test_6() {
-       local POOL_NAME=tank
-       local ZVOL_NAME=fish
-       local SNAP_NAME=pristine
-       local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
-       local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
-       local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-
-       # Create a pool and volume.
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raid0 || fail 2
-       ${ZFS} create -s -V 800M ${FULL_ZVOL_NAME} || fail 3
-       ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
-       label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
-       partition /dev/zvol/${FULL_ZVOL_NAME} primary 1 -1 || fail 4
-       format /dev/zvol/${FULL_ZVOL_NAME}-part1 ext2 || fail 5
-
-       # Mount the ext2 filesystem and copy some data to it.
-       mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 6
-       mount /dev/zvol/${FULL_ZVOL_NAME}-part1 /tmp/${ZVOL_NAME}-part1 \
-               || fail 7
-
-       # Snapshot the pristine ext2 filesystem and mount it read-only.
-       ${ZFS} snapshot ${FULL_SNAP_NAME} || fail 8
-       wait_udev /dev/zvol/${FULL_SNAP_NAME}-part1 30 || fail 8
-       mkdir -p /tmp/${SNAP_NAME}-part1 || fail 9
-       mount /dev/zvol/${FULL_SNAP_NAME}-part1 /tmp/${SNAP_NAME}-part1 \
-               &>/dev/null || fail 10
-
-       # Copy to original volume
-       cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 11
-       sync
-
-       # Verify the copied files match the original files,
-       # and the copied files do NOT appear in the snapshot.
-       diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \
-               &>/dev/null || fail 12
-       diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}-part1/${SRC_DIR##*/} \
-               &>/dev/null && fail 13
-
-       # umount, destroy the snapshot, volume, and pool.
-       umount /tmp/${SNAP_NAME}-part1 || fail 14
-       rmdir /tmp/${SNAP_NAME}-part1 || fail 15
-       ${ZFS} destroy ${FULL_SNAP_NAME} || fail 16
-
-       umount /tmp/${ZVOL_NAME}-part1 || fail 17
-       rmdir /tmp/${ZVOL_NAME}-part1 || fail 18
-       ${ZFS} destroy ${FULL_ZVOL_NAME} || fail 19
-
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 20
-       ${ZFS_SH} -u || fail 21
-       rm -f ${TMP_CACHE} || fail 22
-
-       pass
-}
-run_test 6 "zvol+ext2 snapshot"
-
-# ZVOL clone sanity check
-test_7() {
-       local POOL_NAME=tank
-       local ZVOL_NAME=fish
-       local SNAP_NAME=pristine
-       local CLONE_NAME=clone
-       local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
-       local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
-       local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
-       local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-
-       # Create a pool and volume.
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
-       ${ZFS} create -V 300M ${FULL_ZVOL_NAME} || fail 3
-       ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
-       label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
-       partition /dev/zvol/${FULL_ZVOL_NAME} primary 1 -1 || fail 4
-       format /dev/zvol/${FULL_ZVOL_NAME}-part1 ext2 || fail 5
-
-       # Snapshot the pristine ext2 filesystem.
-       ${ZFS} snapshot ${FULL_SNAP_NAME} || fail 6
-       wait_udev /dev/zvol/${FULL_SNAP_NAME}-part1 30 || fail 7
-
-       # Mount the ext2 filesystem so some data can be copied to it.
-       mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 7
-       mount /dev/zvol/${FULL_ZVOL_NAME}-part1 \
-               /tmp/${ZVOL_NAME}-part1 || fail 8
-
-       # Mount the pristine ext2 snapshot.
-       mkdir -p /tmp/${SNAP_NAME}-part1 || fail 9
-       mount /dev/zvol/${FULL_SNAP_NAME}-part1 \
-               /tmp/${SNAP_NAME}-part1 &>/dev/null || fail 10
-
-       # Copy to original volume.
-       cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 11
-       sync
-
-       # Verify the copied files match the original files,
-       # and the copied files do NOT appear in the snapshot.
-       diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \
-               &>/dev/null || fail 12
-       diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}-part1/${SRC_DIR##*/} \
-               &>/dev/null && fail 13
-
-       # Clone from the original pristine snapshot
-       ${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 14
-       wait_udev /dev/zvol/${FULL_CLONE_NAME}-part1 30 || fail 14
-       mkdir -p /tmp/${CLONE_NAME}-part1 || fail 15
-       mount /dev/zvol/${FULL_CLONE_NAME}-part1 \
-               /tmp/${CLONE_NAME}-part1 || fail 16
-
-       # Verify the clone matches the pristine snapshot,
-       # and the files copied to the original volume are NOT there.
-       diff -ur /tmp/${SNAP_NAME}-part1 /tmp/${CLONE_NAME}-part1 \
-               &>/dev/null || fail 17
-       diff -ur /tmp/${ZVOL_NAME}-part1 /tmp/${CLONE_NAME}-part1 \
-               &>/dev/null && fail 18
-
-       # Copy to cloned volume.
-       cp -RL ${SRC_DIR} /tmp/${CLONE_NAME}-part1 || fail 19
-       sync
-
-       # Verify the clone matches the modified original volume.
-       diff -ur /tmp/${ZVOL_NAME}-part1 /tmp/${CLONE_NAME}-part1 \
-               &>/dev/null || fail 20
-
-       # umount, destroy the snapshot, volume, and pool.
-       umount /tmp/${CLONE_NAME}-part1 || fail 21
-       rmdir /tmp/${CLONE_NAME}-part1 || fail 22
-       ${ZFS} destroy ${FULL_CLONE_NAME} || fail 23
-
-       umount /tmp/${SNAP_NAME}-part1 || fail 24
-       rmdir /tmp/${SNAP_NAME}-part1 || fail 25
-       ${ZFS} destroy ${FULL_SNAP_NAME} || fail 26
-
-       umount /tmp/${ZVOL_NAME}-part1 || fail 27
-       rmdir /tmp/${ZVOL_NAME}-part1 || fail 28
-       ${ZFS} destroy ${FULL_ZVOL_NAME} || fail 29
-
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 30
-       ${ZFS_SH} -u || fail 31
-       rm -f ${TMP_CACHE} || fail 32
-
-       pass
-}
-run_test 7 "zvol+ext2 clone"
-
-# Send/Receive sanity check
-test_8() {
-       local POOL_NAME1=tank1
-       local POOL_NAME2=tank2
-       local ZVOL_NAME=fish
-       local SNAP_NAME=snap
-       local FULL_ZVOL_NAME1=${POOL_NAME1}/${ZVOL_NAME}
-       local FULL_ZVOL_NAME2=${POOL_NAME2}/${ZVOL_NAME}
-       local FULL_SNAP_NAME1=${POOL_NAME1}/${ZVOL_NAME}@${SNAP_NAME}
-       local FULL_SNAP_NAME2=${POOL_NAME2}/${ZVOL_NAME}@${SNAP_NAME}
-       local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-
-       # Create two pools and a volume
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 || fail 2
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 || fail 2
-       ${ZFS} create -V 300M ${FULL_ZVOL_NAME1} || fail 3
-       ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME1} || fail 3
-       label /dev/zvol/${FULL_ZVOL_NAME1} msdos || fail 4
-       partition /dev/zvol/${FULL_ZVOL_NAME1} primary 1 -1 || fail 4
-       format /dev/zvol/${FULL_ZVOL_NAME1}-part1 ext2 || fail 5
-
-       # Mount the ext2 filesystem and copy some data to it.
-       mkdir -p /tmp/${FULL_ZVOL_NAME1}-part1 || fail 6
-       mount /dev/zvol/${FULL_ZVOL_NAME1}-part1 \
-               /tmp/${FULL_ZVOL_NAME1}-part1 || fail 7
-       cp -RL ${SRC_DIR} /tmp/${FULL_ZVOL_NAME1}-part1 || fail 8
-
-       # Unmount, snapshot, mount the ext2 filesystem so it may be sent.
-       # We only unmount to ensure the ext2 filesystem is clean.
-       umount /tmp/${FULL_ZVOL_NAME1}-part1 || fail 9
-       ${ZFS} snapshot ${FULL_SNAP_NAME1} || fail 10
-       wait_udev /dev/zvol/${FULL_SNAP_NAME1} 30 || fail 10
-       mount /dev/zvol/${FULL_ZVOL_NAME1}-part1 \
-               /tmp/${FULL_ZVOL_NAME1}-part1 || 11
-
-       # Send/receive the snapshot from POOL_NAME1 to POOL_NAME2
-       (${ZFS} send ${FULL_SNAP_NAME1} | \
-       ${ZFS} receive ${FULL_ZVOL_NAME2}) || fail 12
-       wait_udev /dev/zvol/${FULL_ZVOL_NAME2}-part1 30 || fail 12
-
-       # Mount the sent ext2 filesystem.
-       mkdir -p /tmp/${FULL_ZVOL_NAME2}-part1 || fail 13
-       mount /dev/zvol/${FULL_ZVOL_NAME2}-part1 \
-               /tmp/${FULL_ZVOL_NAME2}-part1 || fail 14
-
-       # Verify the contents of the volumes match
-       diff -ur /tmp/${FULL_ZVOL_NAME1}-part1 /tmp/${FULL_ZVOL_NAME2}-part1 \
-           &>/dev/null || fail 15
-
-       # Umount, destroy the volume and pool.
-       umount /tmp/${FULL_ZVOL_NAME1}-part1 || fail 16
-       umount /tmp/${FULL_ZVOL_NAME2}-part1 || fail 17
-       rmdir /tmp/${FULL_ZVOL_NAME1}-part1 || fail 18
-       rmdir /tmp/${FULL_ZVOL_NAME2}-part1 || fail 19
-       rmdir /tmp/${POOL_NAME1} || fail 20
-       rmdir /tmp/${POOL_NAME2} || fail 21
-
-       ${ZFS} destroy ${FULL_SNAP_NAME1} || fail 22
-       ${ZFS} destroy ${FULL_SNAP_NAME2} || fail 23
-       ${ZFS} destroy ${FULL_ZVOL_NAME1} || fail 24
-       ${ZFS} destroy ${FULL_ZVOL_NAME2} || fail 25
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 -d || fail 26
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 -d || fail 27
-       ${ZFS_SH} -u || fail 28
-       rm -f ${TMP_CACHE} || fail 29
-
-       pass
-}
-run_test 8 "zfs send/receive"
-
-# zpool event sanity check
-test_9() {
-       local POOL_NAME=tank
-       local ZVOL_NAME=fish
-       local FULL_NAME=${POOL_NAME}/${ZVOL_NAME}
-       local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-       local TMP_EVENTS=`mktemp -p /tmp zpool.events.XXXXXXXX`
-
-       # Create a pool and volume.
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
-       ${ZFS} create -V 300M ${FULL_NAME} || fail 3
-       udev_trigger
-
-       # Dump the events, there should be a pool create event
-       ${ZPOOL} events >${TMP_EVENTS} || fail 4
-       MATCHES=`grep -c sysevent\.fs\.zfs\.pool_create ${TMP_EVENTS}`
-       [ $MATCHES -eq 1 ] || fail 5
-
-       # Clear the events and ensure there are none.
-       ${ZPOOL} events -c >/dev/null || fail 6
-       ${ZPOOL} events >${TMP_EVENTS} || fail 7
-       EVENTS=`wc -l ${TMP_EVENTS} | cut -f1 -d' '`
-       [ $EVENTS -gt 1 ] && fail 8
-
-       ${ZFS} destroy ${FULL_NAME} || fail 9
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 10
-       ${ZFS_SH} -u || fail 11
-       rm -f ${TMP_CACHE} || fail 12
-       rm -f ${TMP_EVENTS} || fail 13
-
-       pass
-}
-run_test 9 "zpool events"
-
-zconfig_add_vdev() {
-       local POOL_NAME=$1
-       local TYPE=$2
-       local DEVICE=$3
-       local TMP_FILE1=`mktemp`
-       local TMP_FILE2=`mktemp`
-       local TMP_FILE3=`mktemp`
-
-       BASE_DEVICE=`basename ${DEVICE}`
-
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE1}
-       ${ZPOOL} add -f ${POOL_NAME} ${TYPE} ${DEVICE} 2>/dev/null || return 1
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2}
-       diff ${TMP_FILE1} ${TMP_FILE2} > ${TMP_FILE3}
-
-       [ `wc -l ${TMP_FILE3}|${AWK} '{print $1}'` -eq 3 ] || return 1
-
-       PARENT_VDEV=`tail -2 ${TMP_FILE3} | head -1 | ${AWK} '{print $NF}'`
-       case $TYPE in
-       cache)
-               [ "${PARENT_VDEV}" = "${TYPE}" ] || return 1
-               ;;
-       log)
-               [ "${PARENT_VDEV}" = "logs" ] || return 1
-               ;;
-       esac
-
-       if ! tail -1 ${TMP_FILE3} |
-           egrep -q "^>[[:space:]]+${BASE_DEVICE}[[:space:]]+ONLINE" ; then
-               return 1
-       fi
-       rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_FILE3}
-
-       return 0
-}
-
-# zpool add and remove sanity check
-test_10() {
-       local POOL_NAME=tank
-       local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
-       local TMP_FILE1=`mktemp`
-       local TMP_FILE2=`mktemp`
-
-       if [ ${SCSI_DEBUG} -eq 0 ] || [ ${HAVE_LSSCSI} -eq 0 ] ; then
-               skip
-               return
-       fi
-
-       test `${LSMOD} | grep -c scsi_debug` -gt 0 && \
-               (${RMMOD} scsi_debug || exit 1)
-
-       /sbin/modprobe scsi_debug dev_size_mb=128 ||
-               die "Error $? creating scsi_debug device"
-       udev_trigger
-
-       SDDEVICE=`${LSSCSI}|${AWK} '/scsi_debug/ { print $6; exit }'`
-       BASE_SDDEVICE=`basename $SDDEVICE`
-
-       # Create a pool
-       ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
-
-       # Add and remove a cache vdev by full path
-       zconfig_add_vdev ${POOL_NAME} cache ${SDDEVICE} || fail 4
-       ${ZPOOL} remove ${POOL_NAME} ${SDDEVICE} || fail 5
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 6
-       cmp ${TMP_FILE1} ${TMP_FILE2} || fail 7
-       sleep 1
-
-       # Add and remove a cache vdev by shorthand path
-       zconfig_add_vdev ${POOL_NAME} cache ${BASE_SDDEVICE} || fail 8
-       ${ZPOOL} remove ${POOL_NAME} ${BASE_SDDEVICE} || fail 9
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 10
-       cmp ${TMP_FILE1} ${TMP_FILE2} || fail 11
-       sleep 1
-
-       # Add and remove a log vdev
-       zconfig_add_vdev ${POOL_NAME} log ${BASE_SDDEVICE} || fail 12
-       ${ZPOOL} remove ${POOL_NAME} ${BASE_SDDEVICE} || fail 13
-       ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 14
-       cmp ${TMP_FILE1} ${TMP_FILE2} || fail 15
-
-       ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 16
-       ${ZFS_SH} -u || fail 17
-       ${RMMOD} scsi_debug || fail 18
-
-       rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_CACHE} || fail 19
-
-       pass
-}
-run_test 10 "zpool add/remove vdev"
-
-exit 0