mirror of
				https://git.proxmox.com/git/mirror_zfs
				synced 2025-11-04 01:41:27 +00:00 
			
		
		
		
	Authored by: Matthew Ahrens <mahrens@delphix.com> Reviewed by: George Wilson <george.wilson@delphix.com> Reviewed by: Paul Dagnelie <pcd@delphix.com> Reviewed by: Dan McDonald <danmcd@omniti.com> Approved by: Robert Mustacchi <rm@joyent.com> Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov> Ported-by: George Melikov <mail@gmelikov.ru> OpenZFS-issue: https://www.illumos.org/issues/7104 OpenZFS-commit: https://github.com/openzfs/openzfs/commit/4b5c8e9 Closes #5679
		
			
				
	
	
		
			697 lines
		
	
	
		
			21 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			697 lines
		
	
	
		
			21 KiB
		
	
	
	
		
			Bash
		
	
	
		
			Executable File
		
	
	
	
	
#!/bin/bash
 | 
						|
#
 | 
						|
# ZFS/ZPOOL configuration test script.
 | 
						|
 | 
						|
basedir="$(dirname $0)"
 | 
						|
 | 
						|
SCRIPT_COMMON=common.sh
 | 
						|
if [ -f "${basedir}/${SCRIPT_COMMON}" ]; then
 | 
						|
. "${basedir}/${SCRIPT_COMMON}"
 | 
						|
else
 | 
						|
echo "Missing helper script ${SCRIPT_COMMON}" && exit 1
 | 
						|
fi
 | 
						|
 | 
						|
PROG=zconfig.sh
 | 
						|
 | 
						|
usage() {
 | 
						|
cat << EOF
 | 
						|
USAGE:
 | 
						|
$0 [hvcts]
 | 
						|
 | 
						|
DESCRIPTION:
 | 
						|
	ZFS/ZPOOL configuration tests
 | 
						|
 | 
						|
OPTIONS:
 | 
						|
	-h      Show this message
 | 
						|
	-v      Verbose
 | 
						|
	-c      Cleanup lo+file devices at start
 | 
						|
	-t <#>  Run listed tests
 | 
						|
	-s <#>  Skip listed tests
 | 
						|
 | 
						|
EOF
 | 
						|
}
 | 
						|
 | 
						|
while getopts 'hvct:s:?' OPTION; do
 | 
						|
	case $OPTION in
 | 
						|
	h)
 | 
						|
		usage
 | 
						|
		exit 1
 | 
						|
		;;
 | 
						|
	v)
 | 
						|
		VERBOSE=1
 | 
						|
		;;
 | 
						|
	c)
 | 
						|
		CLEANUP=1
 | 
						|
		;;
 | 
						|
	t)
 | 
						|
		TESTS_RUN=($OPTARG)
 | 
						|
		;;
 | 
						|
	s)
 | 
						|
		TESTS_SKIP=($OPTARG)
 | 
						|
		;;
 | 
						|
	?)
 | 
						|
		usage
 | 
						|
		exit
 | 
						|
		;;
 | 
						|
	esac
 | 
						|
done
 | 
						|
 | 
						|
if [ $(id -u) != 0 ]; then
 | 
						|
	die "Must run as root"
 | 
						|
fi
 | 
						|
 | 
						|
# Initialize the test suite
 | 
						|
init
 | 
						|
 | 
						|
# Disable the udev rule 90-zfs.rules to prevent the zfs module
 | 
						|
# stack from being loaded due to the detection of a zfs device.
 | 
						|
# This is important because this test scripts require full control
 | 
						|
# over when and how the modules are loaded/unloaded.  A trap is
 | 
						|
# set to ensure the udev rule is correctly replaced on exit.
 | 
						|
RULE=${udevruledir}/90-zfs.rules
 | 
						|
if test -e  ${RULE}; then
 | 
						|
	trap "mv ${RULE}.disabled ${RULE}" INT TERM EXIT
 | 
						|
	mv ${RULE} ${RULE}.disabled
 | 
						|
fi
 | 
						|
 | 
						|
# Perform pre-cleanup is requested
 | 
						|
if [ ${CLEANUP} ]; then
 | 
						|
	${ZFS_SH} -u
 | 
						|
	cleanup_md_devices
 | 
						|
	cleanup_loop_devices
 | 
						|
	rm -f /tmp/zpool.cache.*
 | 
						|
fi
 | 
						|
 | 
						|
# Check if we need to skip the tests that require scsi_debug and lsscsi.
 | 
						|
SCSI_DEBUG=0
 | 
						|
${INFOMOD} scsi_debug &>/dev/null && SCSI_DEBUG=1
 | 
						|
HAVE_LSSCSI=0
 | 
						|
test -f ${LSSCSI} && HAVE_LSSCSI=1
 | 
						|
if [ ${SCSI_DEBUG} -eq 0 ] || [ ${HAVE_LSSCSI} -eq 0 ]; then
 | 
						|
	echo "Skipping test 10 which requires the scsi_debug " \
 | 
						|
		"module and the ${LSSCSI} utility"
 | 
						|
fi
 | 
						|
 | 
						|
# Validate persistent zpool.cache configuration.
 | 
						|
test_1() {
 | 
						|
	local POOL_NAME=test1
 | 
						|
	local TMP_FILE1=`mktemp`
 | 
						|
	local TMP_FILE2=`mktemp`
 | 
						|
	local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
 | 
						|
	# Create a pool save its status for comparison.
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
 | 
						|
 | 
						|
	# Unload/load the module stack and verify the pool persists.
 | 
						|
	${ZFS_SH} -u || fail 4
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 5
 | 
						|
	${ZPOOL} import -c ${TMP_CACHE} ${POOL_NAME} || fail 5
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 6
 | 
						|
	cmp ${TMP_FILE1} ${TMP_FILE2} || fail 7
 | 
						|
 | 
						|
	# Cleanup the test pool and temporary files
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 8
 | 
						|
	rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_CACHE} || fail 9
 | 
						|
	${ZFS_SH} -u || fail 10
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 1 "persistent zpool.cache"
 | 
						|
 | 
						|
# Validate ZFS disk scanning and import w/out zpool.cache configuration.
 | 
						|
test_2() {
 | 
						|
	local POOL_NAME=test2
 | 
						|
	local TMP_FILE1=`mktemp`
 | 
						|
	local TMP_FILE2=`mktemp`
 | 
						|
	local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
 | 
						|
	# Create a pool save its status for comparison.
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
 | 
						|
 | 
						|
	# Unload the module stack, remove the cache file, load the module
 | 
						|
	# stack and attempt to probe the disks to import the pool.  As
 | 
						|
	# a cross check verify the old pool state against the imported.
 | 
						|
	${ZFS_SH} -u || fail 4
 | 
						|
	rm -f ${TMP_CACHE} || fail 5
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 6
 | 
						|
	${ZPOOL} import -d /dev ${POOL_NAME} || fail 8
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 9
 | 
						|
	cmp ${TMP_FILE1} ${TMP_FILE2} || fail 10
 | 
						|
 | 
						|
	# Cleanup the test pool and temporary files
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 11
 | 
						|
	rm -f ${TMP_FILE1} ${TMP_FILE2} || fail 12
 | 
						|
	${ZFS_SH} -u || fail 13
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 2 "scan disks for pools to import"
 | 
						|
 | 
						|
zconfig_zvol_device_stat() {
 | 
						|
	local EXPECT=$1
 | 
						|
	local POOL_NAME=/dev/zvol/$2
 | 
						|
	local ZVOL_NAME=/dev/zvol/$3
 | 
						|
	local SNAP_NAME=/dev/zvol/$4
 | 
						|
	local CLONE_NAME=/dev/zvol/$5
 | 
						|
	local COUNT=0
 | 
						|
 | 
						|
	# Briefly delay for udev
 | 
						|
	udev_trigger
 | 
						|
 | 
						|
	# Pool exists
 | 
						|
	stat ${POOL_NAME} &>/dev/null   && let COUNT=$COUNT+1
 | 
						|
 | 
						|
	# Volume and partitions
 | 
						|
	stat ${ZVOL_NAME}  &>/dev/null  && let COUNT=$COUNT+1
 | 
						|
	stat ${ZVOL_NAME}-part1 &>/dev/null  && let COUNT=$COUNT+1
 | 
						|
	stat ${ZVOL_NAME}-part2 &>/dev/null  && let COUNT=$COUNT+1
 | 
						|
 | 
						|
	# Snapshot with partitions
 | 
						|
	stat ${SNAP_NAME}  &>/dev/null  && let COUNT=$COUNT+1
 | 
						|
	stat ${SNAP_NAME}-part1 &>/dev/null  && let COUNT=$COUNT+1
 | 
						|
	stat ${SNAP_NAME}-part2 &>/dev/null  && let COUNT=$COUNT+1
 | 
						|
 | 
						|
	# Clone with partitions
 | 
						|
	stat ${CLONE_NAME}  &>/dev/null && let COUNT=$COUNT+1
 | 
						|
	stat ${CLONE_NAME}-part1 &>/dev/null && let COUNT=$COUNT+1
 | 
						|
	stat ${CLONE_NAME}-part2 &>/dev/null && let COUNT=$COUNT+1
 | 
						|
 | 
						|
	if [ $EXPECT -ne $COUNT ]; then
 | 
						|
		return 1
 | 
						|
	fi
 | 
						|
 | 
						|
	return 0
 | 
						|
}
 | 
						|
 | 
						|
# zpool import/export device check
 | 
						|
# (1 volume, 2 partitions, 1 snapshot, 1 clone)
 | 
						|
test_3() {
 | 
						|
	local POOL_NAME=tank
 | 
						|
	local ZVOL_NAME=volume
 | 
						|
	local SNAP_NAME=snap
 | 
						|
	local CLONE_NAME=clone
 | 
						|
	local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
 | 
						|
	local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
 | 
						|
	local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
 | 
						|
	local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
 | 
						|
	# Create a pool, volume, partition, snapshot, and clone.
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
 | 
						|
	${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3
 | 
						|
	${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
 | 
						|
	label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
 | 
						|
	partition /dev/zvol/${FULL_ZVOL_NAME} primary 1% 50% || fail 4
 | 
						|
	partition /dev/zvol/${FULL_ZVOL_NAME} primary 51% -1 || fail 4
 | 
						|
	${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5
 | 
						|
	${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6
 | 
						|
 | 
						|
	# Verify the devices were created
 | 
						|
	zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7
 | 
						|
 | 
						|
	# Export the pool
 | 
						|
	${ZPOOL} export ${POOL_NAME} || fail 8
 | 
						|
 | 
						|
	# verify the devices were removed
 | 
						|
	zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9
 | 
						|
 | 
						|
	# Import the pool, wait 1 second for udev
 | 
						|
	${ZPOOL} import ${POOL_NAME} || fail 10
 | 
						|
 | 
						|
	# Verify the devices were created
 | 
						|
	zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11
 | 
						|
 | 
						|
	# Toggle the snapdev and observe snapshot device links toggled
 | 
						|
	${ZFS} set snapdev=hidden ${FULL_ZVOL_NAME} || fail 12
 | 
						|
	
 | 
						|
	zconfig_zvol_device_stat 7 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    "invalid" ${FULL_CLONE_NAME} || fail 13
 | 
						|
 | 
						|
	${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 14
 | 
						|
 | 
						|
	zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 15
 | 
						|
 | 
						|
	# Destroy the pool and consequently the devices
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 16
 | 
						|
 | 
						|
	# verify the devices were removed
 | 
						|
	zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 17
 | 
						|
 | 
						|
	${ZFS_SH} -u || fail 18
 | 
						|
	rm -f ${TMP_CACHE} || fail 19
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 3 "zpool import/export device"
 | 
						|
 | 
						|
# zpool insmod/rmmod device check (1 volume, 1 snapshot, 1 clone)
 | 
						|
test_4() {
 | 
						|
	POOL_NAME=tank
 | 
						|
	ZVOL_NAME=volume
 | 
						|
	SNAP_NAME=snap
 | 
						|
	CLONE_NAME=clone
 | 
						|
	FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
 | 
						|
	FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
 | 
						|
	FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
 | 
						|
	TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
 | 
						|
	# Create a pool, volume, snapshot, and clone
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
 | 
						|
	${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3
 | 
						|
	${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
 | 
						|
	label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
 | 
						|
	partition /dev/zvol/${FULL_ZVOL_NAME} primary 1% 50% || fail 4
 | 
						|
	partition /dev/zvol/${FULL_ZVOL_NAME} primary 51% -1 || fail 4
 | 
						|
	${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5
 | 
						|
	${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6
 | 
						|
 | 
						|
	# Verify the devices were created
 | 
						|
	zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7
 | 
						|
 | 
						|
	# Unload the modules
 | 
						|
	${ZFS_SH} -u || fail 8
 | 
						|
 | 
						|
	# Verify the devices were removed
 | 
						|
	zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9
 | 
						|
 | 
						|
	# Load the modules, list the pools to ensure they are opened
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 10
 | 
						|
	${ZPOOL} import -c ${TMP_CACHE} ${POOL_NAME} || fail 10
 | 
						|
	${ZPOOL} list &>/dev/null
 | 
						|
 | 
						|
	# Verify the devices were created
 | 
						|
	zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11
 | 
						|
 | 
						|
	# Destroy the pool and consequently the devices
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 12
 | 
						|
 | 
						|
	# Verify the devices were removed
 | 
						|
	zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \
 | 
						|
	    ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 13
 | 
						|
 | 
						|
	${ZFS_SH} -u || fail 14
 | 
						|
	rm -f ${TMP_CACHE} || fail 15
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 4 "zpool insmod/rmmod device"
 | 
						|
 | 
						|
# ZVOL volume sanity check
 | 
						|
test_5() {
 | 
						|
	local POOL_NAME=tank
 | 
						|
	local ZVOL_NAME=fish
 | 
						|
	local FULL_NAME=${POOL_NAME}/${ZVOL_NAME}
 | 
						|
	local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
 | 
						|
	# Create a pool and volume.
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raid0 || fail 2
 | 
						|
	${ZFS} create -V 800M ${FULL_NAME} || fail 3
 | 
						|
	label /dev/zvol/${FULL_NAME} msdos || fail 4
 | 
						|
	partition /dev/zvol/${FULL_NAME} primary 1 -1 || fail 4
 | 
						|
	format /dev/zvol/${FULL_NAME}-part1 ext2 || fail 5
 | 
						|
 | 
						|
	# Mount the ext2 filesystem and copy some data to it.
 | 
						|
	mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 6
 | 
						|
	mount /dev/zvol/${FULL_NAME}-part1 /tmp/${ZVOL_NAME}-part1 || fail 7
 | 
						|
	cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 8
 | 
						|
	sync
 | 
						|
 | 
						|
	# Verify the copied files match the original files.
 | 
						|
	diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \
 | 
						|
		&>/dev/null || fail 9
 | 
						|
 | 
						|
	# Remove the files, umount, destroy the volume and pool.
 | 
						|
	rm -Rf /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} || fail 10
 | 
						|
	umount /tmp/${ZVOL_NAME}-part1 || fail 11
 | 
						|
	rmdir /tmp/${ZVOL_NAME}-part1 || fail 12
 | 
						|
 | 
						|
	${ZFS} destroy ${FULL_NAME} || fail 13
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 14
 | 
						|
	${ZFS_SH} -u || fail 15
 | 
						|
	rm -f ${TMP_CACHE} || fail 16
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 5 "zvol+ext2 volume"
 | 
						|
 | 
						|
# ZVOL snapshot sanity check
 | 
						|
test_6() {
 | 
						|
	local POOL_NAME=tank
 | 
						|
	local ZVOL_NAME=fish
 | 
						|
	local SNAP_NAME=pristine
 | 
						|
	local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
 | 
						|
	local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
 | 
						|
	local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
 | 
						|
	# Create a pool and volume.
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raid0 || fail 2
 | 
						|
	${ZFS} create -s -V 800M ${FULL_ZVOL_NAME} || fail 3
 | 
						|
	${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
 | 
						|
	label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
 | 
						|
	partition /dev/zvol/${FULL_ZVOL_NAME} primary 1 -1 || fail 4
 | 
						|
	format /dev/zvol/${FULL_ZVOL_NAME}-part1 ext2 || fail 5
 | 
						|
 | 
						|
	# Mount the ext2 filesystem and copy some data to it.
 | 
						|
	mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 6
 | 
						|
	mount /dev/zvol/${FULL_ZVOL_NAME}-part1 /tmp/${ZVOL_NAME}-part1 \
 | 
						|
		|| fail 7
 | 
						|
 | 
						|
	# Snapshot the pristine ext2 filesystem and mount it read-only.
 | 
						|
	${ZFS} snapshot ${FULL_SNAP_NAME} || fail 8
 | 
						|
	wait_udev /dev/zvol/${FULL_SNAP_NAME}-part1 30 || fail 8
 | 
						|
	mkdir -p /tmp/${SNAP_NAME}-part1 || fail 9
 | 
						|
	mount /dev/zvol/${FULL_SNAP_NAME}-part1 /tmp/${SNAP_NAME}-part1 \
 | 
						|
		&>/dev/null || fail 10
 | 
						|
 | 
						|
	# Copy to original volume
 | 
						|
	cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 11
 | 
						|
	sync
 | 
						|
 | 
						|
	# Verify the copied files match the original files,
 | 
						|
	# and the copied files do NOT appear in the snapshot.
 | 
						|
	diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \
 | 
						|
		&>/dev/null || fail 12
 | 
						|
	diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}-part1/${SRC_DIR##*/} \
 | 
						|
		&>/dev/null && fail 13
 | 
						|
 | 
						|
	# umount, destroy the snapshot, volume, and pool.
 | 
						|
	umount /tmp/${SNAP_NAME}-part1 || fail 14
 | 
						|
	rmdir /tmp/${SNAP_NAME}-part1 || fail 15
 | 
						|
	${ZFS} destroy ${FULL_SNAP_NAME} || fail 16
 | 
						|
 | 
						|
	umount /tmp/${ZVOL_NAME}-part1 || fail 17
 | 
						|
	rmdir /tmp/${ZVOL_NAME}-part1 || fail 18
 | 
						|
	${ZFS} destroy ${FULL_ZVOL_NAME} || fail 19
 | 
						|
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 20
 | 
						|
	${ZFS_SH} -u || fail 21
 | 
						|
	rm -f ${TMP_CACHE} || fail 22
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 6 "zvol+ext2 snapshot"
 | 
						|
 | 
						|
# ZVOL clone sanity check
 | 
						|
test_7() {
 | 
						|
	local POOL_NAME=tank
 | 
						|
	local ZVOL_NAME=fish
 | 
						|
	local SNAP_NAME=pristine
 | 
						|
	local CLONE_NAME=clone
 | 
						|
	local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME}
 | 
						|
	local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME}
 | 
						|
	local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME}
 | 
						|
	local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
 | 
						|
	# Create a pool and volume.
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
 | 
						|
	${ZFS} create -V 300M ${FULL_ZVOL_NAME} || fail 3
 | 
						|
	${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3
 | 
						|
	label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4
 | 
						|
	partition /dev/zvol/${FULL_ZVOL_NAME} primary 1 -1 || fail 4
 | 
						|
	format /dev/zvol/${FULL_ZVOL_NAME}-part1 ext2 || fail 5
 | 
						|
 | 
						|
	# Snapshot the pristine ext2 filesystem.
 | 
						|
	${ZFS} snapshot ${FULL_SNAP_NAME} || fail 6
 | 
						|
	wait_udev /dev/zvol/${FULL_SNAP_NAME}-part1 30 || fail 7
 | 
						|
 | 
						|
	# Mount the ext2 filesystem so some data can be copied to it.
 | 
						|
	mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 7
 | 
						|
	mount /dev/zvol/${FULL_ZVOL_NAME}-part1 \
 | 
						|
		/tmp/${ZVOL_NAME}-part1 || fail 8
 | 
						|
 | 
						|
	# Mount the pristine ext2 snapshot.
 | 
						|
	mkdir -p /tmp/${SNAP_NAME}-part1 || fail 9
 | 
						|
	mount /dev/zvol/${FULL_SNAP_NAME}-part1 \
 | 
						|
		/tmp/${SNAP_NAME}-part1 &>/dev/null || fail 10
 | 
						|
 | 
						|
	# Copy to original volume.
 | 
						|
	cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 11
 | 
						|
	sync
 | 
						|
 | 
						|
	# Verify the copied files match the original files,
 | 
						|
	# and the copied files do NOT appear in the snapshot.
 | 
						|
	diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \
 | 
						|
		&>/dev/null || fail 12
 | 
						|
	diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}-part1/${SRC_DIR##*/} \
 | 
						|
		&>/dev/null && fail 13
 | 
						|
 | 
						|
	# Clone from the original pristine snapshot
 | 
						|
	${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 14
 | 
						|
	wait_udev /dev/zvol/${FULL_CLONE_NAME}-part1 30 || fail 14
 | 
						|
	mkdir -p /tmp/${CLONE_NAME}-part1 || fail 15
 | 
						|
	mount /dev/zvol/${FULL_CLONE_NAME}-part1 \
 | 
						|
		/tmp/${CLONE_NAME}-part1 || fail 16
 | 
						|
 | 
						|
	# Verify the clone matches the pristine snapshot,
 | 
						|
	# and the files copied to the original volume are NOT there.
 | 
						|
	diff -ur /tmp/${SNAP_NAME}-part1 /tmp/${CLONE_NAME}-part1 \
 | 
						|
		&>/dev/null || fail 17
 | 
						|
	diff -ur /tmp/${ZVOL_NAME}-part1 /tmp/${CLONE_NAME}-part1 \
 | 
						|
		&>/dev/null && fail 18
 | 
						|
 | 
						|
	# Copy to cloned volume.
 | 
						|
	cp -RL ${SRC_DIR} /tmp/${CLONE_NAME}-part1 || fail 19
 | 
						|
	sync
 | 
						|
 | 
						|
	# Verify the clone matches the modified original volume.
 | 
						|
	diff -ur /tmp/${ZVOL_NAME}-part1 /tmp/${CLONE_NAME}-part1 \
 | 
						|
		&>/dev/null || fail 20
 | 
						|
 | 
						|
	# umount, destroy the snapshot, volume, and pool.
 | 
						|
	umount /tmp/${CLONE_NAME}-part1 || fail 21
 | 
						|
	rmdir /tmp/${CLONE_NAME}-part1 || fail 22
 | 
						|
	${ZFS} destroy ${FULL_CLONE_NAME} || fail 23
 | 
						|
 | 
						|
	umount /tmp/${SNAP_NAME}-part1 || fail 24
 | 
						|
	rmdir /tmp/${SNAP_NAME}-part1 || fail 25
 | 
						|
	${ZFS} destroy ${FULL_SNAP_NAME} || fail 26
 | 
						|
 | 
						|
	umount /tmp/${ZVOL_NAME}-part1 || fail 27
 | 
						|
	rmdir /tmp/${ZVOL_NAME}-part1 || fail 28
 | 
						|
	${ZFS} destroy ${FULL_ZVOL_NAME} || fail 29
 | 
						|
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 30
 | 
						|
	${ZFS_SH} -u || fail 31
 | 
						|
	rm -f ${TMP_CACHE} || fail 32
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 7 "zvol+ext2 clone"
 | 
						|
 | 
						|
# Send/Receive sanity check
 | 
						|
test_8() {
 | 
						|
	local POOL_NAME1=tank1
 | 
						|
	local POOL_NAME2=tank2
 | 
						|
	local ZVOL_NAME=fish
 | 
						|
	local SNAP_NAME=snap
 | 
						|
	local FULL_ZVOL_NAME1=${POOL_NAME1}/${ZVOL_NAME}
 | 
						|
	local FULL_ZVOL_NAME2=${POOL_NAME2}/${ZVOL_NAME}
 | 
						|
	local FULL_SNAP_NAME1=${POOL_NAME1}/${ZVOL_NAME}@${SNAP_NAME}
 | 
						|
	local FULL_SNAP_NAME2=${POOL_NAME2}/${ZVOL_NAME}@${SNAP_NAME}
 | 
						|
	local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
 | 
						|
	# Create two pools and a volume
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 || fail 2
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 || fail 2
 | 
						|
	${ZFS} create -V 300M ${FULL_ZVOL_NAME1} || fail 3
 | 
						|
	${ZFS} set snapdev=visible ${FULL_ZVOL_NAME1} || fail 3
 | 
						|
	label /dev/zvol/${FULL_ZVOL_NAME1} msdos || fail 4
 | 
						|
	partition /dev/zvol/${FULL_ZVOL_NAME1} primary 1 -1 || fail 4
 | 
						|
	format /dev/zvol/${FULL_ZVOL_NAME1}-part1 ext2 || fail 5
 | 
						|
 | 
						|
	# Mount the ext2 filesystem and copy some data to it.
 | 
						|
	mkdir -p /tmp/${FULL_ZVOL_NAME1}-part1 || fail 6
 | 
						|
	mount /dev/zvol/${FULL_ZVOL_NAME1}-part1 \
 | 
						|
		/tmp/${FULL_ZVOL_NAME1}-part1 || fail 7
 | 
						|
	cp -RL ${SRC_DIR} /tmp/${FULL_ZVOL_NAME1}-part1 || fail 8
 | 
						|
 | 
						|
	# Unmount, snapshot, mount the ext2 filesystem so it may be sent.
 | 
						|
	# We only unmount to ensure the ext2 filesystem is clean.
 | 
						|
	umount /tmp/${FULL_ZVOL_NAME1}-part1 || fail 9
 | 
						|
	${ZFS} snapshot ${FULL_SNAP_NAME1} || fail 10
 | 
						|
	wait_udev /dev/zvol/${FULL_SNAP_NAME1} 30 || fail 10
 | 
						|
	mount /dev/zvol/${FULL_ZVOL_NAME1}-part1 \
 | 
						|
		/tmp/${FULL_ZVOL_NAME1}-part1 || 11
 | 
						|
 | 
						|
	# Send/receive the snapshot from POOL_NAME1 to POOL_NAME2
 | 
						|
	(${ZFS} send ${FULL_SNAP_NAME1} | \
 | 
						|
	${ZFS} receive ${FULL_ZVOL_NAME2}) || fail 12
 | 
						|
	wait_udev /dev/zvol/${FULL_ZVOL_NAME2}-part1 30 || fail 12
 | 
						|
 | 
						|
	# Mount the sent ext2 filesystem.
 | 
						|
	mkdir -p /tmp/${FULL_ZVOL_NAME2}-part1 || fail 13
 | 
						|
	mount /dev/zvol/${FULL_ZVOL_NAME2}-part1 \
 | 
						|
		/tmp/${FULL_ZVOL_NAME2}-part1 || fail 14
 | 
						|
 | 
						|
	# Verify the contents of the volumes match
 | 
						|
	diff -ur /tmp/${FULL_ZVOL_NAME1}-part1 /tmp/${FULL_ZVOL_NAME2}-part1 \
 | 
						|
	    &>/dev/null || fail 15
 | 
						|
 | 
						|
	# Umount, destroy the volume and pool.
 | 
						|
	umount /tmp/${FULL_ZVOL_NAME1}-part1 || fail 16
 | 
						|
	umount /tmp/${FULL_ZVOL_NAME2}-part1 || fail 17
 | 
						|
	rmdir /tmp/${FULL_ZVOL_NAME1}-part1 || fail 18
 | 
						|
	rmdir /tmp/${FULL_ZVOL_NAME2}-part1 || fail 19
 | 
						|
	rmdir /tmp/${POOL_NAME1} || fail 20
 | 
						|
	rmdir /tmp/${POOL_NAME2} || fail 21
 | 
						|
 | 
						|
	${ZFS} destroy ${FULL_SNAP_NAME1} || fail 22
 | 
						|
	${ZFS} destroy ${FULL_SNAP_NAME2} || fail 23
 | 
						|
	${ZFS} destroy ${FULL_ZVOL_NAME1} || fail 24
 | 
						|
	${ZFS} destroy ${FULL_ZVOL_NAME2} || fail 25
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 -d || fail 26
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 -d || fail 27
 | 
						|
	${ZFS_SH} -u || fail 28
 | 
						|
	rm -f ${TMP_CACHE} || fail 29
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 8 "zfs send/receive"
 | 
						|
 | 
						|
# zpool event sanity check
 | 
						|
test_9() {
 | 
						|
	local POOL_NAME=tank
 | 
						|
	local ZVOL_NAME=fish
 | 
						|
	local FULL_NAME=${POOL_NAME}/${ZVOL_NAME}
 | 
						|
	local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
	local TMP_EVENTS=`mktemp -p /tmp zpool.events.XXXXXXXX`
 | 
						|
 | 
						|
	# Create a pool and volume.
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
 | 
						|
	${ZFS} create -V 300M ${FULL_NAME} || fail 3
 | 
						|
	udev_trigger
 | 
						|
 | 
						|
	# Dump the events, there should be a pool create event
 | 
						|
	${ZPOOL} events >${TMP_EVENTS} || fail 4
 | 
						|
	MATCHES=`grep -c sysevent\.fs\.zfs\.pool_create ${TMP_EVENTS}`
 | 
						|
	[ $MATCHES -eq 1 ] || fail 5
 | 
						|
 | 
						|
	# Clear the events and ensure there are none.
 | 
						|
	${ZPOOL} events -c >/dev/null || fail 6
 | 
						|
	${ZPOOL} events >${TMP_EVENTS} || fail 7
 | 
						|
	EVENTS=`wc -l ${TMP_EVENTS} | cut -f1 -d' '`
 | 
						|
	[ $EVENTS -gt 1 ] && fail 8
 | 
						|
 | 
						|
	${ZFS} destroy ${FULL_NAME} || fail 9
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 10
 | 
						|
	${ZFS_SH} -u || fail 11
 | 
						|
	rm -f ${TMP_CACHE} || fail 12
 | 
						|
	rm -f ${TMP_EVENTS} || fail 13
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 9 "zpool events"
 | 
						|
 | 
						|
zconfig_add_vdev() {
 | 
						|
	local POOL_NAME=$1
 | 
						|
	local TYPE=$2
 | 
						|
	local DEVICE=$3
 | 
						|
	local TMP_FILE1=`mktemp`
 | 
						|
	local TMP_FILE2=`mktemp`
 | 
						|
	local TMP_FILE3=`mktemp`
 | 
						|
 | 
						|
	BASE_DEVICE=`basename ${DEVICE}`
 | 
						|
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE1}
 | 
						|
	${ZPOOL} add -f ${POOL_NAME} ${TYPE} ${DEVICE} 2>/dev/null || return 1
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE2}
 | 
						|
	diff ${TMP_FILE1} ${TMP_FILE2} > ${TMP_FILE3}
 | 
						|
 | 
						|
	[ `wc -l ${TMP_FILE3}|${AWK} '{print $1}'` -eq 3 ] || return 1
 | 
						|
 | 
						|
	PARENT_VDEV=`tail -2 ${TMP_FILE3} | head -1 | ${AWK} '{print $NF}'`
 | 
						|
	case $TYPE in
 | 
						|
	cache)
 | 
						|
		[ "${PARENT_VDEV}" = "${TYPE}" ] || return 1
 | 
						|
		;;
 | 
						|
	log)
 | 
						|
		[ "${PARENT_VDEV}" = "logs" ] || return 1
 | 
						|
		;;
 | 
						|
	esac
 | 
						|
 | 
						|
	if ! tail -1 ${TMP_FILE3} |
 | 
						|
	    egrep -q "^>[[:space:]]+${BASE_DEVICE}[[:space:]]+ONLINE" ; then
 | 
						|
		return 1
 | 
						|
	fi
 | 
						|
	rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_FILE3}
 | 
						|
 | 
						|
	return 0
 | 
						|
}
 | 
						|
 | 
						|
# zpool add and remove sanity check
 | 
						|
test_10() {
 | 
						|
	local POOL_NAME=tank
 | 
						|
	local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
 | 
						|
	local TMP_FILE1=`mktemp`
 | 
						|
	local TMP_FILE2=`mktemp`
 | 
						|
 | 
						|
	if [ ${SCSI_DEBUG} -eq 0 ] || [ ${HAVE_LSSCSI} -eq 0 ] ; then
 | 
						|
		skip
 | 
						|
		return
 | 
						|
	fi
 | 
						|
 | 
						|
	test `${LSMOD} | grep -c scsi_debug` -gt 0 && \
 | 
						|
		(${RMMOD} scsi_debug || exit 1)
 | 
						|
 | 
						|
	/sbin/modprobe scsi_debug dev_size_mb=128 ||
 | 
						|
		die "Error $? creating scsi_debug device"
 | 
						|
	udev_trigger
 | 
						|
 | 
						|
	SDDEVICE=`${LSSCSI}|${AWK} '/scsi_debug/ { print $6; exit }'`
 | 
						|
	BASE_SDDEVICE=`basename $SDDEVICE`
 | 
						|
 | 
						|
	# Create a pool
 | 
						|
	${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3
 | 
						|
 | 
						|
	# Add and remove a cache vdev by full path
 | 
						|
	zconfig_add_vdev ${POOL_NAME} cache ${SDDEVICE} || fail 4
 | 
						|
	${ZPOOL} remove ${POOL_NAME} ${SDDEVICE} || fail 5
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 6
 | 
						|
	cmp ${TMP_FILE1} ${TMP_FILE2} || fail 7
 | 
						|
	sleep 1
 | 
						|
 | 
						|
	# Add and remove a cache vdev by shorthand path
 | 
						|
	zconfig_add_vdev ${POOL_NAME} cache ${BASE_SDDEVICE} || fail 8
 | 
						|
	${ZPOOL} remove ${POOL_NAME} ${BASE_SDDEVICE} || fail 9
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 10
 | 
						|
	cmp ${TMP_FILE1} ${TMP_FILE2} || fail 11
 | 
						|
	sleep 1
 | 
						|
 | 
						|
	# Add and remove a log vdev
 | 
						|
	zconfig_add_vdev ${POOL_NAME} log ${BASE_SDDEVICE} || fail 12
 | 
						|
	${ZPOOL} remove ${POOL_NAME} ${BASE_SDDEVICE} || fail 13
 | 
						|
	${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 14
 | 
						|
	cmp ${TMP_FILE1} ${TMP_FILE2} || fail 15
 | 
						|
 | 
						|
	${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 16
 | 
						|
	${ZFS_SH} -u || fail 17
 | 
						|
	${RMMOD} scsi_debug || fail 18
 | 
						|
 | 
						|
	rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_CACHE} || fail 19
 | 
						|
 | 
						|
	pass
 | 
						|
}
 | 
						|
run_test 10 "zpool add/remove vdev"
 | 
						|
 | 
						|
exit 0
 |