mirror of
https://git.proxmox.com/git/mirror_zfs.git
synced 2025-12-01 09:32:08 +03:00
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
110 lines
3.4 KiB
Bash
Executable File
110 lines
3.4 KiB
Bash
Executable File
#!/bin/ksh -p
|
|
#
|
|
# CDDL HEADER START
|
|
#
|
|
# This file and its contents are supplied under the terms of the
|
|
# Common Development and Distribution License ("CDDL"), version 1.0.
|
|
# You may only use this file in accordance with the terms of version
|
|
# 1.0 of the CDDL.
|
|
#
|
|
# A full copy of the text of the CDDL should have accompanied this
|
|
# source. A copy of the CDDL is also available via the Internet at
|
|
# http://www.illumos.org/license/CDDL.
|
|
#
|
|
# CDDL HEADER END
|
|
#
|
|
|
|
#
|
|
# Copyright (c) 2019 by Tim Chase. All rights reserved.
|
|
# Copyright (c) 2019 Lawrence Livermore National Security, LLC.
|
|
#
|
|
|
|
. $STF_SUITE/include/libtest.shlib
|
|
. $STF_SUITE/tests/functional/trim/trim.kshlib
|
|
. $STF_SUITE/tests/functional/trim/trim.cfg
|
|
|
|
#
|
|
# DESCRIPTION:
|
|
# Check various pool geometries stripe, mirror, raidz)
|
|
#
|
|
# STRATEGY:
|
|
# 1. Create a pool on file vdevs to trim.
|
|
# 2. Set 'autotrim=on' on pool.
|
|
# 3. Fill the pool to a known percentage of capacity.
|
|
# 4. Verify the vdevs contain 75% or more allocated blocks.
|
|
# 5. Remove all files making it possible to trim the entire pool.
|
|
# 6. Wait for auto trim to issue trim IOs for the free blocks.
|
|
# 7. Verify the disks contain 30% or less allocated blocks.
|
|
# 8. Repeat for test for striped, mirrored, and RAIDZ pools.
|
|
|
|
verify_runnable "global"
|
|
|
|
log_assert "Set 'autotrim=on' verify pool disks were trimmed"
|
|
|
|
function cleanup
|
|
{
|
|
if poolexists $TESTPOOL; then
|
|
destroy_pool $TESTPOOL
|
|
fi
|
|
|
|
log_must rm -f $TRIM_VDEVS
|
|
|
|
log_must set_tunable64 TRIM_EXTENT_BYTES_MIN $trim_extent_bytes_min
|
|
log_must set_tunable64 TRIM_TXG_BATCH $trim_txg_batch
|
|
log_must set_tunable64 VDEV_MIN_MS_COUNT $vdev_min_ms_count
|
|
}
|
|
log_onexit cleanup
|
|
|
|
# Minimum trim size is decreased to verify all trim sizes.
|
|
typeset trim_extent_bytes_min=$(get_tunable TRIM_EXTENT_BYTES_MIN)
|
|
log_must set_tunable64 TRIM_EXTENT_BYTES_MIN 4096
|
|
|
|
# Reduced TRIM_TXG_BATCH to make trimming more frequent.
|
|
typeset trim_txg_batch=$(get_tunable TRIM_TXG_BATCH)
|
|
log_must set_tunable64 TRIM_TXG_BATCH 8
|
|
|
|
# Increased metaslabs to better simulate larger more realistic devices.
|
|
typeset vdev_min_ms_count=$(get_tunable VDEV_MIN_MS_COUNT)
|
|
log_must set_tunable64 VDEV_MIN_MS_COUNT 32
|
|
|
|
typeset VDEV_MAX_MB=$(( floor(4 * MINVDEVSIZE * 0.75 / 1024 / 1024) ))
|
|
typeset VDEV_MIN_MB=$(( floor(4 * MINVDEVSIZE * 0.30 / 1024 / 1024) ))
|
|
|
|
for type in "" "mirror" "raidz2" "draid"; do
|
|
|
|
if [[ "$type" = "" ]]; then
|
|
VDEVS="$TRIM_VDEV1"
|
|
elif [[ "$type" = "mirror" ]]; then
|
|
VDEVS="$TRIM_VDEV1 $TRIM_VDEV2"
|
|
elif [[ "$type" = "raidz2" ]]; then
|
|
VDEVS="$TRIM_VDEV1 $TRIM_VDEV2 $TRIM_VDEV3"
|
|
elif [[ "$type" = "draid" ]]; then
|
|
VDEVS="$TRIM_VDEV1 $TRIM_VDEV2 $TRIM_VDEV3 $TRIM_VDEV4"
|
|
|
|
# The per-vdev utilization is lower due to the capacity
|
|
# resilverd for the distributed spare.
|
|
VDEV_MAX_MB=$(( floor(4 * MINVDEVSIZE * 0.50 / 1024 / 1024) ))
|
|
fi
|
|
|
|
log_must truncate -s $((4 * MINVDEVSIZE)) $VDEVS
|
|
log_must zpool create -f $TESTPOOL $VDEVS
|
|
log_must zpool set autotrim=on $TESTPOOL
|
|
|
|
typeset availspace=$(get_prop available $TESTPOOL)
|
|
typeset fill_mb=$(( floor(availspace * 0.90 / 1024 / 1024) ))
|
|
|
|
# Fill the pool, verify the vdevs are no longer sparse.
|
|
file_write -o create -f /$TESTPOOL/file -b 1048576 -c $fill_mb -d R
|
|
verify_vdevs "-ge" "$VDEV_MAX_MB" $VDEVS
|
|
|
|
# Remove the file, wait for trim, verify the vdevs are now sparse.
|
|
log_must rm /$TESTPOOL/file
|
|
wait_trim_io $TESTPOOL "ind" 64
|
|
verify_vdevs "-le" "$VDEV_MIN_MB" $VDEVS
|
|
|
|
log_must zpool destroy $TESTPOOL
|
|
log_must rm -f $VDEVS
|
|
done
|
|
|
|
log_pass "Auto trim successfully shrunk vdevs"
|