2019-03-29 19:13:20 +03:00
|
|
|
/*
|
|
|
|
* CDDL HEADER START
|
|
|
|
*
|
|
|
|
* The contents of this file are subject to the terms of the
|
|
|
|
* Common Development and Distribution License (the "License").
|
|
|
|
* You may not use this file except in compliance with the License.
|
|
|
|
*
|
|
|
|
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
|
|
|
|
* or http://www.opensolaris.org/os/licensing.
|
|
|
|
* See the License for the specific language governing permissions
|
|
|
|
* and limitations under the License.
|
|
|
|
*
|
|
|
|
* When distributing Covered Code, include this CDDL HEADER in each
|
|
|
|
* file and include the License file at usr/src/OPENSOLARIS.LICENSE.
|
|
|
|
* If applicable, add the following below this CDDL HEADER, with the
|
|
|
|
* fields enclosed by brackets "[]" replaced with your own identifying
|
|
|
|
* information: Portions Copyright [yyyy] [name of copyright owner]
|
|
|
|
*
|
|
|
|
* CDDL HEADER END
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Copyright (c) 2016 by Delphix. All rights reserved.
|
|
|
|
* Copyright (c) 2019 by Lawrence Livermore National Security, LLC.
|
2021-03-02 21:27:27 +03:00
|
|
|
* Copyright (c) 2021 Hewlett Packard Enterprise Development LP
|
2019-03-29 19:13:20 +03:00
|
|
|
*/
|
|
|
|
|
|
|
|
#include <sys/spa.h>
|
|
|
|
#include <sys/spa_impl.h>
|
|
|
|
#include <sys/txg.h>
|
|
|
|
#include <sys/vdev_impl.h>
|
|
|
|
#include <sys/vdev_trim.h>
|
|
|
|
#include <sys/metaslab_impl.h>
|
|
|
|
#include <sys/dsl_synctask.h>
|
|
|
|
#include <sys/zap.h>
|
|
|
|
#include <sys/dmu_tx.h>
|
2020-06-09 20:15:08 +03:00
|
|
|
#include <sys/arc_impl.h>
|
2019-03-29 19:13:20 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* TRIM is a feature which is used to notify a SSD that some previously
|
|
|
|
* written space is no longer allocated by the pool. This is useful because
|
|
|
|
* writes to a SSD must be performed to blocks which have first been erased.
|
|
|
|
* Ensuring the SSD always has a supply of erased blocks for new writes
|
|
|
|
* helps prevent the performance from deteriorating.
|
|
|
|
*
|
|
|
|
* There are two supported TRIM methods; manual and automatic.
|
|
|
|
*
|
|
|
|
* Manual TRIM:
|
|
|
|
*
|
|
|
|
* A manual TRIM is initiated by running the 'zpool trim' command. A single
|
|
|
|
* 'vdev_trim' thread is created for each leaf vdev, and it is responsible for
|
|
|
|
* managing that vdev TRIM process. This involves iterating over all the
|
|
|
|
* metaslabs, calculating the unallocated space ranges, and then issuing the
|
|
|
|
* required TRIM I/Os.
|
|
|
|
*
|
|
|
|
* While a metaslab is being actively trimmed it is not eligible to perform
|
|
|
|
* new allocations. After traversing all of the metaslabs the thread is
|
|
|
|
* terminated. Finally, both the requested options and current progress of
|
|
|
|
* the TRIM are regularly written to the pool. This allows the TRIM to be
|
|
|
|
* suspended and resumed as needed.
|
|
|
|
*
|
|
|
|
* Automatic TRIM:
|
|
|
|
*
|
|
|
|
* An automatic TRIM is enabled by setting the 'autotrim' pool property
|
|
|
|
* to 'on'. When enabled, a `vdev_autotrim' thread is created for each
|
|
|
|
* top-level (not leaf) vdev in the pool. These threads perform the same
|
|
|
|
* core TRIM process as a manual TRIM, but with a few key differences.
|
|
|
|
*
|
|
|
|
* 1) Automatic TRIM happens continuously in the background and operates
|
|
|
|
* solely on recently freed blocks (ms_trim not ms_allocatable).
|
|
|
|
*
|
|
|
|
* 2) Each thread is associated with a top-level (not leaf) vdev. This has
|
|
|
|
* the benefit of simplifying the threading model, it makes it easier
|
|
|
|
* to coordinate administrative commands, and it ensures only a single
|
|
|
|
* metaslab is disabled at a time. Unlike manual TRIM, this means each
|
|
|
|
* 'vdev_autotrim' thread is responsible for issuing TRIM I/Os for its
|
|
|
|
* children.
|
|
|
|
*
|
|
|
|
* 3) There is no automatic TRIM progress information stored on disk, nor
|
|
|
|
* is it reported by 'zpool status'.
|
|
|
|
*
|
|
|
|
* While the automatic TRIM process is highly effective it is more likely
|
|
|
|
* than a manual TRIM to encounter tiny ranges. Ranges less than or equal to
|
|
|
|
* 'zfs_trim_extent_bytes_min' (32k) are considered too small to efficiently
|
|
|
|
* TRIM and are skipped. This means small amounts of freed space may not
|
|
|
|
* be automatically trimmed.
|
|
|
|
*
|
|
|
|
* Furthermore, devices with attached hot spares and devices being actively
|
|
|
|
* replaced are skipped. This is done to avoid adding additional stress to
|
|
|
|
* a potentially unhealthy device and to minimize the required rebuild time.
|
|
|
|
*
|
|
|
|
* For this reason it may be beneficial to occasionally manually TRIM a pool
|
|
|
|
* even when automatic TRIM is enabled.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Maximum size of TRIM I/O, ranges will be chunked in to 128MiB lengths.
|
|
|
|
*/
|
|
|
|
unsigned int zfs_trim_extent_bytes_max = 128 * 1024 * 1024;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Minimum size of TRIM I/O, extents smaller than 32Kib will be skipped.
|
|
|
|
*/
|
|
|
|
unsigned int zfs_trim_extent_bytes_min = 32 * 1024;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Skip uninitialized metaslabs during the TRIM process. This option is
|
|
|
|
* useful for pools constructed from large thinly-provisioned devices where
|
|
|
|
* TRIM operations are slow. As a pool ages an increasing fraction of
|
|
|
|
* the pools metaslabs will be initialized progressively degrading the
|
|
|
|
* usefulness of this option. This setting is stored when starting a
|
|
|
|
* manual TRIM and will persist for the duration of the requested TRIM.
|
|
|
|
*/
|
|
|
|
unsigned int zfs_trim_metaslab_skip = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Maximum number of queued TRIM I/Os per leaf vdev. The number of
|
|
|
|
* concurrent TRIM I/Os issued to the device is controlled by the
|
|
|
|
* zfs_vdev_trim_min_active and zfs_vdev_trim_max_active module options.
|
|
|
|
*/
|
|
|
|
unsigned int zfs_trim_queue_limit = 10;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The minimum number of transaction groups between automatic trims of a
|
|
|
|
* metaslab. This setting represents a trade-off between issuing more
|
|
|
|
* efficient TRIM operations, by allowing them to be aggregated longer,
|
|
|
|
* and issuing them promptly so the trimmed space is available. Note
|
|
|
|
* that this value is a minimum; metaslabs can be trimmed less frequently
|
|
|
|
* when there are a large number of ranges which need to be trimmed.
|
|
|
|
*
|
|
|
|
* Increasing this value will allow frees to be aggregated for a longer
|
|
|
|
* time. This can result is larger TRIM operations, and increased memory
|
|
|
|
* usage in order to track the ranges to be trimmed. Decreasing this value
|
|
|
|
* has the opposite effect. The default value of 32 was determined though
|
|
|
|
* testing to be a reasonable compromise.
|
|
|
|
*/
|
|
|
|
unsigned int zfs_trim_txg_batch = 32;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The trim_args are a control structure which describe how a leaf vdev
|
|
|
|
* should be trimmed. The core elements are the vdev, the metaslab being
|
|
|
|
* trimmed and a range tree containing the extents to TRIM. All provided
|
|
|
|
* ranges must be within the metaslab.
|
|
|
|
*/
|
|
|
|
typedef struct trim_args {
|
|
|
|
/*
|
|
|
|
* These fields are set by the caller of vdev_trim_ranges().
|
|
|
|
*/
|
|
|
|
vdev_t *trim_vdev; /* Leaf vdev to TRIM */
|
|
|
|
metaslab_t *trim_msp; /* Disabled metaslab */
|
|
|
|
range_tree_t *trim_tree; /* TRIM ranges (in metaslab) */
|
|
|
|
trim_type_t trim_type; /* Manual or auto TRIM */
|
|
|
|
uint64_t trim_extent_bytes_max; /* Maximum TRIM I/O size */
|
|
|
|
uint64_t trim_extent_bytes_min; /* Minimum TRIM I/O size */
|
|
|
|
enum trim_flag trim_flags; /* TRIM flags (secure) */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* These fields are updated by vdev_trim_ranges().
|
|
|
|
*/
|
|
|
|
hrtime_t trim_start_time; /* Start time */
|
|
|
|
uint64_t trim_bytes_done; /* Bytes trimmed */
|
|
|
|
} trim_args_t;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Determines whether a vdev_trim_thread() should be stopped.
|
|
|
|
*/
|
|
|
|
static boolean_t
|
|
|
|
vdev_trim_should_stop(vdev_t *vd)
|
|
|
|
{
|
|
|
|
return (vd->vdev_trim_exit_wanted || !vdev_writeable(vd) ||
|
|
|
|
vd->vdev_detached || vd->vdev_top->vdev_removing);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Determines whether a vdev_autotrim_thread() should be stopped.
|
|
|
|
*/
|
|
|
|
static boolean_t
|
|
|
|
vdev_autotrim_should_stop(vdev_t *tvd)
|
|
|
|
{
|
|
|
|
return (tvd->vdev_autotrim_exit_wanted ||
|
|
|
|
!vdev_writeable(tvd) || tvd->vdev_removing ||
|
|
|
|
spa_get_autotrim(tvd->vdev_spa) == SPA_AUTOTRIM_OFF);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The sync task for updating the on-disk state of a manual TRIM. This
|
|
|
|
* is scheduled by vdev_trim_change_state().
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_zap_update_sync(void *arg, dmu_tx_t *tx)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* We pass in the guid instead of the vdev_t since the vdev may
|
|
|
|
* have been freed prior to the sync task being processed. This
|
|
|
|
* happens when a vdev is detached as we call spa_config_vdev_exit(),
|
|
|
|
* stop the trimming thread, schedule the sync task, and free
|
|
|
|
* the vdev. Later when the scheduled sync task is invoked, it would
|
|
|
|
* find that the vdev has been freed.
|
|
|
|
*/
|
|
|
|
uint64_t guid = *(uint64_t *)arg;
|
|
|
|
uint64_t txg = dmu_tx_get_txg(tx);
|
|
|
|
kmem_free(arg, sizeof (uint64_t));
|
|
|
|
|
|
|
|
vdev_t *vd = spa_lookup_by_guid(tx->tx_pool->dp_spa, guid, B_FALSE);
|
|
|
|
if (vd == NULL || vd->vdev_top->vdev_removing || !vdev_is_concrete(vd))
|
|
|
|
return;
|
|
|
|
|
|
|
|
uint64_t last_offset = vd->vdev_trim_offset[txg & TXG_MASK];
|
|
|
|
vd->vdev_trim_offset[txg & TXG_MASK] = 0;
|
|
|
|
|
|
|
|
VERIFY3U(vd->vdev_leaf_zap, !=, 0);
|
|
|
|
|
|
|
|
objset_t *mos = vd->vdev_spa->spa_meta_objset;
|
|
|
|
|
|
|
|
if (last_offset > 0 || vd->vdev_trim_last_offset == UINT64_MAX) {
|
|
|
|
|
|
|
|
if (vd->vdev_trim_last_offset == UINT64_MAX)
|
|
|
|
last_offset = 0;
|
|
|
|
|
|
|
|
vd->vdev_trim_last_offset = last_offset;
|
|
|
|
VERIFY0(zap_update(mos, vd->vdev_leaf_zap,
|
|
|
|
VDEV_LEAF_ZAP_TRIM_LAST_OFFSET,
|
|
|
|
sizeof (last_offset), 1, &last_offset, tx));
|
|
|
|
}
|
|
|
|
|
|
|
|
if (vd->vdev_trim_action_time > 0) {
|
|
|
|
uint64_t val = (uint64_t)vd->vdev_trim_action_time;
|
|
|
|
VERIFY0(zap_update(mos, vd->vdev_leaf_zap,
|
|
|
|
VDEV_LEAF_ZAP_TRIM_ACTION_TIME, sizeof (val),
|
|
|
|
1, &val, tx));
|
|
|
|
}
|
|
|
|
|
|
|
|
if (vd->vdev_trim_rate > 0) {
|
|
|
|
uint64_t rate = (uint64_t)vd->vdev_trim_rate;
|
|
|
|
|
|
|
|
if (rate == UINT64_MAX)
|
|
|
|
rate = 0;
|
|
|
|
|
|
|
|
VERIFY0(zap_update(mos, vd->vdev_leaf_zap,
|
|
|
|
VDEV_LEAF_ZAP_TRIM_RATE, sizeof (rate), 1, &rate, tx));
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t partial = vd->vdev_trim_partial;
|
|
|
|
if (partial == UINT64_MAX)
|
|
|
|
partial = 0;
|
|
|
|
|
|
|
|
VERIFY0(zap_update(mos, vd->vdev_leaf_zap, VDEV_LEAF_ZAP_TRIM_PARTIAL,
|
|
|
|
sizeof (partial), 1, &partial, tx));
|
|
|
|
|
|
|
|
uint64_t secure = vd->vdev_trim_secure;
|
|
|
|
if (secure == UINT64_MAX)
|
|
|
|
secure = 0;
|
|
|
|
|
|
|
|
VERIFY0(zap_update(mos, vd->vdev_leaf_zap, VDEV_LEAF_ZAP_TRIM_SECURE,
|
|
|
|
sizeof (secure), 1, &secure, tx));
|
|
|
|
|
|
|
|
|
|
|
|
uint64_t trim_state = vd->vdev_trim_state;
|
|
|
|
VERIFY0(zap_update(mos, vd->vdev_leaf_zap, VDEV_LEAF_ZAP_TRIM_STATE,
|
|
|
|
sizeof (trim_state), 1, &trim_state, tx));
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update the on-disk state of a manual TRIM. This is called to request
|
|
|
|
* that a TRIM be started/suspended/canceled, or to change one of the
|
|
|
|
* TRIM options (partial, secure, rate).
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_change_state(vdev_t *vd, vdev_trim_state_t new_state,
|
|
|
|
uint64_t rate, boolean_t partial, boolean_t secure)
|
|
|
|
{
|
|
|
|
ASSERT(MUTEX_HELD(&vd->vdev_trim_lock));
|
|
|
|
spa_t *spa = vd->vdev_spa;
|
|
|
|
|
|
|
|
if (new_state == vd->vdev_trim_state)
|
|
|
|
return;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Copy the vd's guid, this will be freed by the sync task.
|
|
|
|
*/
|
|
|
|
uint64_t *guid = kmem_zalloc(sizeof (uint64_t), KM_SLEEP);
|
|
|
|
*guid = vd->vdev_guid;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we're suspending, then preserve the original start time.
|
|
|
|
*/
|
|
|
|
if (vd->vdev_trim_state != VDEV_TRIM_SUSPENDED) {
|
|
|
|
vd->vdev_trim_action_time = gethrestime_sec();
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we're activating, then preserve the requested rate and trim
|
|
|
|
* method. Setting the last offset and rate to UINT64_MAX is used
|
|
|
|
* as a sentinel to indicate they should be reset to default values.
|
|
|
|
*/
|
|
|
|
if (new_state == VDEV_TRIM_ACTIVE) {
|
|
|
|
if (vd->vdev_trim_state == VDEV_TRIM_COMPLETE ||
|
|
|
|
vd->vdev_trim_state == VDEV_TRIM_CANCELED) {
|
|
|
|
vd->vdev_trim_last_offset = UINT64_MAX;
|
|
|
|
vd->vdev_trim_rate = UINT64_MAX;
|
|
|
|
vd->vdev_trim_partial = UINT64_MAX;
|
|
|
|
vd->vdev_trim_secure = UINT64_MAX;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rate != 0)
|
|
|
|
vd->vdev_trim_rate = rate;
|
|
|
|
|
|
|
|
if (partial != 0)
|
|
|
|
vd->vdev_trim_partial = partial;
|
|
|
|
|
|
|
|
if (secure != 0)
|
|
|
|
vd->vdev_trim_secure = secure;
|
|
|
|
}
|
|
|
|
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
vdev_trim_state_t old_state = vd->vdev_trim_state;
|
|
|
|
boolean_t resumed = (old_state == VDEV_TRIM_SUSPENDED);
|
2019-03-29 19:13:20 +03:00
|
|
|
vd->vdev_trim_state = new_state;
|
|
|
|
|
|
|
|
dmu_tx_t *tx = dmu_tx_create_dd(spa_get_dsl(spa)->dp_mos_dir);
|
|
|
|
VERIFY0(dmu_tx_assign(tx, TXG_WAIT));
|
|
|
|
dsl_sync_task_nowait(spa_get_dsl(spa), vdev_trim_zap_update_sync,
|
2020-09-04 20:29:39 +03:00
|
|
|
guid, tx);
|
2019-03-29 19:13:20 +03:00
|
|
|
|
|
|
|
switch (new_state) {
|
|
|
|
case VDEV_TRIM_ACTIVE:
|
|
|
|
spa_event_notify(spa, vd, NULL,
|
|
|
|
resumed ? ESC_ZFS_TRIM_RESUME : ESC_ZFS_TRIM_START);
|
|
|
|
spa_history_log_internal(spa, "trim", tx,
|
|
|
|
"vdev=%s activated", vd->vdev_path);
|
|
|
|
break;
|
|
|
|
case VDEV_TRIM_SUSPENDED:
|
|
|
|
spa_event_notify(spa, vd, NULL, ESC_ZFS_TRIM_SUSPEND);
|
|
|
|
spa_history_log_internal(spa, "trim", tx,
|
|
|
|
"vdev=%s suspended", vd->vdev_path);
|
|
|
|
break;
|
|
|
|
case VDEV_TRIM_CANCELED:
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
if (old_state == VDEV_TRIM_ACTIVE ||
|
|
|
|
old_state == VDEV_TRIM_SUSPENDED) {
|
|
|
|
spa_event_notify(spa, vd, NULL, ESC_ZFS_TRIM_CANCEL);
|
|
|
|
spa_history_log_internal(spa, "trim", tx,
|
|
|
|
"vdev=%s canceled", vd->vdev_path);
|
|
|
|
}
|
2019-03-29 19:13:20 +03:00
|
|
|
break;
|
|
|
|
case VDEV_TRIM_COMPLETE:
|
|
|
|
spa_event_notify(spa, vd, NULL, ESC_ZFS_TRIM_FINISH);
|
|
|
|
spa_history_log_internal(spa, "trim", tx,
|
|
|
|
"vdev=%s complete", vd->vdev_path);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
panic("invalid state %llu", (unsigned long long)new_state);
|
|
|
|
}
|
|
|
|
|
|
|
|
dmu_tx_commit(tx);
|
2020-03-05 02:07:11 +03:00
|
|
|
|
|
|
|
if (new_state != VDEV_TRIM_ACTIVE)
|
|
|
|
spa_notify_waiters(spa);
|
2019-03-29 19:13:20 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The zio_done_func_t done callback for each manual TRIM issued. It is
|
|
|
|
* responsible for updating the TRIM stats, reissuing failed TRIM I/Os,
|
|
|
|
* and limiting the number of in flight TRIM I/Os.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_cb(zio_t *zio)
|
|
|
|
{
|
|
|
|
vdev_t *vd = zio->io_vd;
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_trim_io_lock);
|
|
|
|
if (zio->io_error == ENXIO && !vdev_writeable(vd)) {
|
|
|
|
/*
|
|
|
|
* The I/O failed because the vdev was unavailable; roll the
|
|
|
|
* last offset back. (This works because spa_sync waits on
|
|
|
|
* spa_txg_zio before it runs sync tasks.)
|
|
|
|
*/
|
|
|
|
uint64_t *offset =
|
|
|
|
&vd->vdev_trim_offset[zio->io_txg & TXG_MASK];
|
|
|
|
*offset = MIN(*offset, zio->io_offset);
|
|
|
|
} else {
|
|
|
|
if (zio->io_error != 0) {
|
|
|
|
vd->vdev_stat.vs_trim_errors++;
|
|
|
|
spa_iostats_trim_add(vd->vdev_spa, TRIM_TYPE_MANUAL,
|
|
|
|
0, 0, 0, 0, 1, zio->io_orig_size);
|
|
|
|
} else {
|
|
|
|
spa_iostats_trim_add(vd->vdev_spa, TRIM_TYPE_MANUAL,
|
|
|
|
1, zio->io_orig_size, 0, 0, 0, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
vd->vdev_trim_bytes_done += zio->io_orig_size;
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT3U(vd->vdev_trim_inflight[TRIM_TYPE_MANUAL], >, 0);
|
|
|
|
vd->vdev_trim_inflight[TRIM_TYPE_MANUAL]--;
|
|
|
|
cv_broadcast(&vd->vdev_trim_io_cv);
|
|
|
|
mutex_exit(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
spa_config_exit(vd->vdev_spa, SCL_STATE_ALL, vd);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The zio_done_func_t done callback for each automatic TRIM issued. It
|
|
|
|
* is responsible for updating the TRIM stats and limiting the number of
|
|
|
|
* in flight TRIM I/Os. Automatic TRIM I/Os are best effort and are
|
|
|
|
* never reissued on failure.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_autotrim_cb(zio_t *zio)
|
|
|
|
{
|
|
|
|
vdev_t *vd = zio->io_vd;
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
if (zio->io_error != 0) {
|
|
|
|
vd->vdev_stat.vs_trim_errors++;
|
|
|
|
spa_iostats_trim_add(vd->vdev_spa, TRIM_TYPE_AUTO,
|
|
|
|
0, 0, 0, 0, 1, zio->io_orig_size);
|
|
|
|
} else {
|
|
|
|
spa_iostats_trim_add(vd->vdev_spa, TRIM_TYPE_AUTO,
|
|
|
|
1, zio->io_orig_size, 0, 0, 0, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT3U(vd->vdev_trim_inflight[TRIM_TYPE_AUTO], >, 0);
|
|
|
|
vd->vdev_trim_inflight[TRIM_TYPE_AUTO]--;
|
|
|
|
cv_broadcast(&vd->vdev_trim_io_cv);
|
|
|
|
mutex_exit(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
spa_config_exit(vd->vdev_spa, SCL_STATE_ALL, vd);
|
|
|
|
}
|
|
|
|
|
2020-06-09 20:15:08 +03:00
|
|
|
/*
|
|
|
|
* The zio_done_func_t done callback for each TRIM issued via
|
|
|
|
* vdev_trim_simple(). It is responsible for updating the TRIM stats and
|
|
|
|
* limiting the number of in flight TRIM I/Os. Simple TRIM I/Os are best
|
|
|
|
* effort and are never reissued on failure.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_simple_cb(zio_t *zio)
|
|
|
|
{
|
|
|
|
vdev_t *vd = zio->io_vd;
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
if (zio->io_error != 0) {
|
|
|
|
vd->vdev_stat.vs_trim_errors++;
|
|
|
|
spa_iostats_trim_add(vd->vdev_spa, TRIM_TYPE_SIMPLE,
|
|
|
|
0, 0, 0, 0, 1, zio->io_orig_size);
|
|
|
|
} else {
|
|
|
|
spa_iostats_trim_add(vd->vdev_spa, TRIM_TYPE_SIMPLE,
|
|
|
|
1, zio->io_orig_size, 0, 0, 0, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT3U(vd->vdev_trim_inflight[TRIM_TYPE_SIMPLE], >, 0);
|
|
|
|
vd->vdev_trim_inflight[TRIM_TYPE_SIMPLE]--;
|
|
|
|
cv_broadcast(&vd->vdev_trim_io_cv);
|
|
|
|
mutex_exit(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
spa_config_exit(vd->vdev_spa, SCL_STATE_ALL, vd);
|
|
|
|
}
|
2019-03-29 19:13:20 +03:00
|
|
|
/*
|
|
|
|
* Returns the average trim rate in bytes/sec for the ta->trim_vdev.
|
|
|
|
*/
|
|
|
|
static uint64_t
|
|
|
|
vdev_trim_calculate_rate(trim_args_t *ta)
|
|
|
|
{
|
|
|
|
return (ta->trim_bytes_done * 1000 /
|
|
|
|
(NSEC2MSEC(gethrtime() - ta->trim_start_time) + 1));
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Issues a physical TRIM and takes care of rate limiting (bytes/sec)
|
|
|
|
* and number of concurrent TRIM I/Os.
|
|
|
|
*/
|
|
|
|
static int
|
|
|
|
vdev_trim_range(trim_args_t *ta, uint64_t start, uint64_t size)
|
|
|
|
{
|
|
|
|
vdev_t *vd = ta->trim_vdev;
|
|
|
|
spa_t *spa = vd->vdev_spa;
|
2020-06-09 20:15:08 +03:00
|
|
|
void *cb;
|
2019-03-29 19:13:20 +03:00
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Limit manual TRIM I/Os to the requested rate. This does not
|
|
|
|
* apply to automatic TRIM since no per vdev rate can be specified.
|
|
|
|
*/
|
|
|
|
if (ta->trim_type == TRIM_TYPE_MANUAL) {
|
|
|
|
while (vd->vdev_trim_rate != 0 && !vdev_trim_should_stop(vd) &&
|
|
|
|
vdev_trim_calculate_rate(ta) > vd->vdev_trim_rate) {
|
2020-09-04 06:04:09 +03:00
|
|
|
cv_timedwait_idle(&vd->vdev_trim_io_cv,
|
2019-03-29 19:13:20 +03:00
|
|
|
&vd->vdev_trim_io_lock, ddi_get_lbolt() +
|
|
|
|
MSEC_TO_TICK(10));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
ta->trim_bytes_done += size;
|
|
|
|
|
|
|
|
/* Limit in flight trimming I/Os */
|
2020-06-09 20:15:08 +03:00
|
|
|
while (vd->vdev_trim_inflight[0] + vd->vdev_trim_inflight[1] +
|
|
|
|
vd->vdev_trim_inflight[2] >= zfs_trim_queue_limit) {
|
2019-03-29 19:13:20 +03:00
|
|
|
cv_wait(&vd->vdev_trim_io_cv, &vd->vdev_trim_io_lock);
|
|
|
|
}
|
|
|
|
vd->vdev_trim_inflight[ta->trim_type]++;
|
|
|
|
mutex_exit(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
dmu_tx_t *tx = dmu_tx_create_dd(spa_get_dsl(spa)->dp_mos_dir);
|
|
|
|
VERIFY0(dmu_tx_assign(tx, TXG_WAIT));
|
|
|
|
uint64_t txg = dmu_tx_get_txg(tx);
|
|
|
|
|
|
|
|
spa_config_enter(spa, SCL_STATE_ALL, vd, RW_READER);
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
|
|
|
|
|
|
|
if (ta->trim_type == TRIM_TYPE_MANUAL &&
|
|
|
|
vd->vdev_trim_offset[txg & TXG_MASK] == 0) {
|
|
|
|
uint64_t *guid = kmem_zalloc(sizeof (uint64_t), KM_SLEEP);
|
|
|
|
*guid = vd->vdev_guid;
|
|
|
|
|
|
|
|
/* This is the first write of this txg. */
|
|
|
|
dsl_sync_task_nowait(spa_get_dsl(spa),
|
2020-09-04 20:29:39 +03:00
|
|
|
vdev_trim_zap_update_sync, guid, tx);
|
2019-03-29 19:13:20 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We know the vdev_t will still be around since all consumers of
|
|
|
|
* vdev_free must stop the trimming first.
|
|
|
|
*/
|
|
|
|
if ((ta->trim_type == TRIM_TYPE_MANUAL &&
|
|
|
|
vdev_trim_should_stop(vd)) ||
|
|
|
|
(ta->trim_type == TRIM_TYPE_AUTO &&
|
|
|
|
vdev_autotrim_should_stop(vd->vdev_top))) {
|
|
|
|
mutex_enter(&vd->vdev_trim_io_lock);
|
|
|
|
vd->vdev_trim_inflight[ta->trim_type]--;
|
|
|
|
mutex_exit(&vd->vdev_trim_io_lock);
|
|
|
|
spa_config_exit(vd->vdev_spa, SCL_STATE_ALL, vd);
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
dmu_tx_commit(tx);
|
|
|
|
return (SET_ERROR(EINTR));
|
|
|
|
}
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
|
|
|
|
if (ta->trim_type == TRIM_TYPE_MANUAL)
|
|
|
|
vd->vdev_trim_offset[txg & TXG_MASK] = start + size;
|
|
|
|
|
2020-06-09 20:15:08 +03:00
|
|
|
if (ta->trim_type == TRIM_TYPE_MANUAL) {
|
|
|
|
cb = vdev_trim_cb;
|
|
|
|
} else if (ta->trim_type == TRIM_TYPE_AUTO) {
|
|
|
|
cb = vdev_autotrim_cb;
|
|
|
|
} else {
|
|
|
|
cb = vdev_trim_simple_cb;
|
|
|
|
}
|
|
|
|
|
2019-03-29 19:13:20 +03:00
|
|
|
zio_nowait(zio_trim(spa->spa_txg_zio[txg & TXG_MASK], vd,
|
2020-06-09 20:15:08 +03:00
|
|
|
start, size, cb, NULL, ZIO_PRIORITY_TRIM, ZIO_FLAG_CANFAIL,
|
|
|
|
ta->trim_flags));
|
2019-03-29 19:13:20 +03:00
|
|
|
/* vdev_trim_cb and vdev_autotrim_cb release SCL_STATE_ALL */
|
|
|
|
|
|
|
|
dmu_tx_commit(tx);
|
|
|
|
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Issues TRIM I/Os for all ranges in the provided ta->trim_tree range tree.
|
|
|
|
* Additional parameters describing how the TRIM should be performed must
|
|
|
|
* be set in the trim_args structure. See the trim_args definition for
|
|
|
|
* additional information.
|
|
|
|
*/
|
|
|
|
static int
|
|
|
|
vdev_trim_ranges(trim_args_t *ta)
|
|
|
|
{
|
|
|
|
vdev_t *vd = ta->trim_vdev;
|
Reduce loaded range tree memory usage
This patch implements a new tree structure for ZFS, and uses it to
store range trees more efficiently.
The new structure is approximately a B-tree, though there are some
small differences from the usual characterizations. The tree has core
nodes and leaf nodes; each contain data elements, which the elements
in the core nodes acting as separators between its children. The
difference between core and leaf nodes is that the core nodes have an
array of children, while leaf nodes don't. Every node in the tree may
be only partially full; in most cases, they are all at least 50% full
(in terms of element count) except for the root node, which can be
less full. Underfull nodes will steal from their neighbors or merge to
remain full enough, while overfull nodes will split in two. The data
elements are contained in tree-controlled buffers; they are copied
into these on insertion, and overwritten on deletion. This means that
the elements are not independently allocated, which reduces overhead,
but also means they can't be shared between trees (and also that
pointers to them are only valid until a side-effectful tree operation
occurs). The overhead varies based on how dense the tree is, but is
usually on the order of about 50% of the element size; the per-node
overheads are very small, and so don't make a significant difference.
The trees can accept arbitrary records; they accept a size and a
comparator to allow them to be used for a variety of purposes.
The new trees replace the AVL trees used in the range trees today.
Currently, the range_seg_t structure contains three 8 byte integers
of payload and two 24 byte avl_tree_node_ts to handle its storage in
both an offset-sorted tree and a size-sorted tree (total size: 64
bytes). In the new model, the range seg structures are usually two 4
byte integers, but a separate one needs to exist for the size-sorted
and offset-sorted tree. Between the raw size, the 50% overhead, and
the double storage, the new btrees are expected to use 8*1.5*2 = 24
bytes per record, or 33.3% as much memory as the AVL trees (this is
for the purposes of storing metaslab range trees; for other purposes,
like scrubs, they use ~50% as much memory).
We reduced the size of the payload in the range segments by teaching
range trees about starting offsets and shifts; since metaslabs have a
fixed starting offset, and they all operate in terms of disk sectors,
we can store the ranges using 4-byte integers as long as the size of
the metaslab divided by the sector size is less than 2^32. For 512-byte
sectors, this is a 2^41 (or 2TB) metaslab, which with the default
settings corresponds to a 256PB disk. 4k sector disks can handle
metaslabs up to 2^46 bytes, or 2^63 byte disks. Since we do not
anticipate disks of this size in the near future, there should be
almost no cases where metaslabs need 64-byte integers to store their
ranges. We do still have the capability to store 64-byte integer ranges
to account for cases where we are storing per-vdev (or per-dnode) trees,
which could reasonably go above the limits discussed. We also do not
store fill information in the compact version of the node, since it
is only used for sorted scrub.
We also optimized the metaslab loading process in various other ways
to offset some inefficiencies in the btree model. While individual
operations (find, insert, remove_from) are faster for the btree than
they are for the avl tree, remove usually requires a find operation,
while in the AVL tree model the element itself suffices. Some clever
changes actually caused an overall speedup in metaslab loading; we use
approximately 40% less cpu to load metaslabs in our tests on Illumos.
Another memory and performance optimization was achieved by changing
what is stored in the size-sorted trees. When a disk is heavily
fragmented, the df algorithm used by default in ZFS will almost always
find a number of small regions in its initial cursor-based search; it
will usually only fall back to the size-sorted tree to find larger
regions. If we increase the size of the cursor-based search slightly,
and don't store segments that are smaller than a tunable size floor
in the size-sorted tree, we can further cut memory usage down to
below 20% of what the AVL trees store. This also results in further
reductions in CPU time spent loading metaslabs.
The 16KiB size floor was chosen because it results in substantial memory
usage reduction while not usually resulting in situations where we can't
find an appropriate chunk with the cursor and are forced to use an
oversized chunk from the size-sorted tree. In addition, even if we do
have to use an oversized chunk from the size-sorted tree, the chunk
would be too small to use for ZIL allocations, so it isn't as big of a
loss as it might otherwise be. And often, more small allocations will
follow the initial one, and the cursor search will now find the
remainder of the chunk we didn't use all of and use it for subsequent
allocations. Practical testing has shown little or no change in
fragmentation as a result of this change.
If the size-sorted tree becomes empty while the offset sorted one still
has entries, it will load all the entries from the offset sorted tree
and disregard the size floor until it is unloaded again. This operation
occurs rarely with the default setting, only on incredibly thoroughly
fragmented pools.
There are some other small changes to zdb to teach it to handle btrees,
but nothing major.
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed by: Sebastien Roy seb@delphix.com
Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9181
2019-10-09 20:36:03 +03:00
|
|
|
zfs_btree_t *t = &ta->trim_tree->rt_root;
|
|
|
|
zfs_btree_index_t idx;
|
2019-03-29 19:13:20 +03:00
|
|
|
uint64_t extent_bytes_max = ta->trim_extent_bytes_max;
|
|
|
|
uint64_t extent_bytes_min = ta->trim_extent_bytes_min;
|
|
|
|
spa_t *spa = vd->vdev_spa;
|
|
|
|
|
|
|
|
ta->trim_start_time = gethrtime();
|
|
|
|
ta->trim_bytes_done = 0;
|
|
|
|
|
Reduce loaded range tree memory usage
This patch implements a new tree structure for ZFS, and uses it to
store range trees more efficiently.
The new structure is approximately a B-tree, though there are some
small differences from the usual characterizations. The tree has core
nodes and leaf nodes; each contain data elements, which the elements
in the core nodes acting as separators between its children. The
difference between core and leaf nodes is that the core nodes have an
array of children, while leaf nodes don't. Every node in the tree may
be only partially full; in most cases, they are all at least 50% full
(in terms of element count) except for the root node, which can be
less full. Underfull nodes will steal from their neighbors or merge to
remain full enough, while overfull nodes will split in two. The data
elements are contained in tree-controlled buffers; they are copied
into these on insertion, and overwritten on deletion. This means that
the elements are not independently allocated, which reduces overhead,
but also means they can't be shared between trees (and also that
pointers to them are only valid until a side-effectful tree operation
occurs). The overhead varies based on how dense the tree is, but is
usually on the order of about 50% of the element size; the per-node
overheads are very small, and so don't make a significant difference.
The trees can accept arbitrary records; they accept a size and a
comparator to allow them to be used for a variety of purposes.
The new trees replace the AVL trees used in the range trees today.
Currently, the range_seg_t structure contains three 8 byte integers
of payload and two 24 byte avl_tree_node_ts to handle its storage in
both an offset-sorted tree and a size-sorted tree (total size: 64
bytes). In the new model, the range seg structures are usually two 4
byte integers, but a separate one needs to exist for the size-sorted
and offset-sorted tree. Between the raw size, the 50% overhead, and
the double storage, the new btrees are expected to use 8*1.5*2 = 24
bytes per record, or 33.3% as much memory as the AVL trees (this is
for the purposes of storing metaslab range trees; for other purposes,
like scrubs, they use ~50% as much memory).
We reduced the size of the payload in the range segments by teaching
range trees about starting offsets and shifts; since metaslabs have a
fixed starting offset, and they all operate in terms of disk sectors,
we can store the ranges using 4-byte integers as long as the size of
the metaslab divided by the sector size is less than 2^32. For 512-byte
sectors, this is a 2^41 (or 2TB) metaslab, which with the default
settings corresponds to a 256PB disk. 4k sector disks can handle
metaslabs up to 2^46 bytes, or 2^63 byte disks. Since we do not
anticipate disks of this size in the near future, there should be
almost no cases where metaslabs need 64-byte integers to store their
ranges. We do still have the capability to store 64-byte integer ranges
to account for cases where we are storing per-vdev (or per-dnode) trees,
which could reasonably go above the limits discussed. We also do not
store fill information in the compact version of the node, since it
is only used for sorted scrub.
We also optimized the metaslab loading process in various other ways
to offset some inefficiencies in the btree model. While individual
operations (find, insert, remove_from) are faster for the btree than
they are for the avl tree, remove usually requires a find operation,
while in the AVL tree model the element itself suffices. Some clever
changes actually caused an overall speedup in metaslab loading; we use
approximately 40% less cpu to load metaslabs in our tests on Illumos.
Another memory and performance optimization was achieved by changing
what is stored in the size-sorted trees. When a disk is heavily
fragmented, the df algorithm used by default in ZFS will almost always
find a number of small regions in its initial cursor-based search; it
will usually only fall back to the size-sorted tree to find larger
regions. If we increase the size of the cursor-based search slightly,
and don't store segments that are smaller than a tunable size floor
in the size-sorted tree, we can further cut memory usage down to
below 20% of what the AVL trees store. This also results in further
reductions in CPU time spent loading metaslabs.
The 16KiB size floor was chosen because it results in substantial memory
usage reduction while not usually resulting in situations where we can't
find an appropriate chunk with the cursor and are forced to use an
oversized chunk from the size-sorted tree. In addition, even if we do
have to use an oversized chunk from the size-sorted tree, the chunk
would be too small to use for ZIL allocations, so it isn't as big of a
loss as it might otherwise be. And often, more small allocations will
follow the initial one, and the cursor search will now find the
remainder of the chunk we didn't use all of and use it for subsequent
allocations. Practical testing has shown little or no change in
fragmentation as a result of this change.
If the size-sorted tree becomes empty while the offset sorted one still
has entries, it will load all the entries from the offset sorted tree
and disregard the size floor until it is unloaded again. This operation
occurs rarely with the default setting, only on incredibly thoroughly
fragmented pools.
There are some other small changes to zdb to teach it to handle btrees,
but nothing major.
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed by: Sebastien Roy seb@delphix.com
Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9181
2019-10-09 20:36:03 +03:00
|
|
|
for (range_seg_t *rs = zfs_btree_first(t, &idx); rs != NULL;
|
|
|
|
rs = zfs_btree_next(t, &idx, &idx)) {
|
|
|
|
uint64_t size = rs_get_end(rs, ta->trim_tree) - rs_get_start(rs,
|
|
|
|
ta->trim_tree);
|
2019-03-29 19:13:20 +03:00
|
|
|
|
|
|
|
if (extent_bytes_min && size < extent_bytes_min) {
|
|
|
|
spa_iostats_trim_add(spa, ta->trim_type,
|
|
|
|
0, 0, 1, size, 0, 0);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Split range into legally-sized physical chunks */
|
|
|
|
uint64_t writes_required = ((size - 1) / extent_bytes_max) + 1;
|
|
|
|
|
|
|
|
for (uint64_t w = 0; w < writes_required; w++) {
|
|
|
|
int error;
|
|
|
|
|
|
|
|
error = vdev_trim_range(ta, VDEV_LABEL_START_SIZE +
|
Reduce loaded range tree memory usage
This patch implements a new tree structure for ZFS, and uses it to
store range trees more efficiently.
The new structure is approximately a B-tree, though there are some
small differences from the usual characterizations. The tree has core
nodes and leaf nodes; each contain data elements, which the elements
in the core nodes acting as separators between its children. The
difference between core and leaf nodes is that the core nodes have an
array of children, while leaf nodes don't. Every node in the tree may
be only partially full; in most cases, they are all at least 50% full
(in terms of element count) except for the root node, which can be
less full. Underfull nodes will steal from their neighbors or merge to
remain full enough, while overfull nodes will split in two. The data
elements are contained in tree-controlled buffers; they are copied
into these on insertion, and overwritten on deletion. This means that
the elements are not independently allocated, which reduces overhead,
but also means they can't be shared between trees (and also that
pointers to them are only valid until a side-effectful tree operation
occurs). The overhead varies based on how dense the tree is, but is
usually on the order of about 50% of the element size; the per-node
overheads are very small, and so don't make a significant difference.
The trees can accept arbitrary records; they accept a size and a
comparator to allow them to be used for a variety of purposes.
The new trees replace the AVL trees used in the range trees today.
Currently, the range_seg_t structure contains three 8 byte integers
of payload and two 24 byte avl_tree_node_ts to handle its storage in
both an offset-sorted tree and a size-sorted tree (total size: 64
bytes). In the new model, the range seg structures are usually two 4
byte integers, but a separate one needs to exist for the size-sorted
and offset-sorted tree. Between the raw size, the 50% overhead, and
the double storage, the new btrees are expected to use 8*1.5*2 = 24
bytes per record, or 33.3% as much memory as the AVL trees (this is
for the purposes of storing metaslab range trees; for other purposes,
like scrubs, they use ~50% as much memory).
We reduced the size of the payload in the range segments by teaching
range trees about starting offsets and shifts; since metaslabs have a
fixed starting offset, and they all operate in terms of disk sectors,
we can store the ranges using 4-byte integers as long as the size of
the metaslab divided by the sector size is less than 2^32. For 512-byte
sectors, this is a 2^41 (or 2TB) metaslab, which with the default
settings corresponds to a 256PB disk. 4k sector disks can handle
metaslabs up to 2^46 bytes, or 2^63 byte disks. Since we do not
anticipate disks of this size in the near future, there should be
almost no cases where metaslabs need 64-byte integers to store their
ranges. We do still have the capability to store 64-byte integer ranges
to account for cases where we are storing per-vdev (or per-dnode) trees,
which could reasonably go above the limits discussed. We also do not
store fill information in the compact version of the node, since it
is only used for sorted scrub.
We also optimized the metaslab loading process in various other ways
to offset some inefficiencies in the btree model. While individual
operations (find, insert, remove_from) are faster for the btree than
they are for the avl tree, remove usually requires a find operation,
while in the AVL tree model the element itself suffices. Some clever
changes actually caused an overall speedup in metaslab loading; we use
approximately 40% less cpu to load metaslabs in our tests on Illumos.
Another memory and performance optimization was achieved by changing
what is stored in the size-sorted trees. When a disk is heavily
fragmented, the df algorithm used by default in ZFS will almost always
find a number of small regions in its initial cursor-based search; it
will usually only fall back to the size-sorted tree to find larger
regions. If we increase the size of the cursor-based search slightly,
and don't store segments that are smaller than a tunable size floor
in the size-sorted tree, we can further cut memory usage down to
below 20% of what the AVL trees store. This also results in further
reductions in CPU time spent loading metaslabs.
The 16KiB size floor was chosen because it results in substantial memory
usage reduction while not usually resulting in situations where we can't
find an appropriate chunk with the cursor and are forced to use an
oversized chunk from the size-sorted tree. In addition, even if we do
have to use an oversized chunk from the size-sorted tree, the chunk
would be too small to use for ZIL allocations, so it isn't as big of a
loss as it might otherwise be. And often, more small allocations will
follow the initial one, and the cursor search will now find the
remainder of the chunk we didn't use all of and use it for subsequent
allocations. Practical testing has shown little or no change in
fragmentation as a result of this change.
If the size-sorted tree becomes empty while the offset sorted one still
has entries, it will load all the entries from the offset sorted tree
and disregard the size floor until it is unloaded again. This operation
occurs rarely with the default setting, only on incredibly thoroughly
fragmented pools.
There are some other small changes to zdb to teach it to handle btrees,
but nothing major.
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed by: Sebastien Roy seb@delphix.com
Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9181
2019-10-09 20:36:03 +03:00
|
|
|
rs_get_start(rs, ta->trim_tree) +
|
|
|
|
(w *extent_bytes_max), MIN(size -
|
|
|
|
(w * extent_bytes_max), extent_bytes_max));
|
2019-03-29 19:13:20 +03:00
|
|
|
if (error != 0) {
|
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return (0);
|
|
|
|
}
|
|
|
|
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
static void
|
|
|
|
vdev_trim_xlate_last_rs_end(void *arg, range_seg64_t *physical_rs)
|
|
|
|
{
|
|
|
|
uint64_t *last_rs_end = (uint64_t *)arg;
|
|
|
|
|
|
|
|
if (physical_rs->rs_end > *last_rs_end)
|
|
|
|
*last_rs_end = physical_rs->rs_end;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
vdev_trim_xlate_progress(void *arg, range_seg64_t *physical_rs)
|
|
|
|
{
|
|
|
|
vdev_t *vd = (vdev_t *)arg;
|
|
|
|
|
|
|
|
uint64_t size = physical_rs->rs_end - physical_rs->rs_start;
|
|
|
|
vd->vdev_trim_bytes_est += size;
|
|
|
|
|
|
|
|
if (vd->vdev_trim_last_offset >= physical_rs->rs_end) {
|
|
|
|
vd->vdev_trim_bytes_done += size;
|
|
|
|
} else if (vd->vdev_trim_last_offset > physical_rs->rs_start &&
|
|
|
|
vd->vdev_trim_last_offset <= physical_rs->rs_end) {
|
|
|
|
vd->vdev_trim_bytes_done +=
|
|
|
|
vd->vdev_trim_last_offset - physical_rs->rs_start;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-03-29 19:13:20 +03:00
|
|
|
/*
|
|
|
|
* Calculates the completion percentage of a manual TRIM.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_calculate_progress(vdev_t *vd)
|
|
|
|
{
|
|
|
|
ASSERT(spa_config_held(vd->vdev_spa, SCL_CONFIG, RW_READER) ||
|
|
|
|
spa_config_held(vd->vdev_spa, SCL_CONFIG, RW_WRITER));
|
|
|
|
ASSERT(vd->vdev_leaf_zap != 0);
|
|
|
|
|
|
|
|
vd->vdev_trim_bytes_est = 0;
|
|
|
|
vd->vdev_trim_bytes_done = 0;
|
|
|
|
|
|
|
|
for (uint64_t i = 0; i < vd->vdev_top->vdev_ms_count; i++) {
|
|
|
|
metaslab_t *msp = vd->vdev_top->vdev_ms[i];
|
|
|
|
mutex_enter(&msp->ms_lock);
|
|
|
|
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
uint64_t ms_free = (msp->ms_size -
|
|
|
|
metaslab_allocated_space(msp)) /
|
|
|
|
vdev_get_ndisks(vd->vdev_top);
|
2019-03-29 19:13:20 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Convert the metaslab range to a physical range
|
|
|
|
* on our vdev. We use this to determine if we are
|
|
|
|
* in the middle of this metaslab range.
|
|
|
|
*/
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
range_seg64_t logical_rs, physical_rs, remain_rs;
|
2019-03-29 19:13:20 +03:00
|
|
|
logical_rs.rs_start = msp->ms_start;
|
|
|
|
logical_rs.rs_end = msp->ms_start + msp->ms_size;
|
|
|
|
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
/* Metaslab space after this offset has not been trimmed. */
|
|
|
|
vdev_xlate(vd, &logical_rs, &physical_rs, &remain_rs);
|
2019-03-29 19:13:20 +03:00
|
|
|
if (vd->vdev_trim_last_offset <= physical_rs.rs_start) {
|
|
|
|
vd->vdev_trim_bytes_est += ms_free;
|
|
|
|
mutex_exit(&msp->ms_lock);
|
|
|
|
continue;
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Metaslab space before this offset has been trimmed */
|
|
|
|
uint64_t last_rs_end = physical_rs.rs_end;
|
|
|
|
if (!vdev_xlate_is_empty(&remain_rs)) {
|
|
|
|
vdev_xlate_walk(vd, &remain_rs,
|
|
|
|
vdev_trim_xlate_last_rs_end, &last_rs_end);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (vd->vdev_trim_last_offset > last_rs_end) {
|
2019-03-29 19:13:20 +03:00
|
|
|
vd->vdev_trim_bytes_done += ms_free;
|
|
|
|
vd->vdev_trim_bytes_est += ms_free;
|
|
|
|
mutex_exit(&msp->ms_lock);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we get here, we're in the middle of trimming this
|
|
|
|
* metaslab. Load it and walk the free tree for more
|
|
|
|
* accurate progress estimation.
|
|
|
|
*/
|
|
|
|
VERIFY0(metaslab_load(msp));
|
|
|
|
|
Reduce loaded range tree memory usage
This patch implements a new tree structure for ZFS, and uses it to
store range trees more efficiently.
The new structure is approximately a B-tree, though there are some
small differences from the usual characterizations. The tree has core
nodes and leaf nodes; each contain data elements, which the elements
in the core nodes acting as separators between its children. The
difference between core and leaf nodes is that the core nodes have an
array of children, while leaf nodes don't. Every node in the tree may
be only partially full; in most cases, they are all at least 50% full
(in terms of element count) except for the root node, which can be
less full. Underfull nodes will steal from their neighbors or merge to
remain full enough, while overfull nodes will split in two. The data
elements are contained in tree-controlled buffers; they are copied
into these on insertion, and overwritten on deletion. This means that
the elements are not independently allocated, which reduces overhead,
but also means they can't be shared between trees (and also that
pointers to them are only valid until a side-effectful tree operation
occurs). The overhead varies based on how dense the tree is, but is
usually on the order of about 50% of the element size; the per-node
overheads are very small, and so don't make a significant difference.
The trees can accept arbitrary records; they accept a size and a
comparator to allow them to be used for a variety of purposes.
The new trees replace the AVL trees used in the range trees today.
Currently, the range_seg_t structure contains three 8 byte integers
of payload and two 24 byte avl_tree_node_ts to handle its storage in
both an offset-sorted tree and a size-sorted tree (total size: 64
bytes). In the new model, the range seg structures are usually two 4
byte integers, but a separate one needs to exist for the size-sorted
and offset-sorted tree. Between the raw size, the 50% overhead, and
the double storage, the new btrees are expected to use 8*1.5*2 = 24
bytes per record, or 33.3% as much memory as the AVL trees (this is
for the purposes of storing metaslab range trees; for other purposes,
like scrubs, they use ~50% as much memory).
We reduced the size of the payload in the range segments by teaching
range trees about starting offsets and shifts; since metaslabs have a
fixed starting offset, and they all operate in terms of disk sectors,
we can store the ranges using 4-byte integers as long as the size of
the metaslab divided by the sector size is less than 2^32. For 512-byte
sectors, this is a 2^41 (or 2TB) metaslab, which with the default
settings corresponds to a 256PB disk. 4k sector disks can handle
metaslabs up to 2^46 bytes, or 2^63 byte disks. Since we do not
anticipate disks of this size in the near future, there should be
almost no cases where metaslabs need 64-byte integers to store their
ranges. We do still have the capability to store 64-byte integer ranges
to account for cases where we are storing per-vdev (or per-dnode) trees,
which could reasonably go above the limits discussed. We also do not
store fill information in the compact version of the node, since it
is only used for sorted scrub.
We also optimized the metaslab loading process in various other ways
to offset some inefficiencies in the btree model. While individual
operations (find, insert, remove_from) are faster for the btree than
they are for the avl tree, remove usually requires a find operation,
while in the AVL tree model the element itself suffices. Some clever
changes actually caused an overall speedup in metaslab loading; we use
approximately 40% less cpu to load metaslabs in our tests on Illumos.
Another memory and performance optimization was achieved by changing
what is stored in the size-sorted trees. When a disk is heavily
fragmented, the df algorithm used by default in ZFS will almost always
find a number of small regions in its initial cursor-based search; it
will usually only fall back to the size-sorted tree to find larger
regions. If we increase the size of the cursor-based search slightly,
and don't store segments that are smaller than a tunable size floor
in the size-sorted tree, we can further cut memory usage down to
below 20% of what the AVL trees store. This also results in further
reductions in CPU time spent loading metaslabs.
The 16KiB size floor was chosen because it results in substantial memory
usage reduction while not usually resulting in situations where we can't
find an appropriate chunk with the cursor and are forced to use an
oversized chunk from the size-sorted tree. In addition, even if we do
have to use an oversized chunk from the size-sorted tree, the chunk
would be too small to use for ZIL allocations, so it isn't as big of a
loss as it might otherwise be. And often, more small allocations will
follow the initial one, and the cursor search will now find the
remainder of the chunk we didn't use all of and use it for subsequent
allocations. Practical testing has shown little or no change in
fragmentation as a result of this change.
If the size-sorted tree becomes empty while the offset sorted one still
has entries, it will load all the entries from the offset sorted tree
and disregard the size floor until it is unloaded again. This operation
occurs rarely with the default setting, only on incredibly thoroughly
fragmented pools.
There are some other small changes to zdb to teach it to handle btrees,
but nothing major.
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed by: Sebastien Roy seb@delphix.com
Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9181
2019-10-09 20:36:03 +03:00
|
|
|
range_tree_t *rt = msp->ms_allocatable;
|
|
|
|
zfs_btree_t *bt = &rt->rt_root;
|
|
|
|
zfs_btree_index_t idx;
|
|
|
|
for (range_seg_t *rs = zfs_btree_first(bt, &idx);
|
|
|
|
rs != NULL; rs = zfs_btree_next(bt, &idx, &idx)) {
|
|
|
|
logical_rs.rs_start = rs_get_start(rs, rt);
|
|
|
|
logical_rs.rs_end = rs_get_end(rs, rt);
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
|
|
|
|
vdev_xlate_walk(vd, &logical_rs,
|
|
|
|
vdev_trim_xlate_progress, vd);
|
2019-03-29 19:13:20 +03:00
|
|
|
}
|
|
|
|
mutex_exit(&msp->ms_lock);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Load from disk the vdev's manual TRIM information. This includes the
|
|
|
|
* state, progress, and options provided when initiating the manual TRIM.
|
|
|
|
*/
|
|
|
|
static int
|
|
|
|
vdev_trim_load(vdev_t *vd)
|
|
|
|
{
|
|
|
|
int err = 0;
|
|
|
|
ASSERT(spa_config_held(vd->vdev_spa, SCL_CONFIG, RW_READER) ||
|
|
|
|
spa_config_held(vd->vdev_spa, SCL_CONFIG, RW_WRITER));
|
|
|
|
ASSERT(vd->vdev_leaf_zap != 0);
|
|
|
|
|
|
|
|
if (vd->vdev_trim_state == VDEV_TRIM_ACTIVE ||
|
|
|
|
vd->vdev_trim_state == VDEV_TRIM_SUSPENDED) {
|
|
|
|
err = zap_lookup(vd->vdev_spa->spa_meta_objset,
|
|
|
|
vd->vdev_leaf_zap, VDEV_LEAF_ZAP_TRIM_LAST_OFFSET,
|
|
|
|
sizeof (vd->vdev_trim_last_offset), 1,
|
|
|
|
&vd->vdev_trim_last_offset);
|
|
|
|
if (err == ENOENT) {
|
|
|
|
vd->vdev_trim_last_offset = 0;
|
|
|
|
err = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (err == 0) {
|
|
|
|
err = zap_lookup(vd->vdev_spa->spa_meta_objset,
|
|
|
|
vd->vdev_leaf_zap, VDEV_LEAF_ZAP_TRIM_RATE,
|
|
|
|
sizeof (vd->vdev_trim_rate), 1,
|
|
|
|
&vd->vdev_trim_rate);
|
|
|
|
if (err == ENOENT) {
|
|
|
|
vd->vdev_trim_rate = 0;
|
|
|
|
err = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (err == 0) {
|
|
|
|
err = zap_lookup(vd->vdev_spa->spa_meta_objset,
|
|
|
|
vd->vdev_leaf_zap, VDEV_LEAF_ZAP_TRIM_PARTIAL,
|
|
|
|
sizeof (vd->vdev_trim_partial), 1,
|
|
|
|
&vd->vdev_trim_partial);
|
|
|
|
if (err == ENOENT) {
|
|
|
|
vd->vdev_trim_partial = 0;
|
|
|
|
err = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (err == 0) {
|
|
|
|
err = zap_lookup(vd->vdev_spa->spa_meta_objset,
|
|
|
|
vd->vdev_leaf_zap, VDEV_LEAF_ZAP_TRIM_SECURE,
|
|
|
|
sizeof (vd->vdev_trim_secure), 1,
|
|
|
|
&vd->vdev_trim_secure);
|
|
|
|
if (err == ENOENT) {
|
|
|
|
vd->vdev_trim_secure = 0;
|
|
|
|
err = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
vdev_trim_calculate_progress(vd);
|
|
|
|
|
|
|
|
return (err);
|
|
|
|
}
|
|
|
|
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
static void
|
|
|
|
vdev_trim_xlate_range_add(void *arg, range_seg64_t *physical_rs)
|
|
|
|
{
|
|
|
|
trim_args_t *ta = arg;
|
|
|
|
vdev_t *vd = ta->trim_vdev;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Only a manual trim will be traversing the vdev sequentially.
|
|
|
|
* For an auto trim all valid ranges should be added.
|
|
|
|
*/
|
|
|
|
if (ta->trim_type == TRIM_TYPE_MANUAL) {
|
|
|
|
|
|
|
|
/* Only add segments that we have not visited yet */
|
|
|
|
if (physical_rs->rs_end <= vd->vdev_trim_last_offset)
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* Pick up where we left off mid-range. */
|
|
|
|
if (vd->vdev_trim_last_offset > physical_rs->rs_start) {
|
|
|
|
ASSERT3U(physical_rs->rs_end, >,
|
|
|
|
vd->vdev_trim_last_offset);
|
|
|
|
physical_rs->rs_start = vd->vdev_trim_last_offset;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT3U(physical_rs->rs_end, >, physical_rs->rs_start);
|
|
|
|
|
|
|
|
range_tree_add(ta->trim_tree, physical_rs->rs_start,
|
|
|
|
physical_rs->rs_end - physical_rs->rs_start);
|
|
|
|
}
|
|
|
|
|
2019-03-29 19:13:20 +03:00
|
|
|
/*
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
* Convert the logical range into physical ranges and add them to the
|
2019-03-29 19:13:20 +03:00
|
|
|
* range tree passed in the trim_args_t.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_range_add(void *arg, uint64_t start, uint64_t size)
|
|
|
|
{
|
|
|
|
trim_args_t *ta = arg;
|
|
|
|
vdev_t *vd = ta->trim_vdev;
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
range_seg64_t logical_rs;
|
2019-03-29 19:13:20 +03:00
|
|
|
logical_rs.rs_start = start;
|
|
|
|
logical_rs.rs_end = start + size;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Every range to be trimmed must be part of ms_allocatable.
|
|
|
|
* When ZFS_DEBUG_TRIM is set load the metaslab to verify this
|
|
|
|
* is always the case.
|
|
|
|
*/
|
|
|
|
if (zfs_flags & ZFS_DEBUG_TRIM) {
|
|
|
|
metaslab_t *msp = ta->trim_msp;
|
|
|
|
VERIFY0(metaslab_load(msp));
|
|
|
|
VERIFY3B(msp->ms_loaded, ==, B_TRUE);
|
Reduce loaded range tree memory usage
This patch implements a new tree structure for ZFS, and uses it to
store range trees more efficiently.
The new structure is approximately a B-tree, though there are some
small differences from the usual characterizations. The tree has core
nodes and leaf nodes; each contain data elements, which the elements
in the core nodes acting as separators between its children. The
difference between core and leaf nodes is that the core nodes have an
array of children, while leaf nodes don't. Every node in the tree may
be only partially full; in most cases, they are all at least 50% full
(in terms of element count) except for the root node, which can be
less full. Underfull nodes will steal from their neighbors or merge to
remain full enough, while overfull nodes will split in two. The data
elements are contained in tree-controlled buffers; they are copied
into these on insertion, and overwritten on deletion. This means that
the elements are not independently allocated, which reduces overhead,
but also means they can't be shared between trees (and also that
pointers to them are only valid until a side-effectful tree operation
occurs). The overhead varies based on how dense the tree is, but is
usually on the order of about 50% of the element size; the per-node
overheads are very small, and so don't make a significant difference.
The trees can accept arbitrary records; they accept a size and a
comparator to allow them to be used for a variety of purposes.
The new trees replace the AVL trees used in the range trees today.
Currently, the range_seg_t structure contains three 8 byte integers
of payload and two 24 byte avl_tree_node_ts to handle its storage in
both an offset-sorted tree and a size-sorted tree (total size: 64
bytes). In the new model, the range seg structures are usually two 4
byte integers, but a separate one needs to exist for the size-sorted
and offset-sorted tree. Between the raw size, the 50% overhead, and
the double storage, the new btrees are expected to use 8*1.5*2 = 24
bytes per record, or 33.3% as much memory as the AVL trees (this is
for the purposes of storing metaslab range trees; for other purposes,
like scrubs, they use ~50% as much memory).
We reduced the size of the payload in the range segments by teaching
range trees about starting offsets and shifts; since metaslabs have a
fixed starting offset, and they all operate in terms of disk sectors,
we can store the ranges using 4-byte integers as long as the size of
the metaslab divided by the sector size is less than 2^32. For 512-byte
sectors, this is a 2^41 (or 2TB) metaslab, which with the default
settings corresponds to a 256PB disk. 4k sector disks can handle
metaslabs up to 2^46 bytes, or 2^63 byte disks. Since we do not
anticipate disks of this size in the near future, there should be
almost no cases where metaslabs need 64-byte integers to store their
ranges. We do still have the capability to store 64-byte integer ranges
to account for cases where we are storing per-vdev (or per-dnode) trees,
which could reasonably go above the limits discussed. We also do not
store fill information in the compact version of the node, since it
is only used for sorted scrub.
We also optimized the metaslab loading process in various other ways
to offset some inefficiencies in the btree model. While individual
operations (find, insert, remove_from) are faster for the btree than
they are for the avl tree, remove usually requires a find operation,
while in the AVL tree model the element itself suffices. Some clever
changes actually caused an overall speedup in metaslab loading; we use
approximately 40% less cpu to load metaslabs in our tests on Illumos.
Another memory and performance optimization was achieved by changing
what is stored in the size-sorted trees. When a disk is heavily
fragmented, the df algorithm used by default in ZFS will almost always
find a number of small regions in its initial cursor-based search; it
will usually only fall back to the size-sorted tree to find larger
regions. If we increase the size of the cursor-based search slightly,
and don't store segments that are smaller than a tunable size floor
in the size-sorted tree, we can further cut memory usage down to
below 20% of what the AVL trees store. This also results in further
reductions in CPU time spent loading metaslabs.
The 16KiB size floor was chosen because it results in substantial memory
usage reduction while not usually resulting in situations where we can't
find an appropriate chunk with the cursor and are forced to use an
oversized chunk from the size-sorted tree. In addition, even if we do
have to use an oversized chunk from the size-sorted tree, the chunk
would be too small to use for ZIL allocations, so it isn't as big of a
loss as it might otherwise be. And often, more small allocations will
follow the initial one, and the cursor search will now find the
remainder of the chunk we didn't use all of and use it for subsequent
allocations. Practical testing has shown little or no change in
fragmentation as a result of this change.
If the size-sorted tree becomes empty while the offset sorted one still
has entries, it will load all the entries from the offset sorted tree
and disregard the size floor until it is unloaded again. This operation
occurs rarely with the default setting, only on incredibly thoroughly
fragmented pools.
There are some other small changes to zdb to teach it to handle btrees,
but nothing major.
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed by: Sebastien Roy seb@delphix.com
Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9181
2019-10-09 20:36:03 +03:00
|
|
|
VERIFY(range_tree_contains(msp->ms_allocatable, start, size));
|
2019-03-29 19:13:20 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT(vd->vdev_ops->vdev_op_leaf);
|
Distributed Spare (dRAID) Feature
This patch adds a new top-level vdev type called dRAID, which stands
for Distributed parity RAID. This pool configuration allows all dRAID
vdevs to participate when rebuilding to a distributed hot spare device.
This can substantially reduce the total time required to restore full
parity to pool with a failed device.
A dRAID pool can be created using the new top-level `draid` type.
Like `raidz`, the desired redundancy is specified after the type:
`draid[1,2,3]`. No additional information is required to create the
pool and reasonable default values will be chosen based on the number
of child vdevs in the dRAID vdev.
zpool create <pool> draid[1,2,3] <vdevs...>
Unlike raidz, additional optional dRAID configuration values can be
provided as part of the draid type as colon separated values. This
allows administrators to fully specify a layout for either performance
or capacity reasons. The supported options include:
zpool create <pool> \
draid[<parity>][:<data>d][:<children>c][:<spares>s] \
<vdevs...>
- draid[parity] - Parity level (default 1)
- draid[:<data>d] - Data devices per group (default 8)
- draid[:<children>c] - Expected number of child vdevs
- draid[:<spares>s] - Distributed hot spares (default 0)
Abbreviated example `zpool status` output for a 68 disk dRAID pool
with two distributed spares using special allocation classes.
```
pool: tank
state: ONLINE
config:
NAME STATE READ WRITE CKSUM
slag7 ONLINE 0 0 0
draid2:8d:68c:2s-0 ONLINE 0 0 0
L0 ONLINE 0 0 0
L1 ONLINE 0 0 0
...
U25 ONLINE 0 0 0
U26 ONLINE 0 0 0
spare-53 ONLINE 0 0 0
U27 ONLINE 0 0 0
draid2-0-0 ONLINE 0 0 0
U28 ONLINE 0 0 0
U29 ONLINE 0 0 0
...
U42 ONLINE 0 0 0
U43 ONLINE 0 0 0
special
mirror-1 ONLINE 0 0 0
L5 ONLINE 0 0 0
U5 ONLINE 0 0 0
mirror-2 ONLINE 0 0 0
L6 ONLINE 0 0 0
U6 ONLINE 0 0 0
spares
draid2-0-0 INUSE currently in use
draid2-0-1 AVAIL
```
When adding test coverage for the new dRAID vdev type the following
options were added to the ztest command. These options are leverages
by zloop.sh to test a wide range of dRAID configurations.
-K draid|raidz|random - kind of RAID to test
-D <value> - dRAID data drives per group
-S <value> - dRAID distributed hot spares
-R <value> - RAID parity (raidz or dRAID)
The zpool_create, zpool_import, redundancy, replacement and fault
test groups have all been updated provide test coverage for the
dRAID feature.
Co-authored-by: Isaac Huang <he.huang@intel.com>
Co-authored-by: Mark Maybee <mmaybee@cray.com>
Co-authored-by: Don Brady <don.brady@delphix.com>
Co-authored-by: Matthew Ahrens <mahrens@delphix.com>
Co-authored-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Mark Maybee <mmaybee@cray.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed-by: Tony Hutter <hutter2@llnl.gov>
Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
Closes #10102
2020-11-14 00:51:51 +03:00
|
|
|
vdev_xlate_walk(vd, &logical_rs, vdev_trim_xlate_range_add, arg);
|
2019-03-29 19:13:20 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Each manual TRIM thread is responsible for trimming the unallocated
|
|
|
|
* space for each leaf vdev. This is accomplished by sequentially iterating
|
|
|
|
* over its top-level metaslabs and issuing TRIM I/O for the space described
|
|
|
|
* by its ms_allocatable. While a metaslab is undergoing trimming it is
|
|
|
|
* not eligible for new allocations.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_thread(void *arg)
|
|
|
|
{
|
|
|
|
vdev_t *vd = arg;
|
|
|
|
spa_t *spa = vd->vdev_spa;
|
|
|
|
trim_args_t ta;
|
|
|
|
int error = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The VDEV_LEAF_ZAP_TRIM_* entries may have been updated by
|
|
|
|
* vdev_trim(). Wait for the updated values to be reflected
|
|
|
|
* in the zap in order to start with the requested settings.
|
|
|
|
*/
|
|
|
|
txg_wait_synced(spa_get_dsl(vd->vdev_spa), 0);
|
|
|
|
|
|
|
|
ASSERT(vdev_is_concrete(vd));
|
|
|
|
spa_config_enter(spa, SCL_CONFIG, FTAG, RW_READER);
|
|
|
|
|
|
|
|
vd->vdev_trim_last_offset = 0;
|
|
|
|
vd->vdev_trim_rate = 0;
|
|
|
|
vd->vdev_trim_partial = 0;
|
|
|
|
vd->vdev_trim_secure = 0;
|
|
|
|
|
|
|
|
VERIFY0(vdev_trim_load(vd));
|
|
|
|
|
|
|
|
ta.trim_vdev = vd;
|
|
|
|
ta.trim_extent_bytes_max = zfs_trim_extent_bytes_max;
|
|
|
|
ta.trim_extent_bytes_min = zfs_trim_extent_bytes_min;
|
Reduce loaded range tree memory usage
This patch implements a new tree structure for ZFS, and uses it to
store range trees more efficiently.
The new structure is approximately a B-tree, though there are some
small differences from the usual characterizations. The tree has core
nodes and leaf nodes; each contain data elements, which the elements
in the core nodes acting as separators between its children. The
difference between core and leaf nodes is that the core nodes have an
array of children, while leaf nodes don't. Every node in the tree may
be only partially full; in most cases, they are all at least 50% full
(in terms of element count) except for the root node, which can be
less full. Underfull nodes will steal from their neighbors or merge to
remain full enough, while overfull nodes will split in two. The data
elements are contained in tree-controlled buffers; they are copied
into these on insertion, and overwritten on deletion. This means that
the elements are not independently allocated, which reduces overhead,
but also means they can't be shared between trees (and also that
pointers to them are only valid until a side-effectful tree operation
occurs). The overhead varies based on how dense the tree is, but is
usually on the order of about 50% of the element size; the per-node
overheads are very small, and so don't make a significant difference.
The trees can accept arbitrary records; they accept a size and a
comparator to allow them to be used for a variety of purposes.
The new trees replace the AVL trees used in the range trees today.
Currently, the range_seg_t structure contains three 8 byte integers
of payload and two 24 byte avl_tree_node_ts to handle its storage in
both an offset-sorted tree and a size-sorted tree (total size: 64
bytes). In the new model, the range seg structures are usually two 4
byte integers, but a separate one needs to exist for the size-sorted
and offset-sorted tree. Between the raw size, the 50% overhead, and
the double storage, the new btrees are expected to use 8*1.5*2 = 24
bytes per record, or 33.3% as much memory as the AVL trees (this is
for the purposes of storing metaslab range trees; for other purposes,
like scrubs, they use ~50% as much memory).
We reduced the size of the payload in the range segments by teaching
range trees about starting offsets and shifts; since metaslabs have a
fixed starting offset, and they all operate in terms of disk sectors,
we can store the ranges using 4-byte integers as long as the size of
the metaslab divided by the sector size is less than 2^32. For 512-byte
sectors, this is a 2^41 (or 2TB) metaslab, which with the default
settings corresponds to a 256PB disk. 4k sector disks can handle
metaslabs up to 2^46 bytes, or 2^63 byte disks. Since we do not
anticipate disks of this size in the near future, there should be
almost no cases where metaslabs need 64-byte integers to store their
ranges. We do still have the capability to store 64-byte integer ranges
to account for cases where we are storing per-vdev (or per-dnode) trees,
which could reasonably go above the limits discussed. We also do not
store fill information in the compact version of the node, since it
is only used for sorted scrub.
We also optimized the metaslab loading process in various other ways
to offset some inefficiencies in the btree model. While individual
operations (find, insert, remove_from) are faster for the btree than
they are for the avl tree, remove usually requires a find operation,
while in the AVL tree model the element itself suffices. Some clever
changes actually caused an overall speedup in metaslab loading; we use
approximately 40% less cpu to load metaslabs in our tests on Illumos.
Another memory and performance optimization was achieved by changing
what is stored in the size-sorted trees. When a disk is heavily
fragmented, the df algorithm used by default in ZFS will almost always
find a number of small regions in its initial cursor-based search; it
will usually only fall back to the size-sorted tree to find larger
regions. If we increase the size of the cursor-based search slightly,
and don't store segments that are smaller than a tunable size floor
in the size-sorted tree, we can further cut memory usage down to
below 20% of what the AVL trees store. This also results in further
reductions in CPU time spent loading metaslabs.
The 16KiB size floor was chosen because it results in substantial memory
usage reduction while not usually resulting in situations where we can't
find an appropriate chunk with the cursor and are forced to use an
oversized chunk from the size-sorted tree. In addition, even if we do
have to use an oversized chunk from the size-sorted tree, the chunk
would be too small to use for ZIL allocations, so it isn't as big of a
loss as it might otherwise be. And often, more small allocations will
follow the initial one, and the cursor search will now find the
remainder of the chunk we didn't use all of and use it for subsequent
allocations. Practical testing has shown little or no change in
fragmentation as a result of this change.
If the size-sorted tree becomes empty while the offset sorted one still
has entries, it will load all the entries from the offset sorted tree
and disregard the size floor until it is unloaded again. This operation
occurs rarely with the default setting, only on incredibly thoroughly
fragmented pools.
There are some other small changes to zdb to teach it to handle btrees,
but nothing major.
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed by: Sebastien Roy seb@delphix.com
Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9181
2019-10-09 20:36:03 +03:00
|
|
|
ta.trim_tree = range_tree_create(NULL, RANGE_SEG64, NULL, 0, 0);
|
2019-03-29 19:13:20 +03:00
|
|
|
ta.trim_type = TRIM_TYPE_MANUAL;
|
|
|
|
ta.trim_flags = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* When a secure TRIM has been requested infer that the intent
|
|
|
|
* is that everything must be trimmed. Override the default
|
|
|
|
* minimum TRIM size to prevent ranges from being skipped.
|
|
|
|
*/
|
|
|
|
if (vd->vdev_trim_secure) {
|
|
|
|
ta.trim_flags |= ZIO_TRIM_SECURE;
|
|
|
|
ta.trim_extent_bytes_min = SPA_MINBLOCKSIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
uint64_t ms_count = 0;
|
|
|
|
for (uint64_t i = 0; !vd->vdev_detached &&
|
|
|
|
i < vd->vdev_top->vdev_ms_count; i++) {
|
|
|
|
metaslab_t *msp = vd->vdev_top->vdev_ms[i];
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we've expanded the top-level vdev or it's our
|
|
|
|
* first pass, calculate our progress.
|
|
|
|
*/
|
|
|
|
if (vd->vdev_top->vdev_ms_count != ms_count) {
|
|
|
|
vdev_trim_calculate_progress(vd);
|
|
|
|
ms_count = vd->vdev_top->vdev_ms_count;
|
|
|
|
}
|
|
|
|
|
|
|
|
spa_config_exit(spa, SCL_CONFIG, FTAG);
|
|
|
|
metaslab_disable(msp);
|
|
|
|
mutex_enter(&msp->ms_lock);
|
|
|
|
VERIFY0(metaslab_load(msp));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If a partial TRIM was requested skip metaslabs which have
|
|
|
|
* never been initialized and thus have never been written.
|
|
|
|
*/
|
|
|
|
if (msp->ms_sm == NULL && vd->vdev_trim_partial) {
|
|
|
|
mutex_exit(&msp->ms_lock);
|
Cap metaslab memory usage
On systems with large amounts of storage and high fragmentation, a huge
amount of space can be used by storing metaslab range trees. Since
metaslabs are only unloaded during a txg sync, and only if they have
been inactive for 8 txgs, it is possible to get into a state where all
of the system's memory is consumed by range trees and metaslabs, and
txgs cannot sync. While ZFS knows how to evict ARC data when needed,
it has no such mechanism for range tree data. This can result in boot
hangs for some system configurations.
First, we add the ability to unload metaslabs outside of syncing
context. Second, we store a multilist of all loaded metaslabs, sorted
by their selection txg, so we can quickly identify the oldest
metaslabs. We use a multilist to reduce lock contention during heavy
write workloads. Finally, we add logic that will unload a metaslab
when we're loading a new metaslab, if we're using more than a certain
fraction of the available memory on range trees.
Reviewed-by: Matt Ahrens <mahrens@delphix.com>
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Sebastien Roy <sebastien.roy@delphix.com>
Reviewed-by: Serapheim Dimitropoulos <serapheim@delphix.com>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9128
2019-08-16 18:08:21 +03:00
|
|
|
metaslab_enable(msp, B_FALSE, B_FALSE);
|
2019-03-29 19:13:20 +03:00
|
|
|
spa_config_enter(spa, SCL_CONFIG, FTAG, RW_READER);
|
|
|
|
vdev_trim_calculate_progress(vd);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
ta.trim_msp = msp;
|
|
|
|
range_tree_walk(msp->ms_allocatable, vdev_trim_range_add, &ta);
|
|
|
|
range_tree_vacate(msp->ms_trim, NULL, NULL);
|
|
|
|
mutex_exit(&msp->ms_lock);
|
|
|
|
|
|
|
|
error = vdev_trim_ranges(&ta);
|
Cap metaslab memory usage
On systems with large amounts of storage and high fragmentation, a huge
amount of space can be used by storing metaslab range trees. Since
metaslabs are only unloaded during a txg sync, and only if they have
been inactive for 8 txgs, it is possible to get into a state where all
of the system's memory is consumed by range trees and metaslabs, and
txgs cannot sync. While ZFS knows how to evict ARC data when needed,
it has no such mechanism for range tree data. This can result in boot
hangs for some system configurations.
First, we add the ability to unload metaslabs outside of syncing
context. Second, we store a multilist of all loaded metaslabs, sorted
by their selection txg, so we can quickly identify the oldest
metaslabs. We use a multilist to reduce lock contention during heavy
write workloads. Finally, we add logic that will unload a metaslab
when we're loading a new metaslab, if we're using more than a certain
fraction of the available memory on range trees.
Reviewed-by: Matt Ahrens <mahrens@delphix.com>
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Sebastien Roy <sebastien.roy@delphix.com>
Reviewed-by: Serapheim Dimitropoulos <serapheim@delphix.com>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9128
2019-08-16 18:08:21 +03:00
|
|
|
metaslab_enable(msp, B_TRUE, B_FALSE);
|
2019-03-29 19:13:20 +03:00
|
|
|
spa_config_enter(spa, SCL_CONFIG, FTAG, RW_READER);
|
|
|
|
|
|
|
|
range_tree_vacate(ta.trim_tree, NULL, NULL);
|
|
|
|
if (error != 0)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
spa_config_exit(spa, SCL_CONFIG, FTAG);
|
|
|
|
mutex_enter(&vd->vdev_trim_io_lock);
|
|
|
|
while (vd->vdev_trim_inflight[0] > 0) {
|
|
|
|
cv_wait(&vd->vdev_trim_io_cv, &vd->vdev_trim_io_lock);
|
|
|
|
}
|
|
|
|
mutex_exit(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
range_tree_destroy(ta.trim_tree);
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
2021-03-02 21:27:27 +03:00
|
|
|
if (!vd->vdev_trim_exit_wanted) {
|
|
|
|
if (vdev_writeable(vd)) {
|
|
|
|
vdev_trim_change_state(vd, VDEV_TRIM_COMPLETE,
|
|
|
|
vd->vdev_trim_rate, vd->vdev_trim_partial,
|
|
|
|
vd->vdev_trim_secure);
|
|
|
|
} else if (vd->vdev_faulted) {
|
|
|
|
vdev_trim_change_state(vd, VDEV_TRIM_CANCELED,
|
|
|
|
vd->vdev_trim_rate, vd->vdev_trim_partial,
|
|
|
|
vd->vdev_trim_secure);
|
|
|
|
}
|
2019-03-29 19:13:20 +03:00
|
|
|
}
|
|
|
|
ASSERT(vd->vdev_trim_thread != NULL || vd->vdev_trim_inflight[0] == 0);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Drop the vdev_trim_lock while we sync out the txg since it's
|
|
|
|
* possible that a device might be trying to come online and must
|
|
|
|
* check to see if it needs to restart a trim. That thread will be
|
|
|
|
* holding the spa_config_lock which would prevent the txg_wait_synced
|
|
|
|
* from completing.
|
|
|
|
*/
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
txg_wait_synced(spa_get_dsl(spa), 0);
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
|
|
|
|
|
|
|
vd->vdev_trim_thread = NULL;
|
|
|
|
cv_broadcast(&vd->vdev_trim_cv);
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
2020-05-15 01:58:09 +03:00
|
|
|
|
|
|
|
thread_exit();
|
2019-03-29 19:13:20 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Initiates a manual TRIM for the vdev_t. Callers must hold vdev_trim_lock,
|
|
|
|
* the vdev_t must be a leaf and cannot already be manually trimming.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_trim(vdev_t *vd, uint64_t rate, boolean_t partial, boolean_t secure)
|
|
|
|
{
|
|
|
|
ASSERT(MUTEX_HELD(&vd->vdev_trim_lock));
|
|
|
|
ASSERT(vd->vdev_ops->vdev_op_leaf);
|
|
|
|
ASSERT(vdev_is_concrete(vd));
|
|
|
|
ASSERT3P(vd->vdev_trim_thread, ==, NULL);
|
|
|
|
ASSERT(!vd->vdev_detached);
|
|
|
|
ASSERT(!vd->vdev_trim_exit_wanted);
|
|
|
|
ASSERT(!vd->vdev_top->vdev_removing);
|
|
|
|
|
|
|
|
vdev_trim_change_state(vd, VDEV_TRIM_ACTIVE, rate, partial, secure);
|
|
|
|
vd->vdev_trim_thread = thread_create(NULL, 0,
|
|
|
|
vdev_trim_thread, vd, 0, &p0, TS_RUN, maxclsyspri);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Wait for the trimming thread to be terminated (canceled or stopped).
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_stop_wait_impl(vdev_t *vd)
|
|
|
|
{
|
|
|
|
ASSERT(MUTEX_HELD(&vd->vdev_trim_lock));
|
|
|
|
|
|
|
|
while (vd->vdev_trim_thread != NULL)
|
|
|
|
cv_wait(&vd->vdev_trim_cv, &vd->vdev_trim_lock);
|
|
|
|
|
|
|
|
ASSERT3P(vd->vdev_trim_thread, ==, NULL);
|
|
|
|
vd->vdev_trim_exit_wanted = B_FALSE;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Wait for vdev trim threads which were listed to cleanly exit.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_trim_stop_wait(spa_t *spa, list_t *vd_list)
|
|
|
|
{
|
|
|
|
vdev_t *vd;
|
|
|
|
|
|
|
|
ASSERT(MUTEX_HELD(&spa_namespace_lock));
|
|
|
|
|
|
|
|
while ((vd = list_remove_head(vd_list)) != NULL) {
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
|
|
|
vdev_trim_stop_wait_impl(vd);
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Stop trimming a device, with the resultant trimming state being tgt_state.
|
|
|
|
* For blocking behavior pass NULL for vd_list. Otherwise, when a list_t is
|
|
|
|
* provided the stopping vdev is inserted in to the list. Callers are then
|
|
|
|
* required to call vdev_trim_stop_wait() to block for all the trim threads
|
|
|
|
* to exit. The caller must hold vdev_trim_lock and must not be writing to
|
|
|
|
* the spa config, as the trimming thread may try to enter the config as a
|
|
|
|
* reader before exiting.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_trim_stop(vdev_t *vd, vdev_trim_state_t tgt_state, list_t *vd_list)
|
|
|
|
{
|
|
|
|
ASSERT(!spa_config_held(vd->vdev_spa, SCL_CONFIG|SCL_STATE, RW_WRITER));
|
|
|
|
ASSERT(MUTEX_HELD(&vd->vdev_trim_lock));
|
|
|
|
ASSERT(vd->vdev_ops->vdev_op_leaf);
|
|
|
|
ASSERT(vdev_is_concrete(vd));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allow cancel requests to proceed even if the trim thread has
|
|
|
|
* stopped.
|
|
|
|
*/
|
|
|
|
if (vd->vdev_trim_thread == NULL && tgt_state != VDEV_TRIM_CANCELED)
|
|
|
|
return;
|
|
|
|
|
|
|
|
vdev_trim_change_state(vd, tgt_state, 0, 0, 0);
|
|
|
|
vd->vdev_trim_exit_wanted = B_TRUE;
|
|
|
|
|
|
|
|
if (vd_list == NULL) {
|
|
|
|
vdev_trim_stop_wait_impl(vd);
|
|
|
|
} else {
|
|
|
|
ASSERT(MUTEX_HELD(&spa_namespace_lock));
|
|
|
|
list_insert_tail(vd_list, vd);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Requests that all listed vdevs stop trimming.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_stop_all_impl(vdev_t *vd, vdev_trim_state_t tgt_state,
|
|
|
|
list_t *vd_list)
|
|
|
|
{
|
|
|
|
if (vd->vdev_ops->vdev_op_leaf && vdev_is_concrete(vd)) {
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
|
|
|
vdev_trim_stop(vd, tgt_state, vd_list);
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (uint64_t i = 0; i < vd->vdev_children; i++) {
|
|
|
|
vdev_trim_stop_all_impl(vd->vdev_child[i], tgt_state,
|
|
|
|
vd_list);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Convenience function to stop trimming of a vdev tree and set all trim
|
|
|
|
* thread pointers to NULL.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_trim_stop_all(vdev_t *vd, vdev_trim_state_t tgt_state)
|
|
|
|
{
|
|
|
|
spa_t *spa = vd->vdev_spa;
|
|
|
|
list_t vd_list;
|
2020-06-09 20:15:08 +03:00
|
|
|
vdev_t *vd_l2cache;
|
2019-03-29 19:13:20 +03:00
|
|
|
|
|
|
|
ASSERT(MUTEX_HELD(&spa_namespace_lock));
|
|
|
|
|
|
|
|
list_create(&vd_list, sizeof (vdev_t),
|
|
|
|
offsetof(vdev_t, vdev_trim_node));
|
|
|
|
|
|
|
|
vdev_trim_stop_all_impl(vd, tgt_state, &vd_list);
|
2020-06-09 20:15:08 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Iterate over cache devices and request stop trimming the
|
|
|
|
* whole device in case we export the pool or remove the cache
|
|
|
|
* device prematurely.
|
|
|
|
*/
|
|
|
|
for (int i = 0; i < spa->spa_l2cache.sav_count; i++) {
|
|
|
|
vd_l2cache = spa->spa_l2cache.sav_vdevs[i];
|
|
|
|
vdev_trim_stop_all_impl(vd_l2cache, tgt_state, &vd_list);
|
|
|
|
}
|
|
|
|
|
2019-03-29 19:13:20 +03:00
|
|
|
vdev_trim_stop_wait(spa, &vd_list);
|
|
|
|
|
|
|
|
if (vd->vdev_spa->spa_sync_on) {
|
|
|
|
/* Make sure that our state has been synced to disk */
|
|
|
|
txg_wait_synced(spa_get_dsl(vd->vdev_spa), 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
list_destroy(&vd_list);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Conditionally restarts a manual TRIM given its on-disk state.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_trim_restart(vdev_t *vd)
|
|
|
|
{
|
|
|
|
ASSERT(MUTEX_HELD(&spa_namespace_lock));
|
|
|
|
ASSERT(!spa_config_held(vd->vdev_spa, SCL_ALL, RW_WRITER));
|
|
|
|
|
|
|
|
if (vd->vdev_leaf_zap != 0) {
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
|
|
|
uint64_t trim_state = VDEV_TRIM_NONE;
|
|
|
|
int err = zap_lookup(vd->vdev_spa->spa_meta_objset,
|
|
|
|
vd->vdev_leaf_zap, VDEV_LEAF_ZAP_TRIM_STATE,
|
|
|
|
sizeof (trim_state), 1, &trim_state);
|
|
|
|
ASSERT(err == 0 || err == ENOENT);
|
|
|
|
vd->vdev_trim_state = trim_state;
|
|
|
|
|
|
|
|
uint64_t timestamp = 0;
|
|
|
|
err = zap_lookup(vd->vdev_spa->spa_meta_objset,
|
|
|
|
vd->vdev_leaf_zap, VDEV_LEAF_ZAP_TRIM_ACTION_TIME,
|
|
|
|
sizeof (timestamp), 1, ×tamp);
|
|
|
|
ASSERT(err == 0 || err == ENOENT);
|
2020-02-27 00:18:07 +03:00
|
|
|
vd->vdev_trim_action_time = timestamp;
|
2019-03-29 19:13:20 +03:00
|
|
|
|
|
|
|
if (vd->vdev_trim_state == VDEV_TRIM_SUSPENDED ||
|
|
|
|
vd->vdev_offline) {
|
|
|
|
/* load progress for reporting, but don't resume */
|
|
|
|
VERIFY0(vdev_trim_load(vd));
|
|
|
|
} else if (vd->vdev_trim_state == VDEV_TRIM_ACTIVE &&
|
|
|
|
vdev_writeable(vd) && !vd->vdev_top->vdev_removing &&
|
|
|
|
vd->vdev_trim_thread == NULL) {
|
|
|
|
VERIFY0(vdev_trim_load(vd));
|
|
|
|
vdev_trim(vd, vd->vdev_trim_rate,
|
|
|
|
vd->vdev_trim_partial, vd->vdev_trim_secure);
|
|
|
|
}
|
|
|
|
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (uint64_t i = 0; i < vd->vdev_children; i++) {
|
|
|
|
vdev_trim_restart(vd->vdev_child[i]);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Used by the automatic TRIM when ZFS_DEBUG_TRIM is set to verify that
|
|
|
|
* every TRIM range is contained within ms_allocatable.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_trim_range_verify(void *arg, uint64_t start, uint64_t size)
|
|
|
|
{
|
|
|
|
trim_args_t *ta = arg;
|
|
|
|
metaslab_t *msp = ta->trim_msp;
|
|
|
|
|
|
|
|
VERIFY3B(msp->ms_loaded, ==, B_TRUE);
|
|
|
|
VERIFY3U(msp->ms_disabled, >, 0);
|
Reduce loaded range tree memory usage
This patch implements a new tree structure for ZFS, and uses it to
store range trees more efficiently.
The new structure is approximately a B-tree, though there are some
small differences from the usual characterizations. The tree has core
nodes and leaf nodes; each contain data elements, which the elements
in the core nodes acting as separators between its children. The
difference between core and leaf nodes is that the core nodes have an
array of children, while leaf nodes don't. Every node in the tree may
be only partially full; in most cases, they are all at least 50% full
(in terms of element count) except for the root node, which can be
less full. Underfull nodes will steal from their neighbors or merge to
remain full enough, while overfull nodes will split in two. The data
elements are contained in tree-controlled buffers; they are copied
into these on insertion, and overwritten on deletion. This means that
the elements are not independently allocated, which reduces overhead,
but also means they can't be shared between trees (and also that
pointers to them are only valid until a side-effectful tree operation
occurs). The overhead varies based on how dense the tree is, but is
usually on the order of about 50% of the element size; the per-node
overheads are very small, and so don't make a significant difference.
The trees can accept arbitrary records; they accept a size and a
comparator to allow them to be used for a variety of purposes.
The new trees replace the AVL trees used in the range trees today.
Currently, the range_seg_t structure contains three 8 byte integers
of payload and two 24 byte avl_tree_node_ts to handle its storage in
both an offset-sorted tree and a size-sorted tree (total size: 64
bytes). In the new model, the range seg structures are usually two 4
byte integers, but a separate one needs to exist for the size-sorted
and offset-sorted tree. Between the raw size, the 50% overhead, and
the double storage, the new btrees are expected to use 8*1.5*2 = 24
bytes per record, or 33.3% as much memory as the AVL trees (this is
for the purposes of storing metaslab range trees; for other purposes,
like scrubs, they use ~50% as much memory).
We reduced the size of the payload in the range segments by teaching
range trees about starting offsets and shifts; since metaslabs have a
fixed starting offset, and they all operate in terms of disk sectors,
we can store the ranges using 4-byte integers as long as the size of
the metaslab divided by the sector size is less than 2^32. For 512-byte
sectors, this is a 2^41 (or 2TB) metaslab, which with the default
settings corresponds to a 256PB disk. 4k sector disks can handle
metaslabs up to 2^46 bytes, or 2^63 byte disks. Since we do not
anticipate disks of this size in the near future, there should be
almost no cases where metaslabs need 64-byte integers to store their
ranges. We do still have the capability to store 64-byte integer ranges
to account for cases where we are storing per-vdev (or per-dnode) trees,
which could reasonably go above the limits discussed. We also do not
store fill information in the compact version of the node, since it
is only used for sorted scrub.
We also optimized the metaslab loading process in various other ways
to offset some inefficiencies in the btree model. While individual
operations (find, insert, remove_from) are faster for the btree than
they are for the avl tree, remove usually requires a find operation,
while in the AVL tree model the element itself suffices. Some clever
changes actually caused an overall speedup in metaslab loading; we use
approximately 40% less cpu to load metaslabs in our tests on Illumos.
Another memory and performance optimization was achieved by changing
what is stored in the size-sorted trees. When a disk is heavily
fragmented, the df algorithm used by default in ZFS will almost always
find a number of small regions in its initial cursor-based search; it
will usually only fall back to the size-sorted tree to find larger
regions. If we increase the size of the cursor-based search slightly,
and don't store segments that are smaller than a tunable size floor
in the size-sorted tree, we can further cut memory usage down to
below 20% of what the AVL trees store. This also results in further
reductions in CPU time spent loading metaslabs.
The 16KiB size floor was chosen because it results in substantial memory
usage reduction while not usually resulting in situations where we can't
find an appropriate chunk with the cursor and are forced to use an
oversized chunk from the size-sorted tree. In addition, even if we do
have to use an oversized chunk from the size-sorted tree, the chunk
would be too small to use for ZIL allocations, so it isn't as big of a
loss as it might otherwise be. And often, more small allocations will
follow the initial one, and the cursor search will now find the
remainder of the chunk we didn't use all of and use it for subsequent
allocations. Practical testing has shown little or no change in
fragmentation as a result of this change.
If the size-sorted tree becomes empty while the offset sorted one still
has entries, it will load all the entries from the offset sorted tree
and disregard the size floor until it is unloaded again. This operation
occurs rarely with the default setting, only on incredibly thoroughly
fragmented pools.
There are some other small changes to zdb to teach it to handle btrees,
but nothing major.
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed by: Sebastien Roy seb@delphix.com
Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9181
2019-10-09 20:36:03 +03:00
|
|
|
VERIFY(range_tree_contains(msp->ms_allocatable, start, size));
|
2019-03-29 19:13:20 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Each automatic TRIM thread is responsible for managing the trimming of a
|
|
|
|
* top-level vdev in the pool. No automatic TRIM state is maintained on-disk.
|
|
|
|
*
|
|
|
|
* N.B. This behavior is different from a manual TRIM where a thread
|
|
|
|
* is created for each leaf vdev, instead of each top-level vdev.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
vdev_autotrim_thread(void *arg)
|
|
|
|
{
|
|
|
|
vdev_t *vd = arg;
|
|
|
|
spa_t *spa = vd->vdev_spa;
|
|
|
|
int shift = 0;
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_autotrim_lock);
|
|
|
|
ASSERT3P(vd->vdev_top, ==, vd);
|
|
|
|
ASSERT3P(vd->vdev_autotrim_thread, !=, NULL);
|
|
|
|
mutex_exit(&vd->vdev_autotrim_lock);
|
|
|
|
spa_config_enter(spa, SCL_CONFIG, FTAG, RW_READER);
|
|
|
|
|
|
|
|
uint64_t extent_bytes_max = zfs_trim_extent_bytes_max;
|
|
|
|
uint64_t extent_bytes_min = zfs_trim_extent_bytes_min;
|
|
|
|
|
|
|
|
while (!vdev_autotrim_should_stop(vd)) {
|
|
|
|
int txgs_per_trim = MAX(zfs_trim_txg_batch, 1);
|
|
|
|
boolean_t issued_trim = B_FALSE;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* All of the metaslabs are divided in to groups of size
|
|
|
|
* num_metaslabs / zfs_trim_txg_batch. Each of these groups
|
|
|
|
* is composed of metaslabs which are spread evenly over the
|
|
|
|
* device.
|
|
|
|
*
|
|
|
|
* For example, when zfs_trim_txg_batch = 32 (default) then
|
|
|
|
* group 0 will contain metaslabs 0, 32, 64, ...;
|
|
|
|
* group 1 will contain metaslabs 1, 33, 65, ...;
|
|
|
|
* group 2 will contain metaslabs 2, 34, 66, ...; and so on.
|
|
|
|
*
|
|
|
|
* On each pass through the while() loop one of these groups
|
|
|
|
* is selected. This is accomplished by using a shift value
|
|
|
|
* to select the starting metaslab, then striding over the
|
|
|
|
* metaslabs using the zfs_trim_txg_batch size. This is
|
|
|
|
* done to accomplish two things.
|
|
|
|
*
|
|
|
|
* 1) By dividing the metaslabs in to groups, and making sure
|
|
|
|
* that each group takes a minimum of one txg to process.
|
|
|
|
* Then zfs_trim_txg_batch controls the minimum number of
|
|
|
|
* txgs which must occur before a metaslab is revisited.
|
|
|
|
*
|
|
|
|
* 2) Selecting non-consecutive metaslabs distributes the
|
|
|
|
* TRIM commands for a group evenly over the entire device.
|
|
|
|
* This can be advantageous for certain types of devices.
|
|
|
|
*/
|
|
|
|
for (uint64_t i = shift % txgs_per_trim; i < vd->vdev_ms_count;
|
|
|
|
i += txgs_per_trim) {
|
|
|
|
metaslab_t *msp = vd->vdev_ms[i];
|
|
|
|
range_tree_t *trim_tree;
|
|
|
|
|
|
|
|
spa_config_exit(spa, SCL_CONFIG, FTAG);
|
|
|
|
metaslab_disable(msp);
|
|
|
|
spa_config_enter(spa, SCL_CONFIG, FTAG, RW_READER);
|
|
|
|
|
|
|
|
mutex_enter(&msp->ms_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Skip the metaslab when it has never been allocated
|
|
|
|
* or when there are no recent frees to trim.
|
|
|
|
*/
|
|
|
|
if (msp->ms_sm == NULL ||
|
|
|
|
range_tree_is_empty(msp->ms_trim)) {
|
|
|
|
mutex_exit(&msp->ms_lock);
|
Cap metaslab memory usage
On systems with large amounts of storage and high fragmentation, a huge
amount of space can be used by storing metaslab range trees. Since
metaslabs are only unloaded during a txg sync, and only if they have
been inactive for 8 txgs, it is possible to get into a state where all
of the system's memory is consumed by range trees and metaslabs, and
txgs cannot sync. While ZFS knows how to evict ARC data when needed,
it has no such mechanism for range tree data. This can result in boot
hangs for some system configurations.
First, we add the ability to unload metaslabs outside of syncing
context. Second, we store a multilist of all loaded metaslabs, sorted
by their selection txg, so we can quickly identify the oldest
metaslabs. We use a multilist to reduce lock contention during heavy
write workloads. Finally, we add logic that will unload a metaslab
when we're loading a new metaslab, if we're using more than a certain
fraction of the available memory on range trees.
Reviewed-by: Matt Ahrens <mahrens@delphix.com>
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Sebastien Roy <sebastien.roy@delphix.com>
Reviewed-by: Serapheim Dimitropoulos <serapheim@delphix.com>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9128
2019-08-16 18:08:21 +03:00
|
|
|
metaslab_enable(msp, B_FALSE, B_FALSE);
|
2019-03-29 19:13:20 +03:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Skip the metaslab when it has already been disabled.
|
|
|
|
* This may happen when a manual TRIM or initialize
|
|
|
|
* operation is running concurrently. In the case
|
|
|
|
* of a manual TRIM, the ms_trim tree will have been
|
|
|
|
* vacated. Only ranges added after the manual TRIM
|
|
|
|
* disabled the metaslab will be included in the tree.
|
|
|
|
* These will be processed when the automatic TRIM
|
|
|
|
* next revisits this metaslab.
|
|
|
|
*/
|
|
|
|
if (msp->ms_disabled > 1) {
|
|
|
|
mutex_exit(&msp->ms_lock);
|
Cap metaslab memory usage
On systems with large amounts of storage and high fragmentation, a huge
amount of space can be used by storing metaslab range trees. Since
metaslabs are only unloaded during a txg sync, and only if they have
been inactive for 8 txgs, it is possible to get into a state where all
of the system's memory is consumed by range trees and metaslabs, and
txgs cannot sync. While ZFS knows how to evict ARC data when needed,
it has no such mechanism for range tree data. This can result in boot
hangs for some system configurations.
First, we add the ability to unload metaslabs outside of syncing
context. Second, we store a multilist of all loaded metaslabs, sorted
by their selection txg, so we can quickly identify the oldest
metaslabs. We use a multilist to reduce lock contention during heavy
write workloads. Finally, we add logic that will unload a metaslab
when we're loading a new metaslab, if we're using more than a certain
fraction of the available memory on range trees.
Reviewed-by: Matt Ahrens <mahrens@delphix.com>
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Sebastien Roy <sebastien.roy@delphix.com>
Reviewed-by: Serapheim Dimitropoulos <serapheim@delphix.com>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9128
2019-08-16 18:08:21 +03:00
|
|
|
metaslab_enable(msp, B_FALSE, B_FALSE);
|
2019-03-29 19:13:20 +03:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate an empty range tree which is swapped in
|
|
|
|
* for the existing ms_trim tree while it is processed.
|
|
|
|
*/
|
Reduce loaded range tree memory usage
This patch implements a new tree structure for ZFS, and uses it to
store range trees more efficiently.
The new structure is approximately a B-tree, though there are some
small differences from the usual characterizations. The tree has core
nodes and leaf nodes; each contain data elements, which the elements
in the core nodes acting as separators between its children. The
difference between core and leaf nodes is that the core nodes have an
array of children, while leaf nodes don't. Every node in the tree may
be only partially full; in most cases, they are all at least 50% full
(in terms of element count) except for the root node, which can be
less full. Underfull nodes will steal from their neighbors or merge to
remain full enough, while overfull nodes will split in two. The data
elements are contained in tree-controlled buffers; they are copied
into these on insertion, and overwritten on deletion. This means that
the elements are not independently allocated, which reduces overhead,
but also means they can't be shared between trees (and also that
pointers to them are only valid until a side-effectful tree operation
occurs). The overhead varies based on how dense the tree is, but is
usually on the order of about 50% of the element size; the per-node
overheads are very small, and so don't make a significant difference.
The trees can accept arbitrary records; they accept a size and a
comparator to allow them to be used for a variety of purposes.
The new trees replace the AVL trees used in the range trees today.
Currently, the range_seg_t structure contains three 8 byte integers
of payload and two 24 byte avl_tree_node_ts to handle its storage in
both an offset-sorted tree and a size-sorted tree (total size: 64
bytes). In the new model, the range seg structures are usually two 4
byte integers, but a separate one needs to exist for the size-sorted
and offset-sorted tree. Between the raw size, the 50% overhead, and
the double storage, the new btrees are expected to use 8*1.5*2 = 24
bytes per record, or 33.3% as much memory as the AVL trees (this is
for the purposes of storing metaslab range trees; for other purposes,
like scrubs, they use ~50% as much memory).
We reduced the size of the payload in the range segments by teaching
range trees about starting offsets and shifts; since metaslabs have a
fixed starting offset, and they all operate in terms of disk sectors,
we can store the ranges using 4-byte integers as long as the size of
the metaslab divided by the sector size is less than 2^32. For 512-byte
sectors, this is a 2^41 (or 2TB) metaslab, which with the default
settings corresponds to a 256PB disk. 4k sector disks can handle
metaslabs up to 2^46 bytes, or 2^63 byte disks. Since we do not
anticipate disks of this size in the near future, there should be
almost no cases where metaslabs need 64-byte integers to store their
ranges. We do still have the capability to store 64-byte integer ranges
to account for cases where we are storing per-vdev (or per-dnode) trees,
which could reasonably go above the limits discussed. We also do not
store fill information in the compact version of the node, since it
is only used for sorted scrub.
We also optimized the metaslab loading process in various other ways
to offset some inefficiencies in the btree model. While individual
operations (find, insert, remove_from) are faster for the btree than
they are for the avl tree, remove usually requires a find operation,
while in the AVL tree model the element itself suffices. Some clever
changes actually caused an overall speedup in metaslab loading; we use
approximately 40% less cpu to load metaslabs in our tests on Illumos.
Another memory and performance optimization was achieved by changing
what is stored in the size-sorted trees. When a disk is heavily
fragmented, the df algorithm used by default in ZFS will almost always
find a number of small regions in its initial cursor-based search; it
will usually only fall back to the size-sorted tree to find larger
regions. If we increase the size of the cursor-based search slightly,
and don't store segments that are smaller than a tunable size floor
in the size-sorted tree, we can further cut memory usage down to
below 20% of what the AVL trees store. This also results in further
reductions in CPU time spent loading metaslabs.
The 16KiB size floor was chosen because it results in substantial memory
usage reduction while not usually resulting in situations where we can't
find an appropriate chunk with the cursor and are forced to use an
oversized chunk from the size-sorted tree. In addition, even if we do
have to use an oversized chunk from the size-sorted tree, the chunk
would be too small to use for ZIL allocations, so it isn't as big of a
loss as it might otherwise be. And often, more small allocations will
follow the initial one, and the cursor search will now find the
remainder of the chunk we didn't use all of and use it for subsequent
allocations. Practical testing has shown little or no change in
fragmentation as a result of this change.
If the size-sorted tree becomes empty while the offset sorted one still
has entries, it will load all the entries from the offset sorted tree
and disregard the size floor until it is unloaded again. This operation
occurs rarely with the default setting, only on incredibly thoroughly
fragmented pools.
There are some other small changes to zdb to teach it to handle btrees,
but nothing major.
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed by: Sebastien Roy seb@delphix.com
Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9181
2019-10-09 20:36:03 +03:00
|
|
|
trim_tree = range_tree_create(NULL, RANGE_SEG64, NULL,
|
|
|
|
0, 0);
|
2019-03-29 19:13:20 +03:00
|
|
|
range_tree_swap(&msp->ms_trim, &trim_tree);
|
|
|
|
ASSERT(range_tree_is_empty(msp->ms_trim));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* There are two cases when constructing the per-vdev
|
|
|
|
* trim trees for a metaslab. If the top-level vdev
|
|
|
|
* has no children then it is also a leaf and should
|
|
|
|
* be trimmed. Otherwise our children are the leaves
|
|
|
|
* and a trim tree should be constructed for each.
|
|
|
|
*/
|
|
|
|
trim_args_t *tap;
|
|
|
|
uint64_t children = vd->vdev_children;
|
|
|
|
if (children == 0) {
|
|
|
|
children = 1;
|
|
|
|
tap = kmem_zalloc(sizeof (trim_args_t) *
|
|
|
|
children, KM_SLEEP);
|
|
|
|
tap[0].trim_vdev = vd;
|
|
|
|
} else {
|
|
|
|
tap = kmem_zalloc(sizeof (trim_args_t) *
|
|
|
|
children, KM_SLEEP);
|
|
|
|
|
|
|
|
for (uint64_t c = 0; c < children; c++) {
|
|
|
|
tap[c].trim_vdev = vd->vdev_child[c];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (uint64_t c = 0; c < children; c++) {
|
|
|
|
trim_args_t *ta = &tap[c];
|
|
|
|
vdev_t *cvd = ta->trim_vdev;
|
|
|
|
|
|
|
|
ta->trim_msp = msp;
|
|
|
|
ta->trim_extent_bytes_max = extent_bytes_max;
|
|
|
|
ta->trim_extent_bytes_min = extent_bytes_min;
|
|
|
|
ta->trim_type = TRIM_TYPE_AUTO;
|
|
|
|
ta->trim_flags = 0;
|
|
|
|
|
|
|
|
if (cvd->vdev_detached ||
|
|
|
|
!vdev_writeable(cvd) ||
|
|
|
|
!cvd->vdev_has_trim ||
|
|
|
|
cvd->vdev_trim_thread != NULL) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* When a device has an attached hot spare, or
|
|
|
|
* is being replaced it will not be trimmed.
|
|
|
|
* This is done to avoid adding additional
|
|
|
|
* stress to a potentially unhealthy device,
|
|
|
|
* and to minimize the required rebuild time.
|
|
|
|
*/
|
|
|
|
if (!cvd->vdev_ops->vdev_op_leaf)
|
|
|
|
continue;
|
|
|
|
|
Reduce loaded range tree memory usage
This patch implements a new tree structure for ZFS, and uses it to
store range trees more efficiently.
The new structure is approximately a B-tree, though there are some
small differences from the usual characterizations. The tree has core
nodes and leaf nodes; each contain data elements, which the elements
in the core nodes acting as separators between its children. The
difference between core and leaf nodes is that the core nodes have an
array of children, while leaf nodes don't. Every node in the tree may
be only partially full; in most cases, they are all at least 50% full
(in terms of element count) except for the root node, which can be
less full. Underfull nodes will steal from their neighbors or merge to
remain full enough, while overfull nodes will split in two. The data
elements are contained in tree-controlled buffers; they are copied
into these on insertion, and overwritten on deletion. This means that
the elements are not independently allocated, which reduces overhead,
but also means they can't be shared between trees (and also that
pointers to them are only valid until a side-effectful tree operation
occurs). The overhead varies based on how dense the tree is, but is
usually on the order of about 50% of the element size; the per-node
overheads are very small, and so don't make a significant difference.
The trees can accept arbitrary records; they accept a size and a
comparator to allow them to be used for a variety of purposes.
The new trees replace the AVL trees used in the range trees today.
Currently, the range_seg_t structure contains three 8 byte integers
of payload and two 24 byte avl_tree_node_ts to handle its storage in
both an offset-sorted tree and a size-sorted tree (total size: 64
bytes). In the new model, the range seg structures are usually two 4
byte integers, but a separate one needs to exist for the size-sorted
and offset-sorted tree. Between the raw size, the 50% overhead, and
the double storage, the new btrees are expected to use 8*1.5*2 = 24
bytes per record, or 33.3% as much memory as the AVL trees (this is
for the purposes of storing metaslab range trees; for other purposes,
like scrubs, they use ~50% as much memory).
We reduced the size of the payload in the range segments by teaching
range trees about starting offsets and shifts; since metaslabs have a
fixed starting offset, and they all operate in terms of disk sectors,
we can store the ranges using 4-byte integers as long as the size of
the metaslab divided by the sector size is less than 2^32. For 512-byte
sectors, this is a 2^41 (or 2TB) metaslab, which with the default
settings corresponds to a 256PB disk. 4k sector disks can handle
metaslabs up to 2^46 bytes, or 2^63 byte disks. Since we do not
anticipate disks of this size in the near future, there should be
almost no cases where metaslabs need 64-byte integers to store their
ranges. We do still have the capability to store 64-byte integer ranges
to account for cases where we are storing per-vdev (or per-dnode) trees,
which could reasonably go above the limits discussed. We also do not
store fill information in the compact version of the node, since it
is only used for sorted scrub.
We also optimized the metaslab loading process in various other ways
to offset some inefficiencies in the btree model. While individual
operations (find, insert, remove_from) are faster for the btree than
they are for the avl tree, remove usually requires a find operation,
while in the AVL tree model the element itself suffices. Some clever
changes actually caused an overall speedup in metaslab loading; we use
approximately 40% less cpu to load metaslabs in our tests on Illumos.
Another memory and performance optimization was achieved by changing
what is stored in the size-sorted trees. When a disk is heavily
fragmented, the df algorithm used by default in ZFS will almost always
find a number of small regions in its initial cursor-based search; it
will usually only fall back to the size-sorted tree to find larger
regions. If we increase the size of the cursor-based search slightly,
and don't store segments that are smaller than a tunable size floor
in the size-sorted tree, we can further cut memory usage down to
below 20% of what the AVL trees store. This also results in further
reductions in CPU time spent loading metaslabs.
The 16KiB size floor was chosen because it results in substantial memory
usage reduction while not usually resulting in situations where we can't
find an appropriate chunk with the cursor and are forced to use an
oversized chunk from the size-sorted tree. In addition, even if we do
have to use an oversized chunk from the size-sorted tree, the chunk
would be too small to use for ZIL allocations, so it isn't as big of a
loss as it might otherwise be. And often, more small allocations will
follow the initial one, and the cursor search will now find the
remainder of the chunk we didn't use all of and use it for subsequent
allocations. Practical testing has shown little or no change in
fragmentation as a result of this change.
If the size-sorted tree becomes empty while the offset sorted one still
has entries, it will load all the entries from the offset sorted tree
and disregard the size floor until it is unloaded again. This operation
occurs rarely with the default setting, only on incredibly thoroughly
fragmented pools.
There are some other small changes to zdb to teach it to handle btrees,
but nothing major.
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Matt Ahrens <matt@delphix.com>
Reviewed by: Sebastien Roy seb@delphix.com
Reviewed-by: Igor Kozhukhov <igor@dilos.org>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9181
2019-10-09 20:36:03 +03:00
|
|
|
ta->trim_tree = range_tree_create(NULL,
|
|
|
|
RANGE_SEG64, NULL, 0, 0);
|
2019-03-29 19:13:20 +03:00
|
|
|
range_tree_walk(trim_tree,
|
|
|
|
vdev_trim_range_add, ta);
|
|
|
|
}
|
|
|
|
|
|
|
|
mutex_exit(&msp->ms_lock);
|
|
|
|
spa_config_exit(spa, SCL_CONFIG, FTAG);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Issue the TRIM I/Os for all ranges covered by the
|
|
|
|
* TRIM trees. These ranges are safe to TRIM because
|
|
|
|
* no new allocations will be performed until the call
|
|
|
|
* to metaslab_enabled() below.
|
|
|
|
*/
|
|
|
|
for (uint64_t c = 0; c < children; c++) {
|
|
|
|
trim_args_t *ta = &tap[c];
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Always yield to a manual TRIM if one has
|
|
|
|
* been started for the child vdev.
|
|
|
|
*/
|
|
|
|
if (ta->trim_tree == NULL ||
|
|
|
|
ta->trim_vdev->vdev_trim_thread != NULL) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* After this point metaslab_enable() must be
|
|
|
|
* called with the sync flag set. This is done
|
|
|
|
* here because vdev_trim_ranges() is allowed
|
|
|
|
* to be interrupted (EINTR) before issuing all
|
|
|
|
* of the required TRIM I/Os.
|
|
|
|
*/
|
|
|
|
issued_trim = B_TRUE;
|
|
|
|
|
|
|
|
int error = vdev_trim_ranges(ta);
|
|
|
|
if (error)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Verify every range which was trimmed is still
|
|
|
|
* contained within the ms_allocatable tree.
|
|
|
|
*/
|
|
|
|
if (zfs_flags & ZFS_DEBUG_TRIM) {
|
|
|
|
mutex_enter(&msp->ms_lock);
|
|
|
|
VERIFY0(metaslab_load(msp));
|
|
|
|
VERIFY3P(tap[0].trim_msp, ==, msp);
|
|
|
|
range_tree_walk(trim_tree,
|
|
|
|
vdev_trim_range_verify, &tap[0]);
|
|
|
|
mutex_exit(&msp->ms_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
range_tree_vacate(trim_tree, NULL, NULL);
|
|
|
|
range_tree_destroy(trim_tree);
|
|
|
|
|
Cap metaslab memory usage
On systems with large amounts of storage and high fragmentation, a huge
amount of space can be used by storing metaslab range trees. Since
metaslabs are only unloaded during a txg sync, and only if they have
been inactive for 8 txgs, it is possible to get into a state where all
of the system's memory is consumed by range trees and metaslabs, and
txgs cannot sync. While ZFS knows how to evict ARC data when needed,
it has no such mechanism for range tree data. This can result in boot
hangs for some system configurations.
First, we add the ability to unload metaslabs outside of syncing
context. Second, we store a multilist of all loaded metaslabs, sorted
by their selection txg, so we can quickly identify the oldest
metaslabs. We use a multilist to reduce lock contention during heavy
write workloads. Finally, we add logic that will unload a metaslab
when we're loading a new metaslab, if we're using more than a certain
fraction of the available memory on range trees.
Reviewed-by: Matt Ahrens <mahrens@delphix.com>
Reviewed-by: George Wilson <gwilson@delphix.com>
Reviewed-by: Sebastien Roy <sebastien.roy@delphix.com>
Reviewed-by: Serapheim Dimitropoulos <serapheim@delphix.com>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Signed-off-by: Paul Dagnelie <pcd@delphix.com>
Closes #9128
2019-08-16 18:08:21 +03:00
|
|
|
metaslab_enable(msp, issued_trim, B_FALSE);
|
2019-03-29 19:13:20 +03:00
|
|
|
spa_config_enter(spa, SCL_CONFIG, FTAG, RW_READER);
|
|
|
|
|
|
|
|
for (uint64_t c = 0; c < children; c++) {
|
|
|
|
trim_args_t *ta = &tap[c];
|
|
|
|
|
|
|
|
if (ta->trim_tree == NULL)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
range_tree_vacate(ta->trim_tree, NULL, NULL);
|
|
|
|
range_tree_destroy(ta->trim_tree);
|
|
|
|
}
|
|
|
|
|
|
|
|
kmem_free(tap, sizeof (trim_args_t) * children);
|
|
|
|
}
|
|
|
|
|
|
|
|
spa_config_exit(spa, SCL_CONFIG, FTAG);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* After completing the group of metaslabs wait for the next
|
|
|
|
* open txg. This is done to make sure that a minimum of
|
|
|
|
* zfs_trim_txg_batch txgs will occur before these metaslabs
|
|
|
|
* are trimmed again.
|
|
|
|
*/
|
|
|
|
txg_wait_open(spa_get_dsl(spa), 0, issued_trim);
|
|
|
|
|
|
|
|
shift++;
|
|
|
|
spa_config_enter(spa, SCL_CONFIG, FTAG, RW_READER);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (uint64_t c = 0; c < vd->vdev_children; c++) {
|
|
|
|
vdev_t *cvd = vd->vdev_child[c];
|
|
|
|
mutex_enter(&cvd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
while (cvd->vdev_trim_inflight[1] > 0) {
|
|
|
|
cv_wait(&cvd->vdev_trim_io_cv,
|
|
|
|
&cvd->vdev_trim_io_lock);
|
|
|
|
}
|
|
|
|
mutex_exit(&cvd->vdev_trim_io_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
spa_config_exit(spa, SCL_CONFIG, FTAG);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* When exiting because the autotrim property was set to off, then
|
|
|
|
* abandon any unprocessed ms_trim ranges to reclaim the memory.
|
|
|
|
*/
|
|
|
|
if (spa_get_autotrim(spa) == SPA_AUTOTRIM_OFF) {
|
|
|
|
for (uint64_t i = 0; i < vd->vdev_ms_count; i++) {
|
|
|
|
metaslab_t *msp = vd->vdev_ms[i];
|
|
|
|
|
|
|
|
mutex_enter(&msp->ms_lock);
|
|
|
|
range_tree_vacate(msp->ms_trim, NULL, NULL);
|
|
|
|
mutex_exit(&msp->ms_lock);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_autotrim_lock);
|
|
|
|
ASSERT(vd->vdev_autotrim_thread != NULL);
|
|
|
|
vd->vdev_autotrim_thread = NULL;
|
|
|
|
cv_broadcast(&vd->vdev_autotrim_cv);
|
|
|
|
mutex_exit(&vd->vdev_autotrim_lock);
|
2020-08-05 20:17:07 +03:00
|
|
|
|
|
|
|
thread_exit();
|
2019-03-29 19:13:20 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Starts an autotrim thread, if needed, for each top-level vdev which can be
|
|
|
|
* trimmed. A top-level vdev which has been evacuated will never be trimmed.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_autotrim(spa_t *spa)
|
|
|
|
{
|
|
|
|
vdev_t *root_vd = spa->spa_root_vdev;
|
|
|
|
|
|
|
|
for (uint64_t i = 0; i < root_vd->vdev_children; i++) {
|
|
|
|
vdev_t *tvd = root_vd->vdev_child[i];
|
|
|
|
|
|
|
|
mutex_enter(&tvd->vdev_autotrim_lock);
|
|
|
|
if (vdev_writeable(tvd) && !tvd->vdev_removing &&
|
|
|
|
tvd->vdev_autotrim_thread == NULL) {
|
|
|
|
ASSERT3P(tvd->vdev_top, ==, tvd);
|
|
|
|
|
|
|
|
tvd->vdev_autotrim_thread = thread_create(NULL, 0,
|
|
|
|
vdev_autotrim_thread, tvd, 0, &p0, TS_RUN,
|
|
|
|
maxclsyspri);
|
|
|
|
ASSERT(tvd->vdev_autotrim_thread != NULL);
|
|
|
|
}
|
|
|
|
mutex_exit(&tvd->vdev_autotrim_lock);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Wait for the vdev_autotrim_thread associated with the passed top-level
|
|
|
|
* vdev to be terminated (canceled or stopped).
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_autotrim_stop_wait(vdev_t *tvd)
|
|
|
|
{
|
|
|
|
mutex_enter(&tvd->vdev_autotrim_lock);
|
|
|
|
if (tvd->vdev_autotrim_thread != NULL) {
|
|
|
|
tvd->vdev_autotrim_exit_wanted = B_TRUE;
|
|
|
|
|
|
|
|
while (tvd->vdev_autotrim_thread != NULL) {
|
|
|
|
cv_wait(&tvd->vdev_autotrim_cv,
|
|
|
|
&tvd->vdev_autotrim_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT3P(tvd->vdev_autotrim_thread, ==, NULL);
|
|
|
|
tvd->vdev_autotrim_exit_wanted = B_FALSE;
|
|
|
|
}
|
|
|
|
mutex_exit(&tvd->vdev_autotrim_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Wait for all of the vdev_autotrim_thread associated with the pool to
|
|
|
|
* be terminated (canceled or stopped).
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_autotrim_stop_all(spa_t *spa)
|
|
|
|
{
|
|
|
|
vdev_t *root_vd = spa->spa_root_vdev;
|
|
|
|
|
|
|
|
for (uint64_t i = 0; i < root_vd->vdev_children; i++)
|
|
|
|
vdev_autotrim_stop_wait(root_vd->vdev_child[i]);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Conditionally restart all of the vdev_autotrim_thread's for the pool.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_autotrim_restart(spa_t *spa)
|
|
|
|
{
|
|
|
|
ASSERT(MUTEX_HELD(&spa_namespace_lock));
|
|
|
|
|
|
|
|
if (spa->spa_autotrim)
|
|
|
|
vdev_autotrim(spa);
|
|
|
|
}
|
|
|
|
|
2020-06-09 20:15:08 +03:00
|
|
|
static void
|
|
|
|
vdev_trim_l2arc_thread(void *arg)
|
|
|
|
{
|
|
|
|
vdev_t *vd = arg;
|
|
|
|
spa_t *spa = vd->vdev_spa;
|
|
|
|
l2arc_dev_t *dev = l2arc_vdev_get(vd);
|
|
|
|
trim_args_t ta;
|
|
|
|
range_seg64_t physical_rs;
|
|
|
|
|
|
|
|
ASSERT(vdev_is_concrete(vd));
|
|
|
|
spa_config_enter(spa, SCL_CONFIG, FTAG, RW_READER);
|
|
|
|
|
|
|
|
vd->vdev_trim_last_offset = 0;
|
|
|
|
vd->vdev_trim_rate = 0;
|
|
|
|
vd->vdev_trim_partial = 0;
|
|
|
|
vd->vdev_trim_secure = 0;
|
|
|
|
|
|
|
|
bzero(&ta, sizeof (ta));
|
|
|
|
ta.trim_vdev = vd;
|
|
|
|
ta.trim_tree = range_tree_create(NULL, RANGE_SEG64, NULL, 0, 0);
|
|
|
|
ta.trim_type = TRIM_TYPE_MANUAL;
|
|
|
|
ta.trim_extent_bytes_max = zfs_trim_extent_bytes_max;
|
|
|
|
ta.trim_extent_bytes_min = SPA_MINBLOCKSIZE;
|
|
|
|
ta.trim_flags = 0;
|
|
|
|
|
|
|
|
physical_rs.rs_start = vd->vdev_trim_bytes_done = 0;
|
|
|
|
physical_rs.rs_end = vd->vdev_trim_bytes_est =
|
|
|
|
vdev_get_min_asize(vd);
|
|
|
|
|
|
|
|
range_tree_add(ta.trim_tree, physical_rs.rs_start,
|
|
|
|
physical_rs.rs_end - physical_rs.rs_start);
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
|
|
|
vdev_trim_change_state(vd, VDEV_TRIM_ACTIVE, 0, 0, 0);
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
|
|
|
|
(void) vdev_trim_ranges(&ta);
|
|
|
|
|
|
|
|
spa_config_exit(spa, SCL_CONFIG, FTAG);
|
|
|
|
mutex_enter(&vd->vdev_trim_io_lock);
|
|
|
|
while (vd->vdev_trim_inflight[TRIM_TYPE_MANUAL] > 0) {
|
|
|
|
cv_wait(&vd->vdev_trim_io_cv, &vd->vdev_trim_io_lock);
|
|
|
|
}
|
|
|
|
mutex_exit(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
range_tree_vacate(ta.trim_tree, NULL, NULL);
|
|
|
|
range_tree_destroy(ta.trim_tree);
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
|
|
|
if (!vd->vdev_trim_exit_wanted && vdev_writeable(vd)) {
|
|
|
|
vdev_trim_change_state(vd, VDEV_TRIM_COMPLETE,
|
|
|
|
vd->vdev_trim_rate, vd->vdev_trim_partial,
|
|
|
|
vd->vdev_trim_secure);
|
|
|
|
}
|
|
|
|
ASSERT(vd->vdev_trim_thread != NULL ||
|
|
|
|
vd->vdev_trim_inflight[TRIM_TYPE_MANUAL] == 0);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Drop the vdev_trim_lock while we sync out the txg since it's
|
|
|
|
* possible that a device might be trying to come online and
|
|
|
|
* must check to see if it needs to restart a trim. That thread
|
|
|
|
* will be holding the spa_config_lock which would prevent the
|
|
|
|
* txg_wait_synced from completing. Same strategy as in
|
|
|
|
* vdev_trim_thread().
|
|
|
|
*/
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
txg_wait_synced(spa_get_dsl(vd->vdev_spa), 0);
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update the header of the cache device here, before
|
|
|
|
* broadcasting vdev_trim_cv which may lead to the removal
|
|
|
|
* of the device. The same applies for setting l2ad_trim_all to
|
|
|
|
* false.
|
|
|
|
*/
|
|
|
|
spa_config_enter(vd->vdev_spa, SCL_L2ARC, vd,
|
|
|
|
RW_READER);
|
|
|
|
bzero(dev->l2ad_dev_hdr, dev->l2ad_dev_hdr_asize);
|
|
|
|
l2arc_dev_hdr_update(dev);
|
|
|
|
spa_config_exit(vd->vdev_spa, SCL_L2ARC, vd);
|
|
|
|
|
|
|
|
vd->vdev_trim_thread = NULL;
|
|
|
|
if (vd->vdev_trim_state == VDEV_TRIM_COMPLETE)
|
|
|
|
dev->l2ad_trim_all = B_FALSE;
|
|
|
|
|
|
|
|
cv_broadcast(&vd->vdev_trim_cv);
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
|
|
|
|
thread_exit();
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Punches out TRIM threads for the L2ARC devices in a spa and assigns them
|
|
|
|
* to vd->vdev_trim_thread variable. This facilitates the management of
|
|
|
|
* trimming the whole cache device using TRIM_TYPE_MANUAL upon addition
|
|
|
|
* to a pool or pool creation or when the header of the device is invalid.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
vdev_trim_l2arc(spa_t *spa)
|
|
|
|
{
|
|
|
|
ASSERT(MUTEX_HELD(&spa_namespace_lock));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Locate the spa's l2arc devices and kick off TRIM threads.
|
|
|
|
*/
|
|
|
|
for (int i = 0; i < spa->spa_l2cache.sav_count; i++) {
|
|
|
|
vdev_t *vd = spa->spa_l2cache.sav_vdevs[i];
|
|
|
|
l2arc_dev_t *dev = l2arc_vdev_get(vd);
|
|
|
|
|
|
|
|
if (dev == NULL || !dev->l2ad_trim_all) {
|
|
|
|
/*
|
|
|
|
* Don't attempt TRIM if the vdev is UNAVAIL or if the
|
|
|
|
* cache device was not marked for whole device TRIM
|
|
|
|
* (ie l2arc_trim_ahead = 0, or the L2ARC device header
|
|
|
|
* is valid with trim_state = VDEV_TRIM_COMPLETE and
|
|
|
|
* l2ad_log_entries > 0).
|
|
|
|
*/
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_trim_lock);
|
|
|
|
ASSERT(vd->vdev_ops->vdev_op_leaf);
|
|
|
|
ASSERT(vdev_is_concrete(vd));
|
|
|
|
ASSERT3P(vd->vdev_trim_thread, ==, NULL);
|
|
|
|
ASSERT(!vd->vdev_detached);
|
|
|
|
ASSERT(!vd->vdev_trim_exit_wanted);
|
|
|
|
ASSERT(!vd->vdev_top->vdev_removing);
|
|
|
|
vdev_trim_change_state(vd, VDEV_TRIM_ACTIVE, 0, 0, 0);
|
|
|
|
vd->vdev_trim_thread = thread_create(NULL, 0,
|
|
|
|
vdev_trim_l2arc_thread, vd, 0, &p0, TS_RUN, maxclsyspri);
|
|
|
|
mutex_exit(&vd->vdev_trim_lock);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* A wrapper which calls vdev_trim_ranges(). It is intended to be called
|
|
|
|
* on leaf vdevs.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
vdev_trim_simple(vdev_t *vd, uint64_t start, uint64_t size)
|
|
|
|
{
|
|
|
|
trim_args_t ta;
|
|
|
|
range_seg64_t physical_rs;
|
|
|
|
int error;
|
|
|
|
physical_rs.rs_start = start;
|
|
|
|
physical_rs.rs_end = start + size;
|
|
|
|
|
|
|
|
ASSERT(vdev_is_concrete(vd));
|
|
|
|
ASSERT(vd->vdev_ops->vdev_op_leaf);
|
|
|
|
ASSERT(!vd->vdev_detached);
|
|
|
|
ASSERT(!vd->vdev_top->vdev_removing);
|
|
|
|
|
|
|
|
bzero(&ta, sizeof (ta));
|
|
|
|
ta.trim_vdev = vd;
|
|
|
|
ta.trim_tree = range_tree_create(NULL, RANGE_SEG64, NULL, 0, 0);
|
|
|
|
ta.trim_type = TRIM_TYPE_SIMPLE;
|
|
|
|
ta.trim_extent_bytes_max = zfs_trim_extent_bytes_max;
|
|
|
|
ta.trim_extent_bytes_min = SPA_MINBLOCKSIZE;
|
|
|
|
ta.trim_flags = 0;
|
|
|
|
|
|
|
|
ASSERT3U(physical_rs.rs_end, >=, physical_rs.rs_start);
|
|
|
|
|
|
|
|
if (physical_rs.rs_end > physical_rs.rs_start) {
|
|
|
|
range_tree_add(ta.trim_tree, physical_rs.rs_start,
|
|
|
|
physical_rs.rs_end - physical_rs.rs_start);
|
|
|
|
} else {
|
|
|
|
ASSERT3U(physical_rs.rs_end, ==, physical_rs.rs_start);
|
|
|
|
}
|
|
|
|
|
|
|
|
error = vdev_trim_ranges(&ta);
|
|
|
|
|
|
|
|
mutex_enter(&vd->vdev_trim_io_lock);
|
|
|
|
while (vd->vdev_trim_inflight[TRIM_TYPE_SIMPLE] > 0) {
|
|
|
|
cv_wait(&vd->vdev_trim_io_cv, &vd->vdev_trim_io_lock);
|
|
|
|
}
|
|
|
|
mutex_exit(&vd->vdev_trim_io_lock);
|
|
|
|
|
|
|
|
range_tree_vacate(ta.trim_tree, NULL, NULL);
|
|
|
|
range_tree_destroy(ta.trim_tree);
|
|
|
|
|
|
|
|
return (error);
|
|
|
|
}
|
|
|
|
|
2019-03-29 19:13:20 +03:00
|
|
|
EXPORT_SYMBOL(vdev_trim);
|
|
|
|
EXPORT_SYMBOL(vdev_trim_stop);
|
|
|
|
EXPORT_SYMBOL(vdev_trim_stop_all);
|
|
|
|
EXPORT_SYMBOL(vdev_trim_stop_wait);
|
|
|
|
EXPORT_SYMBOL(vdev_trim_restart);
|
|
|
|
EXPORT_SYMBOL(vdev_autotrim);
|
|
|
|
EXPORT_SYMBOL(vdev_autotrim_stop_all);
|
|
|
|
EXPORT_SYMBOL(vdev_autotrim_stop_wait);
|
|
|
|
EXPORT_SYMBOL(vdev_autotrim_restart);
|
2020-06-09 20:15:08 +03:00
|
|
|
EXPORT_SYMBOL(vdev_trim_l2arc);
|
|
|
|
EXPORT_SYMBOL(vdev_trim_simple);
|
2019-03-29 19:13:20 +03:00
|
|
|
|
|
|
|
/* BEGIN CSTYLED */
|
2019-09-06 00:49:49 +03:00
|
|
|
ZFS_MODULE_PARAM(zfs_trim, zfs_trim_, extent_bytes_max, UINT, ZMOD_RW,
|
2019-03-29 19:13:20 +03:00
|
|
|
"Max size of TRIM commands, larger will be split");
|
|
|
|
|
2019-09-06 00:49:49 +03:00
|
|
|
ZFS_MODULE_PARAM(zfs_trim, zfs_trim_, extent_bytes_min, UINT, ZMOD_RW,
|
2019-03-29 19:13:20 +03:00
|
|
|
"Min size of TRIM commands, smaller will be skipped");
|
|
|
|
|
2019-09-06 00:49:49 +03:00
|
|
|
ZFS_MODULE_PARAM(zfs_trim, zfs_trim_, metaslab_skip, UINT, ZMOD_RW,
|
2019-03-29 19:13:20 +03:00
|
|
|
"Skip metaslabs which have never been initialized");
|
|
|
|
|
2019-09-06 00:49:49 +03:00
|
|
|
ZFS_MODULE_PARAM(zfs_trim, zfs_trim_, txg_batch, UINT, ZMOD_RW,
|
2019-03-29 19:13:20 +03:00
|
|
|
"Min number of txgs to aggregate frees before issuing TRIM");
|
|
|
|
|
2019-09-06 00:49:49 +03:00
|
|
|
ZFS_MODULE_PARAM(zfs_trim, zfs_trim_, queue_limit, UINT, ZMOD_RW,
|
2019-03-29 19:13:20 +03:00
|
|
|
"Max queued TRIMs outstanding per leaf vdev");
|
|
|
|
/* END CSTYLED */
|