mirror of
https://git.proxmox.com/git/mirror_zfs.git
synced 2026-01-25 10:12:13 +03:00
Prefer VERIFY0P(n) over VERIFY3P(n, ==, NULL)
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov> Reviewed-by: Alexander Motin <alexander.motin@TrueNAS.com> Signed-off-by: Rob Norris <robn@despairlabs.com> Sponsored-by: https://despairlabs.com/sponsor/ Closes #17591
This commit is contained in:
parent
f7bdd84328
commit
82d6f7b047
@ -176,7 +176,7 @@ static int
|
||||
sublivelist_verify_blkptr(void *arg, const blkptr_t *bp, boolean_t free,
|
||||
dmu_tx_t *tx)
|
||||
{
|
||||
ASSERT3P(tx, ==, NULL);
|
||||
ASSERT0P(tx);
|
||||
struct sublivelist_verify *sv = arg;
|
||||
sublivelist_verify_block_refcnt_t current = {
|
||||
.svbr_blk = *bp,
|
||||
@ -1455,7 +1455,7 @@ get_obsolete_refcount(vdev_t *vd)
|
||||
refcount++;
|
||||
}
|
||||
} else {
|
||||
ASSERT3P(vd->vdev_obsolete_sm, ==, NULL);
|
||||
ASSERT0P(vd->vdev_obsolete_sm);
|
||||
ASSERT0(obsolete_sm_object);
|
||||
}
|
||||
for (unsigned c = 0; c < vd->vdev_children; c++) {
|
||||
@ -1792,7 +1792,7 @@ print_vdev_indirect(vdev_t *vd)
|
||||
vdev_indirect_births_t *vib = vd->vdev_indirect_births;
|
||||
|
||||
if (vim == NULL) {
|
||||
ASSERT3P(vib, ==, NULL);
|
||||
ASSERT0P(vib);
|
||||
return;
|
||||
}
|
||||
|
||||
@ -3347,7 +3347,7 @@ open_objset(const char *path, const void *tag, objset_t **osp)
|
||||
uint64_t sa_attrs = 0;
|
||||
uint64_t version = 0;
|
||||
|
||||
VERIFY3P(sa_os, ==, NULL);
|
||||
VERIFY0P(sa_os);
|
||||
|
||||
/*
|
||||
* We can't own an objset if it's redacted. Therefore, we do this
|
||||
@ -7016,7 +7016,7 @@ deleted_livelists_count_blocks(spa_t *spa, zdb_cb_t *zbc)
|
||||
static void
|
||||
dump_livelist_cb(dsl_deadlist_t *ll, void *arg)
|
||||
{
|
||||
ASSERT3P(arg, ==, NULL);
|
||||
ASSERT0P(arg);
|
||||
global_feature_count[SPA_FEATURE_LIVELIST]++;
|
||||
dump_blkptr_list(ll, "Deleted Livelist");
|
||||
dsl_deadlist_iterate(ll, sublivelist_verify_lightweight, NULL);
|
||||
@ -7913,7 +7913,7 @@ verify_checkpoint_vdev_spacemaps(spa_t *checkpoint, spa_t *current)
|
||||
for (uint64_t c = ckpoint_rvd->vdev_children;
|
||||
c < current_rvd->vdev_children; c++) {
|
||||
vdev_t *current_vd = current_rvd->vdev_child[c];
|
||||
VERIFY3P(current_vd->vdev_checkpoint_sm, ==, NULL);
|
||||
VERIFY0P(current_vd->vdev_checkpoint_sm);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@ -923,7 +923,7 @@ zfs_do_clone(int argc, char **argv)
|
||||
return (!!ret);
|
||||
|
||||
usage:
|
||||
ASSERT3P(zhp, ==, NULL);
|
||||
ASSERT0P(zhp);
|
||||
nvlist_free(props);
|
||||
usage(B_FALSE);
|
||||
return (-1);
|
||||
|
||||
@ -828,8 +828,8 @@ static char *short_opts = NULL;
|
||||
static void
|
||||
init_options(void)
|
||||
{
|
||||
ASSERT3P(long_opts, ==, NULL);
|
||||
ASSERT3P(short_opts, ==, NULL);
|
||||
ASSERT0P(long_opts);
|
||||
ASSERT0P(short_opts);
|
||||
|
||||
int count = sizeof (option_table) / sizeof (option_table[0]);
|
||||
long_opts = umem_alloc(sizeof (struct option) * count, UMEM_NOFAIL);
|
||||
@ -1685,7 +1685,7 @@ ztest_rll_init(rll_t *rll)
|
||||
static void
|
||||
ztest_rll_destroy(rll_t *rll)
|
||||
{
|
||||
ASSERT3P(rll->rll_writer, ==, NULL);
|
||||
ASSERT0P(rll->rll_writer);
|
||||
ASSERT0(rll->rll_readers);
|
||||
mutex_destroy(&rll->rll_lock);
|
||||
cv_destroy(&rll->rll_cv);
|
||||
@ -1719,7 +1719,7 @@ ztest_rll_unlock(rll_t *rll)
|
||||
rll->rll_writer = NULL;
|
||||
} else {
|
||||
ASSERT3S(rll->rll_readers, >, 0);
|
||||
ASSERT3P(rll->rll_writer, ==, NULL);
|
||||
ASSERT0P(rll->rll_writer);
|
||||
rll->rll_readers--;
|
||||
}
|
||||
|
||||
|
||||
@ -111,7 +111,7 @@ spl_mutex_lockdep_on_maybe(kmutex_t *mp) \
|
||||
#undef mutex_destroy
|
||||
#define mutex_destroy(mp) \
|
||||
{ \
|
||||
VERIFY3P(mutex_owner(mp), ==, NULL); \
|
||||
VERIFY0P(mutex_owner(mp)); \
|
||||
}
|
||||
|
||||
#define mutex_tryenter(mp) \
|
||||
|
||||
@ -302,7 +302,7 @@ abd_iter_at_end(struct abd_iter *aiter)
|
||||
void
|
||||
abd_iter_advance(struct abd_iter *aiter, size_t amount)
|
||||
{
|
||||
ASSERT3P(aiter->iter_mapaddr, ==, NULL);
|
||||
ASSERT0P(aiter->iter_mapaddr);
|
||||
ASSERT0(aiter->iter_mapsize);
|
||||
|
||||
if (abd_iter_at_end(aiter))
|
||||
@ -315,7 +315,7 @@ abd_iter_advance(struct abd_iter *aiter, size_t amount)
|
||||
void
|
||||
abd_iter_map(struct abd_iter *aiter)
|
||||
{
|
||||
ASSERT3P(aiter->iter_mapaddr, ==, NULL);
|
||||
ASSERT0P(aiter->iter_mapaddr);
|
||||
ASSERT0(aiter->iter_mapsize);
|
||||
|
||||
if (abd_iter_at_end(aiter))
|
||||
|
||||
@ -1769,7 +1769,7 @@ zpool_find_import_cached(libpc_handle_t *hdl, importargs_t *iarg)
|
||||
fnvlist_add_nvlist(pools, nvpair_name(pair),
|
||||
fnvpair_value_nvlist(pair));
|
||||
|
||||
VERIFY3P(nvlist_next_nvpair(nv, pair), ==, NULL);
|
||||
VERIFY0P(nvlist_next_nvpair(nv, pair));
|
||||
|
||||
iarg->guid = saved_guid;
|
||||
iarg->poolname = saved_poolname;
|
||||
|
||||
@ -265,7 +265,7 @@ nv_priv_alloc_embedded(nvpriv_t *priv)
|
||||
static int
|
||||
nvt_tab_alloc(nvpriv_t *priv, uint64_t buckets)
|
||||
{
|
||||
ASSERT3P(priv->nvp_hashtable, ==, NULL);
|
||||
ASSERT0P(priv->nvp_hashtable);
|
||||
ASSERT0(priv->nvp_nbuckets);
|
||||
ASSERT0(priv->nvp_nentries);
|
||||
|
||||
@ -334,7 +334,7 @@ nvt_lookup_name_type(const nvlist_t *nvl, const char *name, data_type_t type)
|
||||
i_nvp_t **tab = priv->nvp_hashtable;
|
||||
|
||||
if (tab == NULL) {
|
||||
ASSERT3P(priv->nvp_list, ==, NULL);
|
||||
ASSERT0P(priv->nvp_list);
|
||||
ASSERT0(priv->nvp_nbuckets);
|
||||
ASSERT0(priv->nvp_nentries);
|
||||
return (NULL);
|
||||
@ -540,7 +540,7 @@ nvt_add_nvpair(nvlist_t *nvl, nvpair_t *nvp)
|
||||
|
||||
/* insert link at the beginning of the bucket */
|
||||
i_nvp_t *new_entry = NVPAIR2I_NVP(nvp);
|
||||
ASSERT3P(new_entry->nvi_hashtable_next, ==, NULL);
|
||||
ASSERT0P(new_entry->nvi_hashtable_next);
|
||||
new_entry->nvi_hashtable_next = bucket;
|
||||
// cppcheck-suppress nullPointerRedundantCheck
|
||||
tab[index] = new_entry;
|
||||
|
||||
@ -160,7 +160,7 @@ kmem_cache_create(const char *name, size_t bufsize, size_t align,
|
||||
{
|
||||
kmem_cache_t *cache;
|
||||
|
||||
ASSERT3P(vmp, ==, NULL);
|
||||
ASSERT0P(vmp);
|
||||
|
||||
cache = kmem_alloc(sizeof (*cache), KM_SLEEP);
|
||||
strlcpy(cache->kc_name, name, sizeof (cache->kc_name));
|
||||
|
||||
@ -256,7 +256,7 @@ sysevent_worker(void *arg __unused)
|
||||
* free `ze`, so just inline the free() here -- events have already
|
||||
* been drained.
|
||||
*/
|
||||
VERIFY3P(ze->ze_zevent, ==, NULL);
|
||||
VERIFY0P(ze->ze_zevent);
|
||||
kmem_free(ze, sizeof (zfs_zevent_t));
|
||||
|
||||
kthread_exit();
|
||||
|
||||
@ -507,7 +507,7 @@ abd_iter_at_end(struct abd_iter *aiter)
|
||||
void
|
||||
abd_iter_advance(struct abd_iter *aiter, size_t amount)
|
||||
{
|
||||
ASSERT3P(aiter->iter_mapaddr, ==, NULL);
|
||||
ASSERT0P(aiter->iter_mapaddr);
|
||||
ASSERT0(aiter->iter_mapsize);
|
||||
|
||||
/* There's nothing left to advance to, so do nothing */
|
||||
@ -526,7 +526,7 @@ abd_iter_map(struct abd_iter *aiter)
|
||||
{
|
||||
void *paddr;
|
||||
|
||||
ASSERT3P(aiter->iter_mapaddr, ==, NULL);
|
||||
ASSERT0P(aiter->iter_mapaddr);
|
||||
ASSERT0(aiter->iter_mapsize);
|
||||
|
||||
/* There's nothing left to iterate over, so do nothing */
|
||||
|
||||
@ -1236,7 +1236,7 @@ vdev_geom_io_done(zio_t *zio)
|
||||
struct bio *bp = zio->io_bio;
|
||||
|
||||
if (zio->io_type != ZIO_TYPE_READ && zio->io_type != ZIO_TYPE_WRITE) {
|
||||
ASSERT3P(bp, ==, NULL);
|
||||
ASSERT0P(bp);
|
||||
return;
|
||||
}
|
||||
|
||||
|
||||
@ -1632,7 +1632,7 @@ zfs_acl_ids_create(znode_t *dzp, int flag, vattr_t *vap, cred_t *cr,
|
||||
if (zfsvfs->z_replay == B_FALSE)
|
||||
ASSERT_VOP_ELOCKED(ZTOV(dzp), __func__);
|
||||
} else
|
||||
ASSERT3P(dzp->z_vnode, ==, NULL);
|
||||
ASSERT0P(dzp->z_vnode);
|
||||
memset(acl_ids, 0, sizeof (zfs_acl_ids_t));
|
||||
acl_ids->z_mode = MAKEIMODE(vap->va_type, vap->va_mode);
|
||||
|
||||
@ -2014,7 +2014,7 @@ top:
|
||||
|
||||
error = zfs_aclset_common(zp, aclp, cr, tx);
|
||||
ASSERT0(error);
|
||||
ASSERT3P(zp->z_acl_cached, ==, NULL);
|
||||
ASSERT0P(zp->z_acl_cached);
|
||||
zp->z_acl_cached = aclp;
|
||||
|
||||
if (fuid_dirtied)
|
||||
|
||||
@ -357,7 +357,7 @@ zfsctl_create(zfsvfs_t *zfsvfs)
|
||||
vnode_t *rvp;
|
||||
uint64_t crtime[2];
|
||||
|
||||
ASSERT3P(zfsvfs->z_ctldir, ==, NULL);
|
||||
ASSERT0P(zfsvfs->z_ctldir);
|
||||
|
||||
snapdir = sfs_alloc_node(sizeof (*snapdir), "snapshot", ZFSCTL_INO_ROOT,
|
||||
ZFSCTL_INO_SNAPDIR);
|
||||
@ -1367,7 +1367,7 @@ zfsctl_snapshot_unmount(const char *snapname, int flags __unused)
|
||||
|
||||
int err = getzfsvfs(snapname, &zfsvfs);
|
||||
if (err != 0) {
|
||||
ASSERT3P(zfsvfs, ==, NULL);
|
||||
ASSERT0P(zfsvfs);
|
||||
return (0);
|
||||
}
|
||||
vfsp = zfsvfs->z_vfs;
|
||||
|
||||
@ -1091,7 +1091,7 @@ zfsvfs_setup(zfsvfs_t *zfsvfs, boolean_t mounting)
|
||||
if (mounting) {
|
||||
boolean_t readonly;
|
||||
|
||||
ASSERT3P(zfsvfs->z_kstat.dk_kstats, ==, NULL);
|
||||
ASSERT0P(zfsvfs->z_kstat.dk_kstats);
|
||||
error = dataset_kstats_create(&zfsvfs->z_kstat, zfsvfs->z_os);
|
||||
if (error)
|
||||
return (error);
|
||||
|
||||
@ -1101,7 +1101,7 @@ zfs_create(znode_t *dzp, const char *name, vattr_t *vap, int excl, int mode,
|
||||
zfs_exit(zfsvfs, FTAG);
|
||||
return (error);
|
||||
}
|
||||
ASSERT3P(zp, ==, NULL);
|
||||
ASSERT0P(zp);
|
||||
|
||||
/*
|
||||
* Create a new file object and update the directory
|
||||
@ -1482,7 +1482,7 @@ zfs_mkdir(znode_t *dzp, const char *dirname, vattr_t *vap, znode_t **zpp,
|
||||
zfs_exit(zfsvfs, FTAG);
|
||||
return (error);
|
||||
}
|
||||
ASSERT3P(zp, ==, NULL);
|
||||
ASSERT0P(zp);
|
||||
|
||||
if ((error = zfs_zaccess(dzp, ACE_ADD_SUBDIRECTORY, 0, B_FALSE, cr,
|
||||
mnt_ns))) {
|
||||
|
||||
@ -161,15 +161,15 @@ zfs_znode_cache_destructor(void *buf, void *arg)
|
||||
znode_t *zp = buf;
|
||||
|
||||
ASSERT(!POINTER_IS_VALID(zp->z_zfsvfs));
|
||||
ASSERT3P(zp->z_vnode, ==, NULL);
|
||||
ASSERT0P(zp->z_vnode);
|
||||
ASSERT(!list_link_active(&zp->z_link_node));
|
||||
mutex_destroy(&zp->z_lock);
|
||||
mutex_destroy(&zp->z_acl_lock);
|
||||
rw_destroy(&zp->z_xattr_lock);
|
||||
zfs_rangelock_fini(&zp->z_rangelock);
|
||||
|
||||
ASSERT3P(zp->z_acl_cached, ==, NULL);
|
||||
ASSERT3P(zp->z_xattr_cached, ==, NULL);
|
||||
ASSERT0P(zp->z_acl_cached);
|
||||
ASSERT0P(zp->z_xattr_cached);
|
||||
}
|
||||
|
||||
|
||||
@ -195,7 +195,7 @@ zfs_znode_init(void)
|
||||
/*
|
||||
* Initialize zcache
|
||||
*/
|
||||
ASSERT3P(znode_uma_zone, ==, NULL);
|
||||
ASSERT0P(znode_uma_zone);
|
||||
znode_uma_zone = uma_zcreate("zfs_znode_cache",
|
||||
sizeof (znode_t), zfs_znode_cache_constructor_smr,
|
||||
zfs_znode_cache_destructor_smr, NULL, NULL, 0, 0);
|
||||
@ -224,7 +224,7 @@ zfs_znode_init(void)
|
||||
/*
|
||||
* Initialize zcache
|
||||
*/
|
||||
ASSERT3P(znode_cache, ==, NULL);
|
||||
ASSERT0P(znode_cache);
|
||||
znode_cache = kmem_cache_create("zfs_znode_cache",
|
||||
sizeof (znode_t), 0, zfs_znode_cache_constructor,
|
||||
zfs_znode_cache_destructor, NULL, NULL, NULL, KMC_RECLAIMABLE);
|
||||
@ -353,8 +353,8 @@ zfs_znode_sa_init(zfsvfs_t *zfsvfs, znode_t *zp,
|
||||
ASSERT(!POINTER_IS_VALID(zp->z_zfsvfs) || (zfsvfs == zp->z_zfsvfs));
|
||||
ASSERT(MUTEX_HELD(ZFS_OBJ_MUTEX(zfsvfs, zp->z_id)));
|
||||
|
||||
ASSERT3P(zp->z_sa_hdl, ==, NULL);
|
||||
ASSERT3P(zp->z_acl_cached, ==, NULL);
|
||||
ASSERT0P(zp->z_sa_hdl);
|
||||
ASSERT0P(zp->z_acl_cached);
|
||||
if (sa_hdl == NULL) {
|
||||
VERIFY0(sa_handle_get_from_db(zfsvfs->z_os, db, zp,
|
||||
SA_HDL_SHARED, &zp->z_sa_hdl));
|
||||
@ -1127,7 +1127,7 @@ zfs_rezget(znode_t *zp)
|
||||
}
|
||||
rw_exit(&zp->z_xattr_lock);
|
||||
|
||||
ASSERT3P(zp->z_sa_hdl, ==, NULL);
|
||||
ASSERT0P(zp->z_sa_hdl);
|
||||
err = sa_buf_hold(zfsvfs->z_os, obj_num, NULL, &db);
|
||||
if (err) {
|
||||
ZFS_OBJ_HOLD_EXIT(zfsvfs, obj_num);
|
||||
@ -1298,7 +1298,7 @@ zfs_znode_free(znode_t *zp)
|
||||
zfsvfs_t *zfsvfs = zp->z_zfsvfs;
|
||||
char *symlink;
|
||||
|
||||
ASSERT3P(zp->z_sa_hdl, ==, NULL);
|
||||
ASSERT0P(zp->z_sa_hdl);
|
||||
zp->z_vnode = NULL;
|
||||
mutex_enter(&zfsvfs->z_znodes_lock);
|
||||
POINTER_INVALIDATE(&zp->z_zfsvfs);
|
||||
|
||||
@ -1415,7 +1415,7 @@ zvol_os_free(zvol_state_t *zv)
|
||||
struct zvol_state_geom *zsg = &zv->zv_zso->zso_geom;
|
||||
struct g_provider *pp __maybe_unused = zsg->zsg_provider;
|
||||
|
||||
ASSERT3P(pp->private, ==, NULL);
|
||||
ASSERT0P(pp->private);
|
||||
|
||||
g_topology_lock();
|
||||
zvol_geom_destroy(zv);
|
||||
@ -1425,7 +1425,7 @@ zvol_os_free(zvol_state_t *zv)
|
||||
struct cdev *dev = zsd->zsd_cdev;
|
||||
|
||||
if (dev != NULL) {
|
||||
ASSERT3P(dev->si_drv2, ==, NULL);
|
||||
ASSERT0P(dev->si_drv2);
|
||||
destroy_dev(dev);
|
||||
knlist_clear(&zsd->zsd_selinfo.si_note, 0);
|
||||
knlist_destroy(&zsd->zsd_selinfo.si_note);
|
||||
@ -1493,11 +1493,11 @@ zvol_os_create_minor(const char *name)
|
||||
|
||||
zv->zv_objset = os;
|
||||
|
||||
ASSERT3P(zv->zv_kstat.dk_kstats, ==, NULL);
|
||||
ASSERT0P(zv->zv_kstat.dk_kstats);
|
||||
error = dataset_kstats_create(&zv->zv_kstat, zv->zv_objset);
|
||||
if (error)
|
||||
goto out_dmu_objset_disown;
|
||||
ASSERT3P(zv->zv_zilog, ==, NULL);
|
||||
ASSERT0P(zv->zv_zilog);
|
||||
zv->zv_zilog = zil_open(os, zvol_get_data, &zv->zv_kstat.dk_zil_sums);
|
||||
if (spa_writeable(dmu_objset_spa(os))) {
|
||||
if (zil_replay_disable)
|
||||
|
||||
@ -104,7 +104,7 @@ __cv_destroy(kcondvar_t *cvp)
|
||||
while (cv_destroy_wakeup(cvp) == 0)
|
||||
wait_event_timeout(cvp->cv_destroy, cv_destroy_wakeup(cvp), 1);
|
||||
|
||||
ASSERT3P(cvp->cv_mutex, ==, NULL);
|
||||
ASSERT0P(cvp->cv_mutex);
|
||||
ASSERT3S(atomic_read(&cvp->cv_refs), ==, 0);
|
||||
ASSERT3S(atomic_read(&cvp->cv_waiters), ==, 0);
|
||||
ASSERT3S(waitqueue_active(&cvp->cv_event), ==, 0);
|
||||
|
||||
@ -709,7 +709,7 @@ zone_get_hostid(void *zone)
|
||||
{
|
||||
uint32_t hostid;
|
||||
|
||||
ASSERT3P(zone, ==, NULL);
|
||||
ASSERT0P(zone);
|
||||
|
||||
if (spl_hostid != 0)
|
||||
return ((uint32_t)(spl_hostid & HW_HOSTID_MASK));
|
||||
|
||||
@ -161,7 +161,7 @@ tsd_hash_add(tsd_hash_table_t *table, uint_t key, pid_t pid, void *value)
|
||||
ulong_t hash;
|
||||
int rc = 0;
|
||||
|
||||
ASSERT3P(tsd_hash_search(table, key, pid), ==, NULL);
|
||||
ASSERT0P(tsd_hash_search(table, key, pid));
|
||||
|
||||
/* New entry allocate structure, set value, and add to hash */
|
||||
entry = kmem_alloc(sizeof (tsd_hash_entry_t), KM_PUSHPAGE);
|
||||
|
||||
@ -863,9 +863,9 @@ abd_iter_advance(struct abd_iter *aiter, size_t amount)
|
||||
* Ensure that last chunk is not in use. abd_iterate_*() must clear
|
||||
* this state (directly or abd_iter_unmap()) before advancing.
|
||||
*/
|
||||
ASSERT3P(aiter->iter_mapaddr, ==, NULL);
|
||||
ASSERT0P(aiter->iter_mapaddr);
|
||||
ASSERT0(aiter->iter_mapsize);
|
||||
ASSERT3P(aiter->iter_page, ==, NULL);
|
||||
ASSERT0P(aiter->iter_page);
|
||||
ASSERT0(aiter->iter_page_doff);
|
||||
ASSERT0(aiter->iter_page_dsize);
|
||||
|
||||
@ -897,7 +897,7 @@ abd_iter_map(struct abd_iter *aiter)
|
||||
void *paddr;
|
||||
size_t offset = 0;
|
||||
|
||||
ASSERT3P(aiter->iter_mapaddr, ==, NULL);
|
||||
ASSERT0P(aiter->iter_mapaddr);
|
||||
ASSERT0(aiter->iter_mapsize);
|
||||
|
||||
/* There's nothing left to iterate over, so do nothing */
|
||||
|
||||
@ -552,7 +552,7 @@ vdev_bio_associate_blkg(struct bio *bio)
|
||||
#endif
|
||||
|
||||
ASSERT3P(q, !=, NULL);
|
||||
ASSERT3P(bio->bi_blkg, ==, NULL);
|
||||
ASSERT0P(bio->bi_blkg);
|
||||
|
||||
if (q->root_blkg && vdev_blkg_tryget(q->root_blkg))
|
||||
bio->bi_blkg = q->root_blkg;
|
||||
@ -574,7 +574,7 @@ vdev_bio_set_dev(struct bio *bio, struct block_device *bdev)
|
||||
bio->bi_bdev = bdev;
|
||||
|
||||
ASSERT3P(q, !=, NULL);
|
||||
ASSERT3P(bio->bi_blkg, ==, NULL);
|
||||
ASSERT0P(bio->bi_blkg);
|
||||
|
||||
if (q->root_blkg && vdev_blkg_tryget(q->root_blkg))
|
||||
bio->bi_blkg = q->root_blkg;
|
||||
@ -806,7 +806,7 @@ vbio_completion(struct bio *bio)
|
||||
* here; instead we stash vbio on the zio and take care of it in the
|
||||
* done callback.
|
||||
*/
|
||||
ASSERT3P(zio->io_bio, ==, NULL);
|
||||
ASSERT0P(zio->io_bio);
|
||||
zio->io_bio = vbio;
|
||||
|
||||
zio_delay_interrupt(zio);
|
||||
|
||||
@ -494,9 +494,9 @@ zfsctl_inode_alloc(zfsvfs_t *zfsvfs, uint64_t id,
|
||||
if (!creation)
|
||||
now = current_time(ip);
|
||||
zp = ITOZ(ip);
|
||||
ASSERT3P(zp->z_dirlocks, ==, NULL);
|
||||
ASSERT3P(zp->z_acl_cached, ==, NULL);
|
||||
ASSERT3P(zp->z_xattr_cached, ==, NULL);
|
||||
ASSERT0P(zp->z_dirlocks);
|
||||
ASSERT0P(zp->z_acl_cached);
|
||||
ASSERT0P(zp->z_xattr_cached);
|
||||
zp->z_id = id;
|
||||
zp->z_unlinked = B_FALSE;
|
||||
zp->z_atime_dirty = B_FALSE;
|
||||
|
||||
@ -883,7 +883,7 @@ zfsvfs_setup(zfsvfs_t *zfsvfs, boolean_t mounting)
|
||||
* operations out since we closed the ZIL.
|
||||
*/
|
||||
if (mounting) {
|
||||
ASSERT3P(zfsvfs->z_kstat.dk_kstats, ==, NULL);
|
||||
ASSERT0P(zfsvfs->z_kstat.dk_kstats);
|
||||
error = dataset_kstats_create(&zfsvfs->z_kstat, zfsvfs->z_os);
|
||||
if (error)
|
||||
return (error);
|
||||
|
||||
@ -3157,7 +3157,7 @@ top:
|
||||
* zfs_link_create() to add back the same entry, but with a new
|
||||
* dnode (szp), should not fail.
|
||||
*/
|
||||
ASSERT3P(tzp, ==, NULL);
|
||||
ASSERT0P(tzp);
|
||||
goto commit_link_tzp;
|
||||
}
|
||||
|
||||
|
||||
@ -144,9 +144,9 @@ zfs_znode_cache_destructor(void *buf, void *arg)
|
||||
rw_destroy(&zp->z_xattr_lock);
|
||||
zfs_rangelock_fini(&zp->z_rangelock);
|
||||
|
||||
ASSERT3P(zp->z_dirlocks, ==, NULL);
|
||||
ASSERT3P(zp->z_acl_cached, ==, NULL);
|
||||
ASSERT3P(zp->z_xattr_cached, ==, NULL);
|
||||
ASSERT0P(zp->z_dirlocks);
|
||||
ASSERT0P(zp->z_acl_cached);
|
||||
ASSERT0P(zp->z_xattr_cached);
|
||||
}
|
||||
|
||||
static int
|
||||
@ -531,8 +531,8 @@ zfs_znode_alloc(zfsvfs_t *zfsvfs, dmu_buf_t *db, int blksz,
|
||||
|
||||
zp = ITOZ(ip);
|
||||
ASSERT0P(zp->z_dirlocks);
|
||||
ASSERT3P(zp->z_acl_cached, ==, NULL);
|
||||
ASSERT3P(zp->z_xattr_cached, ==, NULL);
|
||||
ASSERT0P(zp->z_acl_cached);
|
||||
ASSERT0P(zp->z_xattr_cached);
|
||||
zp->z_unlinked = B_FALSE;
|
||||
zp->z_atime_dirty = B_FALSE;
|
||||
zp->z_is_ctldir = B_FALSE;
|
||||
|
||||
@ -1494,7 +1494,7 @@ zpl_posix_acl_free(void *arg)
|
||||
acl_rel_head = NULL;
|
||||
if (cmpxchg(&acl_rel_tail, &a->next,
|
||||
&acl_rel_head) == &a->next) {
|
||||
ASSERT3P(a->next, ==, NULL);
|
||||
ASSERT0P(a->next);
|
||||
a->next = freelist;
|
||||
freelist = a;
|
||||
break;
|
||||
@ -1544,7 +1544,7 @@ zpl_posix_acl_release_impl(struct posix_acl *acl)
|
||||
a->time = ddi_get_lbolt();
|
||||
/* atomically points tail to us and get the previous tail */
|
||||
prev = xchg(&acl_rel_tail, &a->next);
|
||||
ASSERT3P(*prev, ==, NULL);
|
||||
ASSERT0P(*prev);
|
||||
*prev = a;
|
||||
/* if it was empty before, schedule the free task */
|
||||
if (prev == &acl_rel_head)
|
||||
|
||||
@ -1426,7 +1426,7 @@ zvol_os_free(zvol_state_t *zv)
|
||||
ASSERT(!RW_LOCK_HELD(&zv->zv_suspend_lock));
|
||||
ASSERT(!MUTEX_HELD(&zv->zv_state_lock));
|
||||
ASSERT0(zv->zv_open_count);
|
||||
ASSERT3P(zv->zv_zso->zvo_disk->private_data, ==, NULL);
|
||||
ASSERT0P(zv->zv_zso->zvo_disk->private_data);
|
||||
|
||||
rw_destroy(&zv->zv_suspend_lock);
|
||||
zfs_rangelock_fini(&zv->zv_rangelock);
|
||||
@ -1649,11 +1649,11 @@ zvol_os_create_minor(const char *name)
|
||||
blk_queue_flag_set(QUEUE_FLAG_SCSI_PASSTHROUGH, zv->zv_zso->zvo_queue);
|
||||
#endif
|
||||
|
||||
ASSERT3P(zv->zv_kstat.dk_kstats, ==, NULL);
|
||||
ASSERT0P(zv->zv_kstat.dk_kstats);
|
||||
error = dataset_kstats_create(&zv->zv_kstat, zv->zv_objset);
|
||||
if (error)
|
||||
goto out_dmu_objset_disown;
|
||||
ASSERT3P(zv->zv_zilog, ==, NULL);
|
||||
ASSERT0P(zv->zv_zilog);
|
||||
zv->zv_zilog = zil_open(os, zvol_get_data, &zv->zv_kstat.dk_zil_sums);
|
||||
if (spa_writeable(dmu_objset_spa(os))) {
|
||||
if (zil_replay_disable)
|
||||
|
||||
@ -2239,8 +2239,8 @@ arc_evictable_space_increment(arc_buf_hdr_t *hdr, arc_state_t *state)
|
||||
ASSERT(HDR_HAS_L1HDR(hdr));
|
||||
|
||||
if (GHOST_STATE(state)) {
|
||||
ASSERT3P(hdr->b_l1hdr.b_buf, ==, NULL);
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_buf);
|
||||
ASSERT0P(hdr->b_l1hdr.b_pabd);
|
||||
ASSERT(!HDR_HAS_RABD(hdr));
|
||||
(void) zfs_refcount_add_many(&state->arcs_esize[type],
|
||||
HDR_GET_LSIZE(hdr), hdr);
|
||||
@ -2278,8 +2278,8 @@ arc_evictable_space_decrement(arc_buf_hdr_t *hdr, arc_state_t *state)
|
||||
ASSERT(HDR_HAS_L1HDR(hdr));
|
||||
|
||||
if (GHOST_STATE(state)) {
|
||||
ASSERT3P(hdr->b_l1hdr.b_buf, ==, NULL);
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_buf);
|
||||
ASSERT0P(hdr->b_l1hdr.b_pabd);
|
||||
ASSERT(!HDR_HAS_RABD(hdr));
|
||||
(void) zfs_refcount_remove_many(&state->arcs_esize[type],
|
||||
HDR_GET_LSIZE(hdr), hdr);
|
||||
@ -2319,7 +2319,7 @@ add_reference(arc_buf_hdr_t *hdr, const void *tag)
|
||||
if (!HDR_EMPTY(hdr) && !MUTEX_HELD(HDR_LOCK(hdr))) {
|
||||
ASSERT(state == arc_anon);
|
||||
ASSERT(zfs_refcount_is_zero(&hdr->b_l1hdr.b_refcnt));
|
||||
ASSERT3P(hdr->b_l1hdr.b_buf, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_buf);
|
||||
}
|
||||
|
||||
if ((zfs_refcount_add(&hdr->b_l1hdr.b_refcnt, tag) == 1) &&
|
||||
@ -2503,7 +2503,7 @@ arc_change_state(arc_state_t *new_state, arc_buf_hdr_t *hdr)
|
||||
(void) zfs_refcount_add_many(
|
||||
&new_state->arcs_size[type],
|
||||
HDR_GET_LSIZE(hdr), hdr);
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_pabd);
|
||||
ASSERT(!HDR_HAS_RABD(hdr));
|
||||
} else {
|
||||
|
||||
@ -2547,7 +2547,7 @@ arc_change_state(arc_state_t *new_state, arc_buf_hdr_t *hdr)
|
||||
if (update_old && old_state != arc_l2c_only) {
|
||||
ASSERT(HDR_HAS_L1HDR(hdr));
|
||||
if (GHOST_STATE(old_state)) {
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_pabd);
|
||||
ASSERT(!HDR_HAS_RABD(hdr));
|
||||
|
||||
/*
|
||||
@ -2758,7 +2758,7 @@ arc_buf_alloc_impl(arc_buf_hdr_t *hdr, spa_t *spa, const zbookmark_phys_t *zb,
|
||||
VERIFY(hdr->b_type == ARC_BUFC_DATA ||
|
||||
hdr->b_type == ARC_BUFC_METADATA);
|
||||
ASSERT3P(ret, !=, NULL);
|
||||
ASSERT3P(*ret, ==, NULL);
|
||||
ASSERT0P(*ret);
|
||||
IMPLY(encrypted, compressed);
|
||||
|
||||
buf = *ret = kmem_cache_alloc(buf_cache, KM_PUSHPAGE);
|
||||
@ -2982,7 +2982,7 @@ static void
|
||||
arc_share_buf(arc_buf_hdr_t *hdr, arc_buf_t *buf)
|
||||
{
|
||||
ASSERT(arc_can_share(hdr, buf));
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_pabd);
|
||||
ASSERT(!ARC_BUF_ENCRYPTED(buf));
|
||||
ASSERT(HDR_EMPTY_OR_LOCKED(hdr));
|
||||
|
||||
@ -3201,14 +3201,14 @@ arc_hdr_alloc_abd(arc_buf_hdr_t *hdr, int alloc_flags)
|
||||
|
||||
if (alloc_rdata) {
|
||||
size = HDR_GET_PSIZE(hdr);
|
||||
ASSERT3P(hdr->b_crypt_hdr.b_rabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_crypt_hdr.b_rabd);
|
||||
hdr->b_crypt_hdr.b_rabd = arc_get_data_abd(hdr, size, hdr,
|
||||
alloc_flags);
|
||||
ASSERT3P(hdr->b_crypt_hdr.b_rabd, !=, NULL);
|
||||
ARCSTAT_INCR(arcstat_raw_size, size);
|
||||
} else {
|
||||
size = arc_hdr_size(hdr);
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_pabd);
|
||||
hdr->b_l1hdr.b_pabd = arc_get_data_abd(hdr, size, hdr,
|
||||
alloc_flags);
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, !=, NULL);
|
||||
@ -3290,7 +3290,7 @@ arc_hdr_alloc(uint64_t spa, int32_t psize, int32_t lsize,
|
||||
|
||||
ASSERT(HDR_EMPTY(hdr));
|
||||
#ifdef ZFS_DEBUG
|
||||
ASSERT3P(hdr->b_l1hdr.b_freeze_cksum, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_freeze_cksum);
|
||||
#endif
|
||||
HDR_SET_PSIZE(hdr, psize);
|
||||
HDR_SET_LSIZE(hdr, lsize);
|
||||
@ -3351,12 +3351,12 @@ arc_hdr_realloc(arc_buf_hdr_t *hdr, kmem_cache_t *old, kmem_cache_t *new)
|
||||
nhdr->b_l1hdr.b_state = arc_l2c_only;
|
||||
|
||||
/* Verify previous threads set to NULL before freeing */
|
||||
ASSERT3P(nhdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(nhdr->b_l1hdr.b_pabd);
|
||||
ASSERT(!HDR_HAS_RABD(hdr));
|
||||
} else {
|
||||
ASSERT3P(hdr->b_l1hdr.b_buf, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_buf);
|
||||
#ifdef ZFS_DEBUG
|
||||
ASSERT3P(hdr->b_l1hdr.b_freeze_cksum, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_freeze_cksum);
|
||||
#endif
|
||||
|
||||
/*
|
||||
@ -3375,7 +3375,7 @@ arc_hdr_realloc(arc_buf_hdr_t *hdr, kmem_cache_t *old, kmem_cache_t *new)
|
||||
* might try to be accessed, even though it was removed.
|
||||
*/
|
||||
VERIFY(!HDR_L2_WRITING(hdr));
|
||||
VERIFY3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
VERIFY0P(hdr->b_l1hdr.b_pabd);
|
||||
ASSERT(!HDR_HAS_RABD(hdr));
|
||||
|
||||
arc_hdr_clear_flags(nhdr, ARC_FLAG_HAS_L1HDR);
|
||||
@ -3698,12 +3698,12 @@ arc_hdr_destroy(arc_buf_hdr_t *hdr)
|
||||
arc_hdr_free_abd(hdr, B_TRUE);
|
||||
}
|
||||
|
||||
ASSERT3P(hdr->b_hash_next, ==, NULL);
|
||||
ASSERT0P(hdr->b_hash_next);
|
||||
if (HDR_HAS_L1HDR(hdr)) {
|
||||
ASSERT(!multilist_link_active(&hdr->b_l1hdr.b_arc_node));
|
||||
ASSERT3P(hdr->b_l1hdr.b_acb, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_acb);
|
||||
#ifdef ZFS_DEBUG
|
||||
ASSERT3P(hdr->b_l1hdr.b_freeze_cksum, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_freeze_cksum);
|
||||
#endif
|
||||
kmem_cache_free(hdr_full_cache, hdr);
|
||||
} else {
|
||||
@ -3771,7 +3771,7 @@ arc_evict_hdr(arc_buf_hdr_t *hdr, uint64_t *real_evicted)
|
||||
ASSERT(MUTEX_HELD(HDR_LOCK(hdr)));
|
||||
ASSERT(HDR_HAS_L1HDR(hdr));
|
||||
ASSERT(!HDR_IO_IN_PROGRESS(hdr));
|
||||
ASSERT3P(hdr->b_l1hdr.b_buf, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_buf);
|
||||
ASSERT0(zfs_refcount_count(&hdr->b_l1hdr.b_refcnt));
|
||||
|
||||
*real_evicted = 0;
|
||||
@ -6132,14 +6132,14 @@ top:
|
||||
}
|
||||
|
||||
if (GHOST_STATE(hdr->b_l1hdr.b_state)) {
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_pabd);
|
||||
ASSERT(!HDR_HAS_RABD(hdr));
|
||||
ASSERT(!HDR_IO_IN_PROGRESS(hdr));
|
||||
ASSERT0(zfs_refcount_count(
|
||||
&hdr->b_l1hdr.b_refcnt));
|
||||
ASSERT3P(hdr->b_l1hdr.b_buf, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_buf);
|
||||
#ifdef ZFS_DEBUG
|
||||
ASSERT3P(hdr->b_l1hdr.b_freeze_cksum, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_freeze_cksum);
|
||||
#endif
|
||||
} else if (HDR_IO_IN_PROGRESS(hdr)) {
|
||||
/*
|
||||
@ -6233,7 +6233,7 @@ top:
|
||||
acb->acb_nobuf = no_buf;
|
||||
acb->acb_zb = *zb;
|
||||
|
||||
ASSERT3P(hdr->b_l1hdr.b_acb, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_acb);
|
||||
hdr->b_l1hdr.b_acb = acb;
|
||||
|
||||
if (HDR_HAS_L2HDR(hdr) &&
|
||||
@ -6717,7 +6717,7 @@ arc_release(arc_buf_t *buf, const void *tag)
|
||||
|
||||
nhdr = arc_hdr_alloc(spa, psize, lsize, protected,
|
||||
compress, hdr->b_complevel, type);
|
||||
ASSERT3P(nhdr->b_l1hdr.b_buf, ==, NULL);
|
||||
ASSERT0P(nhdr->b_l1hdr.b_buf);
|
||||
ASSERT0(zfs_refcount_count(&nhdr->b_l1hdr.b_refcnt));
|
||||
VERIFY3U(nhdr->b_type, ==, type);
|
||||
ASSERT(!HDR_SHARED_DATA(nhdr));
|
||||
@ -6804,7 +6804,7 @@ arc_write_ready(zio_t *zio)
|
||||
if (HDR_HAS_RABD(hdr))
|
||||
arc_hdr_free_abd(hdr, B_TRUE);
|
||||
}
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_pabd);
|
||||
ASSERT(!HDR_HAS_RABD(hdr));
|
||||
ASSERT(!HDR_SHARED_DATA(hdr));
|
||||
ASSERT(!arc_buf_is_shared(buf));
|
||||
@ -6948,7 +6948,7 @@ arc_write_done(zio_t *zio)
|
||||
arc_buf_t *buf = callback->awcb_buf;
|
||||
arc_buf_hdr_t *hdr = buf->b_hdr;
|
||||
|
||||
ASSERT3P(hdr->b_l1hdr.b_acb, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_acb);
|
||||
|
||||
if (zio->io_error == 0) {
|
||||
arc_hdr_verify(hdr, zio->io_bp);
|
||||
@ -6994,7 +6994,7 @@ arc_write_done(zio_t *zio)
|
||||
arc_hdr_destroy(exists);
|
||||
mutex_exit(hash_lock);
|
||||
exists = buf_hash_insert(hdr, &hash_lock);
|
||||
ASSERT3P(exists, ==, NULL);
|
||||
ASSERT0P(exists);
|
||||
} else if (zio->io_flags & ZIO_FLAG_NOPWRITE) {
|
||||
/* nopwrite */
|
||||
ASSERT(zio->io_prop.zp_nopwrite);
|
||||
@ -7044,7 +7044,7 @@ arc_write(zio_t *pio, spa_t *spa, uint64_t txg,
|
||||
ASSERT3P(done, !=, NULL);
|
||||
ASSERT(!HDR_IO_ERROR(hdr));
|
||||
ASSERT(!HDR_IO_IN_PROGRESS(hdr));
|
||||
ASSERT3P(hdr->b_l1hdr.b_acb, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_acb);
|
||||
ASSERT3P(hdr->b_l1hdr.b_buf, !=, NULL);
|
||||
if (uncached)
|
||||
arc_hdr_set_flags(hdr, ARC_FLAG_UNCACHED);
|
||||
@ -7113,7 +7113,7 @@ arc_write(zio_t *pio, spa_t *spa, uint64_t txg,
|
||||
arc_hdr_set_compress(hdr, ZIO_COMPRESS_OFF);
|
||||
|
||||
ASSERT(!arc_buf_is_shared(buf));
|
||||
ASSERT3P(hdr->b_l1hdr.b_pabd, ==, NULL);
|
||||
ASSERT0P(hdr->b_l1hdr.b_pabd);
|
||||
|
||||
zio = zio_write(pio, spa, txg, bp,
|
||||
abd_get_from_buf(buf->b_data, HDR_GET_LSIZE(hdr)),
|
||||
|
||||
@ -478,7 +478,7 @@ bpobj_iterate_impl(bpobj_t *initial_bpo, bpobj_itor_t func, void *arg,
|
||||
* We have unprocessed subobjs. Process the next one.
|
||||
*/
|
||||
ASSERT(bpo->bpo_havecomp);
|
||||
ASSERT3P(bpobj_size, ==, NULL);
|
||||
ASSERT0P(bpobj_size);
|
||||
|
||||
/* Add the last subobj to stack. */
|
||||
int64_t i = bpi->bpi_unprocessed_subobjs - 1;
|
||||
|
||||
@ -1110,7 +1110,7 @@ zfs_btree_add_idx(zfs_btree_t *tree, const void *value,
|
||||
if (where->bti_node == NULL) {
|
||||
ASSERT3U(tree->bt_num_elems, ==, 1);
|
||||
ASSERT3S(tree->bt_height, ==, -1);
|
||||
ASSERT3P(tree->bt_root, ==, NULL);
|
||||
ASSERT0P(tree->bt_root);
|
||||
ASSERT0(where->bti_offset);
|
||||
|
||||
tree->bt_num_nodes++;
|
||||
@ -1947,7 +1947,7 @@ void
|
||||
zfs_btree_destroy(zfs_btree_t *tree)
|
||||
{
|
||||
ASSERT0(tree->bt_num_elems);
|
||||
ASSERT3P(tree->bt_root, ==, NULL);
|
||||
ASSERT0P(tree->bt_root);
|
||||
}
|
||||
|
||||
/* Verify that every child of this node has the correct parent pointer. */
|
||||
@ -1969,10 +1969,10 @@ static void
|
||||
zfs_btree_verify_pointers(zfs_btree_t *tree)
|
||||
{
|
||||
if (tree->bt_height == -1) {
|
||||
VERIFY3P(tree->bt_root, ==, NULL);
|
||||
VERIFY0P(tree->bt_root);
|
||||
return;
|
||||
}
|
||||
VERIFY3P(tree->bt_root->bth_parent, ==, NULL);
|
||||
VERIFY0P(tree->bt_root->bth_parent);
|
||||
zfs_btree_verify_pointers_helper(tree, tree->bt_root);
|
||||
}
|
||||
|
||||
|
||||
@ -1259,7 +1259,7 @@ dbuf_clear_data(dmu_buf_impl_t *db)
|
||||
{
|
||||
ASSERT(MUTEX_HELD(&db->db_mtx));
|
||||
dbuf_evict_user(db);
|
||||
ASSERT3P(db->db_buf, ==, NULL);
|
||||
ASSERT0P(db->db_buf);
|
||||
db->db.db_data = NULL;
|
||||
if (db->db_state != DB_NOFILL) {
|
||||
db->db_state = DB_UNCACHED;
|
||||
@ -1390,7 +1390,7 @@ dbuf_read_done(zio_t *zio, const zbookmark_phys_t *zb, const blkptr_t *bp,
|
||||
/* i/o error */
|
||||
ASSERT(zio == NULL || zio->io_error != 0);
|
||||
ASSERT(db->db_blkid != DMU_BONUS_BLKID);
|
||||
ASSERT3P(db->db_buf, ==, NULL);
|
||||
ASSERT0P(db->db_buf);
|
||||
db->db_state = DB_UNCACHED;
|
||||
DTRACE_SET_STATE(db, "i/o error");
|
||||
} else if (db->db_level == 0 && db->db_freed_in_flight) {
|
||||
@ -1994,7 +1994,7 @@ dbuf_free_range(dnode_t *dn, uint64_t start_blkid, uint64_t end_blkid,
|
||||
|
||||
mutex_enter(&dn->dn_dbufs_mtx);
|
||||
db = avl_find(&dn->dn_dbufs, db_search, &where);
|
||||
ASSERT3P(db, ==, NULL);
|
||||
ASSERT0P(db);
|
||||
|
||||
db = avl_nearest(&dn->dn_dbufs, where, AVL_AFTER);
|
||||
|
||||
@ -2896,8 +2896,8 @@ dmu_buf_will_clone_or_dio(dmu_buf_t *db_fake, dmu_tx_t *tx)
|
||||
dbuf_clear_data(db);
|
||||
}
|
||||
|
||||
ASSERT3P(db->db_buf, ==, NULL);
|
||||
ASSERT3P(db->db.db_data, ==, NULL);
|
||||
ASSERT0P(db->db_buf);
|
||||
ASSERT0P(db->db.db_data);
|
||||
|
||||
db->db_state = DB_NOFILL;
|
||||
DTRACE_SET_STATE(db,
|
||||
@ -3960,7 +3960,7 @@ dbuf_hold_impl(dnode_t *dn, uint8_t level, uint64_t blkid,
|
||||
if (fail_uncached)
|
||||
return (SET_ERROR(ENOENT));
|
||||
|
||||
ASSERT3P(parent, ==, NULL);
|
||||
ASSERT0P(parent);
|
||||
err = dbuf_findbp(dn, level, blkid, fail_sparse, &parent, &bp);
|
||||
if (fail_sparse) {
|
||||
if (err == 0 && bp && BP_IS_HOLE(bp))
|
||||
@ -4596,7 +4596,7 @@ static blkptr_t *
|
||||
dbuf_lightweight_bp(dbuf_dirty_record_t *dr)
|
||||
{
|
||||
/* This must be a lightweight dirty record. */
|
||||
ASSERT3P(dr->dr_dbuf, ==, NULL);
|
||||
ASSERT0P(dr->dr_dbuf);
|
||||
dnode_t *dn = dr->dr_dnode;
|
||||
|
||||
if (dn->dn_phys->dn_nlevels == 1) {
|
||||
@ -4739,7 +4739,7 @@ dbuf_sync_leaf(dbuf_dirty_record_t *dr, dmu_tx_t *tx)
|
||||
*/
|
||||
if (db->db_state == DB_UNCACHED) {
|
||||
/* This buffer has been freed since it was dirtied */
|
||||
ASSERT3P(db->db.db_data, ==, NULL);
|
||||
ASSERT0P(db->db.db_data);
|
||||
} else if (db->db_state == DB_FILL) {
|
||||
/* This buffer was freed and is now being re-filled */
|
||||
ASSERT(db->db.db_data != dr->dt.dl.dr_data);
|
||||
@ -4756,9 +4756,9 @@ dbuf_sync_leaf(dbuf_dirty_record_t *dr, dmu_tx_t *tx)
|
||||
*/
|
||||
dbuf_dirty_record_t *dr_head =
|
||||
list_head(&db->db_dirty_records);
|
||||
ASSERT3P(db->db_buf, ==, NULL);
|
||||
ASSERT3P(db->db.db_data, ==, NULL);
|
||||
ASSERT3P(dr_head->dt.dl.dr_data, ==, NULL);
|
||||
ASSERT0P(db->db_buf);
|
||||
ASSERT0P(db->db.db_data);
|
||||
ASSERT0P(dr_head->dt.dl.dr_data);
|
||||
ASSERT3U(dr_head->dt.dl.dr_override_state, ==, DR_OVERRIDDEN);
|
||||
} else {
|
||||
ASSERT(db->db_state == DB_CACHED || db->db_state == DB_NOFILL);
|
||||
|
||||
@ -1011,7 +1011,7 @@ ddt_free(const ddt_t *ddt, ddt_entry_t *dde)
|
||||
{
|
||||
if (dde->dde_io != NULL) {
|
||||
for (int p = 0; p < DDT_NPHYS(ddt); p++)
|
||||
ASSERT3P(dde->dde_io->dde_lead_zio[p], ==, NULL);
|
||||
ASSERT0P(dde->dde_io->dde_lead_zio[p]);
|
||||
|
||||
if (dde->dde_io->dde_repair_abd != NULL)
|
||||
abd_free(dde->dde_io->dde_repair_abd);
|
||||
@ -2395,7 +2395,7 @@ ddt_sync(spa_t *spa, uint64_t txg)
|
||||
* scan's root zio here so that we can wait for any scan IOs in
|
||||
* addition to the regular ddt IOs.
|
||||
*/
|
||||
ASSERT3P(scn->scn_zio_root, ==, NULL);
|
||||
ASSERT0P(scn->scn_zio_root);
|
||||
scn->scn_zio_root = rio;
|
||||
|
||||
for (enum zio_checksum c = 0; c < ZIO_CHECKSUM_FUNCTIONS; c++) {
|
||||
|
||||
@ -194,7 +194,7 @@ void
|
||||
ddt_log_begin(ddt_t *ddt, size_t nentries, dmu_tx_t *tx, ddt_log_update_t *dlu)
|
||||
{
|
||||
ASSERT3U(nentries, >, 0);
|
||||
ASSERT3P(dlu->dlu_dbp, ==, NULL);
|
||||
ASSERT0P(dlu->dlu_dbp);
|
||||
|
||||
if (ddt->ddt_log_active->ddl_object == 0)
|
||||
ddt_log_create(ddt, tx);
|
||||
@ -748,8 +748,8 @@ ddt_log_load(ddt_t *ddt)
|
||||
void
|
||||
ddt_log_alloc(ddt_t *ddt)
|
||||
{
|
||||
ASSERT3P(ddt->ddt_log_active, ==, NULL);
|
||||
ASSERT3P(ddt->ddt_log_flushing, ==, NULL);
|
||||
ASSERT0P(ddt->ddt_log_active);
|
||||
ASSERT0P(ddt->ddt_log_flushing);
|
||||
|
||||
avl_create(&ddt->ddt_log[0].ddl_tree, ddt_key_compare,
|
||||
sizeof (ddt_log_entry_t), offsetof(ddt_log_entry_t, ddle_node));
|
||||
|
||||
@ -95,9 +95,9 @@ dmu_write_direct_done(zio_t *zio)
|
||||
abd_free(zio->io_abd);
|
||||
|
||||
mutex_enter(&db->db_mtx);
|
||||
ASSERT3P(db->db_buf, ==, NULL);
|
||||
ASSERT3P(dr->dt.dl.dr_data, ==, NULL);
|
||||
ASSERT3P(db->db.db_data, ==, NULL);
|
||||
ASSERT0P(db->db_buf);
|
||||
ASSERT0P(dr->dt.dl.dr_data);
|
||||
ASSERT0P(db->db.db_data);
|
||||
db->db_state = DB_UNCACHED;
|
||||
mutex_exit(&db->db_mtx);
|
||||
|
||||
|
||||
@ -90,7 +90,7 @@ dmu_object_alloc_impl(objset_t *os, dmu_object_type_t ot, int blocksize,
|
||||
if (allocated_dnode != NULL) {
|
||||
ASSERT3P(tag, !=, NULL);
|
||||
} else {
|
||||
ASSERT3P(tag, ==, NULL);
|
||||
ASSERT0P(tag);
|
||||
tag = FTAG;
|
||||
}
|
||||
|
||||
|
||||
@ -866,7 +866,7 @@ dmu_recv_begin_sync(void *arg, dmu_tx_t *tx)
|
||||
*/
|
||||
if (dcp == NULL && drrb->drr_fromguid == 0 &&
|
||||
drba->drba_origin == NULL) {
|
||||
ASSERT3P(dcp, ==, NULL);
|
||||
ASSERT0P(dcp);
|
||||
dcp = &dummy_dcp;
|
||||
|
||||
if (featureflags & DMU_BACKUP_FEATURE_RAW)
|
||||
@ -881,7 +881,7 @@ dmu_recv_begin_sync(void *arg, dmu_tx_t *tx)
|
||||
if (drba->drba_cookie->drc_fromsnapobj != 0) {
|
||||
VERIFY0(dsl_dataset_hold_obj(dp,
|
||||
drba->drba_cookie->drc_fromsnapobj, FTAG, &snap));
|
||||
ASSERT3P(dcp, ==, NULL);
|
||||
ASSERT0P(dcp);
|
||||
}
|
||||
if (drc->drc_heal) {
|
||||
/* When healing we want to use the provided snapshot */
|
||||
@ -905,7 +905,7 @@ dmu_recv_begin_sync(void *arg, dmu_tx_t *tx)
|
||||
if (drba->drba_origin != NULL) {
|
||||
VERIFY0(dsl_dataset_hold(dp, drba->drba_origin,
|
||||
FTAG, &origin));
|
||||
ASSERT3P(dcp, ==, NULL);
|
||||
ASSERT0P(dcp);
|
||||
}
|
||||
|
||||
/* Create new dataset. */
|
||||
@ -2792,7 +2792,7 @@ receive_read_payload_and_next_header(dmu_recv_cookie_t *drc, int len, void *buf)
|
||||
drc->drc_rrd->bytes_read = drc->drc_bytes_read;
|
||||
}
|
||||
} else {
|
||||
ASSERT3P(buf, ==, NULL);
|
||||
ASSERT0P(buf);
|
||||
}
|
||||
|
||||
drc->drc_prev_cksum = drc->drc_cksum;
|
||||
@ -3450,7 +3450,7 @@ dmu_recv_stream(dmu_recv_cookie_t *drc, offset_t *voffp)
|
||||
break;
|
||||
}
|
||||
|
||||
ASSERT3P(drc->drc_rrd, ==, NULL);
|
||||
ASSERT0P(drc->drc_rrd);
|
||||
drc->drc_rrd = drc->drc_next_rrd;
|
||||
drc->drc_next_rrd = NULL;
|
||||
/* Allocates and loads header into drc->drc_next_rrd */
|
||||
@ -3468,7 +3468,7 @@ dmu_recv_stream(dmu_recv_cookie_t *drc, offset_t *voffp)
|
||||
drc->drc_rrd = NULL;
|
||||
}
|
||||
|
||||
ASSERT3P(drc->drc_rrd, ==, NULL);
|
||||
ASSERT0P(drc->drc_rrd);
|
||||
drc->drc_rrd = kmem_zalloc(sizeof (*drc->drc_rrd), KM_SLEEP);
|
||||
drc->drc_rrd->eos_marker = B_TRUE;
|
||||
bqueue_enqueue_flush(&rwa->q, drc->drc_rrd, 1);
|
||||
|
||||
@ -1067,7 +1067,7 @@ dmu_redact_snap(const char *snapname, nvlist_t *redactnvl,
|
||||
}
|
||||
if (err != 0)
|
||||
goto out;
|
||||
VERIFY3P(nvlist_next_nvpair(redactnvl, pair), ==, NULL);
|
||||
VERIFY0P(nvlist_next_nvpair(redactnvl, pair));
|
||||
|
||||
boolean_t resuming = B_FALSE;
|
||||
zfs_bookmark_phys_t bookmark;
|
||||
|
||||
@ -962,7 +962,7 @@ do_dump(dmu_send_cookie_t *dscp, struct send_range *range)
|
||||
char *data = NULL;
|
||||
if (srdp->abd != NULL) {
|
||||
data = abd_to_buf(srdp->abd);
|
||||
ASSERT3P(srdp->abuf, ==, NULL);
|
||||
ASSERT0P(srdp->abuf);
|
||||
} else if (srdp->abuf != NULL) {
|
||||
data = srdp->abuf->b_data;
|
||||
}
|
||||
@ -2514,7 +2514,7 @@ dmu_send_impl(struct dmu_send_params *dspp)
|
||||
* list in the stream.
|
||||
*/
|
||||
if (dspp->numfromredactsnaps != NUM_SNAPS_NOT_REDACTED) {
|
||||
ASSERT3P(from_rl, ==, NULL);
|
||||
ASSERT0P(from_rl);
|
||||
fnvlist_add_uint64_array(nvl, BEGINNV_REDACT_FROM_SNAPS,
|
||||
dspp->fromredactsnaps, (uint_t)dspp->numfromredactsnaps);
|
||||
if (dspp->numfromredactsnaps > 0) {
|
||||
@ -2891,7 +2891,7 @@ dmu_send(const char *tosnap, const char *fromsnap, boolean_t embedok,
|
||||
&fromds);
|
||||
|
||||
if (err != 0) {
|
||||
ASSERT3P(fromds, ==, NULL);
|
||||
ASSERT0P(fromds);
|
||||
} else {
|
||||
/*
|
||||
* We need to make a deep copy of the redact
|
||||
|
||||
@ -214,7 +214,7 @@ dnode_dest(void *arg, void *unused)
|
||||
|
||||
for (int i = 0; i < TXG_SIZE; i++) {
|
||||
ASSERT(!multilist_link_active(&dn->dn_dirty_link[i]));
|
||||
ASSERT3P(dn->dn_free_ranges[i], ==, NULL);
|
||||
ASSERT0P(dn->dn_free_ranges[i]);
|
||||
list_destroy(&dn->dn_dirty_records[i]);
|
||||
ASSERT0(dn->dn_next_nblkptr[i]);
|
||||
ASSERT0(dn->dn_next_nlevels[i]);
|
||||
@ -231,10 +231,10 @@ dnode_dest(void *arg, void *unused)
|
||||
ASSERT0(dn->dn_assigned_txg);
|
||||
ASSERT0(dn->dn_dirty_txg);
|
||||
ASSERT0(dn->dn_dirtyctx);
|
||||
ASSERT3P(dn->dn_dirtyctx_firstset, ==, NULL);
|
||||
ASSERT3P(dn->dn_bonus, ==, NULL);
|
||||
ASSERT0P(dn->dn_dirtyctx_firstset);
|
||||
ASSERT0P(dn->dn_bonus);
|
||||
ASSERT(!dn->dn_have_spill);
|
||||
ASSERT3P(dn->dn_zio, ==, NULL);
|
||||
ASSERT0P(dn->dn_zio);
|
||||
ASSERT0(dn->dn_oldused);
|
||||
ASSERT0(dn->dn_oldflags);
|
||||
ASSERT0(dn->dn_olduid);
|
||||
@ -780,7 +780,7 @@ dnode_allocate(dnode_t *dn, dmu_object_type_t ot, int blocksize, int ibs,
|
||||
ASSERT0(dn->dn_next_maxblkid[i]);
|
||||
ASSERT(!multilist_link_active(&dn->dn_dirty_link[i]));
|
||||
ASSERT3P(list_head(&dn->dn_dirty_records[i]), ==, NULL);
|
||||
ASSERT3P(dn->dn_free_ranges[i], ==, NULL);
|
||||
ASSERT0P(dn->dn_free_ranges[i]);
|
||||
}
|
||||
|
||||
dn->dn_type = ot;
|
||||
|
||||
@ -868,7 +868,7 @@ dnode_sync(dnode_t *dn, dmu_tx_t *tx)
|
||||
dbuf_sync_list(list, dn->dn_phys->dn_nlevels - 1, tx);
|
||||
|
||||
if (!DMU_OBJECT_IS_SPECIAL(dn->dn_object)) {
|
||||
ASSERT3P(list_head(list), ==, NULL);
|
||||
ASSERT0P(list_head(list));
|
||||
dnode_rele(dn, (void *)(uintptr_t)tx->tx_txg);
|
||||
}
|
||||
|
||||
|
||||
@ -501,7 +501,7 @@ dsl_bookmark_create_sync_impl_snap(const char *bookmark, const char *snapshot,
|
||||
sizeof (uint64_t) * num_redact_snaps);
|
||||
local_rl->rl_phys->rlp_num_snaps = num_redact_snaps;
|
||||
if (bookmark_redacted) {
|
||||
ASSERT3P(redaction_list, ==, NULL);
|
||||
ASSERT0P(redaction_list);
|
||||
local_rl->rl_phys->rlp_last_blkid = UINT64_MAX;
|
||||
local_rl->rl_phys->rlp_last_object = UINT64_MAX;
|
||||
dsl_redaction_list_long_rele(local_rl, tag);
|
||||
|
||||
@ -1912,7 +1912,7 @@ dsl_dataset_create_crypt_sync(uint64_t dsobj, dsl_dir_t *dd,
|
||||
|
||||
/* clones always use their origin's wrapping key */
|
||||
if (dsl_dir_is_clone(dd)) {
|
||||
ASSERT3P(dcp, ==, NULL);
|
||||
ASSERT0P(dcp);
|
||||
|
||||
/*
|
||||
* If this is an encrypted clone we just need to clone the
|
||||
|
||||
@ -350,7 +350,7 @@ dsl_destroy_snapshot_sync_impl(dsl_dataset_t *ds, boolean_t defer, dmu_tx_t *tx)
|
||||
dsl_dataset_deactivate_feature(ds, f, tx);
|
||||
}
|
||||
if (dsl_dataset_phys(ds)->ds_prev_snap_obj != 0) {
|
||||
ASSERT3P(ds->ds_prev, ==, NULL);
|
||||
ASSERT0P(ds->ds_prev);
|
||||
VERIFY0(dsl_dataset_hold_obj(dp,
|
||||
dsl_dataset_phys(ds)->ds_prev_snap_obj, FTAG, &ds_prev));
|
||||
after_branch_point =
|
||||
@ -465,7 +465,7 @@ dsl_destroy_snapshot_sync_impl(dsl_dataset_t *ds, boolean_t defer, dmu_tx_t *tx)
|
||||
&used, &comp, &uncomp);
|
||||
dsl_dataset_phys(ds_next)->ds_unique_bytes += used;
|
||||
dsl_dataset_rele(ds_nextnext, FTAG);
|
||||
ASSERT3P(ds_next->ds_prev, ==, NULL);
|
||||
ASSERT0P(ds_next->ds_prev);
|
||||
|
||||
/* Collapse range in this head. */
|
||||
dsl_dataset_t *hds;
|
||||
|
||||
@ -5141,7 +5141,7 @@ dsl_scan_io_queue_vdev_xfer(vdev_t *svd, vdev_t *tvd)
|
||||
mutex_enter(&svd->vdev_scan_io_queue_lock);
|
||||
mutex_enter(&tvd->vdev_scan_io_queue_lock);
|
||||
|
||||
VERIFY3P(tvd->vdev_scan_io_queue, ==, NULL);
|
||||
VERIFY0P(tvd->vdev_scan_io_queue);
|
||||
tvd->vdev_scan_io_queue = svd->vdev_scan_io_queue;
|
||||
svd->vdev_scan_io_queue = NULL;
|
||||
if (tvd->vdev_scan_io_queue != NULL)
|
||||
|
||||
@ -5739,7 +5739,7 @@ metaslab_unalloc_dva(spa_t *spa, const dva_t *dva, uint64_t txg)
|
||||
ASSERT(!vd->vdev_removing);
|
||||
ASSERT(vdev_is_concrete(vd));
|
||||
ASSERT0(vd->vdev_indirect_config.vic_mapping_object);
|
||||
ASSERT3P(vd->vdev_indirect_mapping, ==, NULL);
|
||||
ASSERT0P(vd->vdev_indirect_mapping);
|
||||
|
||||
if (DVA_GET_GANG(dva))
|
||||
size = vdev_gang_header_asize(vd);
|
||||
|
||||
@ -377,7 +377,7 @@ zfs_range_tree_add_impl(void *arg, uint64_t start, uint64_t size, uint64_t fill)
|
||||
return;
|
||||
}
|
||||
|
||||
ASSERT3P(rs, ==, NULL);
|
||||
ASSERT0P(rs);
|
||||
|
||||
/*
|
||||
* Determine whether or not we will have to merge with our neighbors.
|
||||
@ -867,7 +867,7 @@ zfs_range_tree_remove_xor_add_segment(uint64_t start, uint64_t end,
|
||||
|
||||
next = zfs_btree_next(&removefrom->rt_root, &where, &where);
|
||||
}
|
||||
VERIFY3P(curr, ==, NULL);
|
||||
VERIFY0P(curr);
|
||||
|
||||
if (start != end) {
|
||||
VERIFY3U(start, <, end);
|
||||
|
||||
@ -2799,7 +2799,7 @@ spa_passivate_log(spa_t *spa)
|
||||
vdev_t *tvd = rvd->vdev_child[c];
|
||||
|
||||
if (tvd->vdev_islog) {
|
||||
ASSERT3P(tvd->vdev_log_mg, ==, NULL);
|
||||
ASSERT0P(tvd->vdev_log_mg);
|
||||
metaslab_group_passivate(tvd->vdev_mg);
|
||||
slog_found = B_TRUE;
|
||||
}
|
||||
@ -2822,7 +2822,7 @@ spa_activate_log(spa_t *spa)
|
||||
vdev_t *tvd = rvd->vdev_child[c];
|
||||
|
||||
if (tvd->vdev_islog) {
|
||||
ASSERT3P(tvd->vdev_log_mg, ==, NULL);
|
||||
ASSERT0P(tvd->vdev_log_mg);
|
||||
metaslab_group_activate(tvd->vdev_mg);
|
||||
}
|
||||
}
|
||||
@ -3259,7 +3259,7 @@ spa_livelist_delete_cb(void *arg, zthr_t *z)
|
||||
static void
|
||||
spa_start_livelist_destroy_thread(spa_t *spa)
|
||||
{
|
||||
ASSERT3P(spa->spa_livelist_delete_zthr, ==, NULL);
|
||||
ASSERT0P(spa->spa_livelist_delete_zthr);
|
||||
spa->spa_livelist_delete_zthr =
|
||||
zthr_create("z_livelist_destroy",
|
||||
spa_livelist_delete_cb_check, spa_livelist_delete_cb, spa,
|
||||
@ -3469,7 +3469,7 @@ spa_start_livelist_condensing_thread(spa_t *spa)
|
||||
spa->spa_to_condense.syncing = B_FALSE;
|
||||
spa->spa_to_condense.cancelled = B_FALSE;
|
||||
|
||||
ASSERT3P(spa->spa_livelist_condense_zthr, ==, NULL);
|
||||
ASSERT0P(spa->spa_livelist_condense_zthr);
|
||||
spa->spa_livelist_condense_zthr =
|
||||
zthr_create("z_livelist_condense",
|
||||
spa_livelist_condense_cb_check,
|
||||
@ -3486,7 +3486,7 @@ spa_spawn_aux_threads(spa_t *spa)
|
||||
spa_start_livelist_destroy_thread(spa);
|
||||
spa_start_livelist_condensing_thread(spa);
|
||||
|
||||
ASSERT3P(spa->spa_checkpoint_discard_zthr, ==, NULL);
|
||||
ASSERT0P(spa->spa_checkpoint_discard_zthr);
|
||||
spa->spa_checkpoint_discard_zthr =
|
||||
zthr_create("z_checkpoint_discard",
|
||||
spa_checkpoint_discard_thread_check,
|
||||
@ -5913,7 +5913,7 @@ spa_load_best(spa_t *spa, spa_load_state_t state, uint64_t max_request,
|
||||
nvlist_free(config);
|
||||
|
||||
if (state == SPA_LOAD_RECOVER) {
|
||||
ASSERT3P(loadinfo, ==, NULL);
|
||||
ASSERT0P(loadinfo);
|
||||
spa_import_progress_remove(spa_guid(spa));
|
||||
return (rewind_error);
|
||||
} else {
|
||||
|
||||
@ -1101,10 +1101,10 @@ vdev_free(vdev_t *vd)
|
||||
{
|
||||
spa_t *spa = vd->vdev_spa;
|
||||
|
||||
ASSERT3P(vd->vdev_initialize_thread, ==, NULL);
|
||||
ASSERT3P(vd->vdev_trim_thread, ==, NULL);
|
||||
ASSERT3P(vd->vdev_autotrim_thread, ==, NULL);
|
||||
ASSERT3P(vd->vdev_rebuild_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_initialize_thread);
|
||||
ASSERT0P(vd->vdev_trim_thread);
|
||||
ASSERT0P(vd->vdev_autotrim_thread);
|
||||
ASSERT0P(vd->vdev_rebuild_thread);
|
||||
|
||||
/*
|
||||
* Scan queues are normally destroyed at the end of a scan. If the
|
||||
@ -1309,9 +1309,9 @@ vdev_top_transfer(vdev_t *svd, vdev_t *tvd)
|
||||
ASSERT0(tvd->vdev_indirect_config.vic_births_object);
|
||||
ASSERT0(tvd->vdev_indirect_config.vic_mapping_object);
|
||||
ASSERT3U(tvd->vdev_indirect_config.vic_prev_indirect_vdev, ==, -1ULL);
|
||||
ASSERT3P(tvd->vdev_indirect_mapping, ==, NULL);
|
||||
ASSERT3P(tvd->vdev_indirect_births, ==, NULL);
|
||||
ASSERT3P(tvd->vdev_obsolete_sm, ==, NULL);
|
||||
ASSERT0P(tvd->vdev_indirect_mapping);
|
||||
ASSERT0P(tvd->vdev_indirect_births);
|
||||
ASSERT0P(tvd->vdev_obsolete_sm);
|
||||
ASSERT0(tvd->vdev_noalloc);
|
||||
ASSERT0(tvd->vdev_removing);
|
||||
ASSERT0(tvd->vdev_rebuilding);
|
||||
@ -3945,7 +3945,7 @@ vdev_load(vdev_t *vd)
|
||||
if (error == 0 && checkpoint_sm_obj != 0) {
|
||||
objset_t *mos = spa_meta_objset(vd->vdev_spa);
|
||||
ASSERT(vd->vdev_asize != 0);
|
||||
ASSERT3P(vd->vdev_checkpoint_sm, ==, NULL);
|
||||
ASSERT0P(vd->vdev_checkpoint_sm);
|
||||
|
||||
error = space_map_open(&vd->vdev_checkpoint_sm,
|
||||
mos, checkpoint_sm_obj, 0, vd->vdev_asize,
|
||||
@ -3993,7 +3993,7 @@ vdev_load(vdev_t *vd)
|
||||
if (error == 0 && obsolete_sm_object != 0) {
|
||||
objset_t *mos = vd->vdev_spa->spa_meta_objset;
|
||||
ASSERT(vd->vdev_asize != 0);
|
||||
ASSERT3P(vd->vdev_obsolete_sm, ==, NULL);
|
||||
ASSERT0P(vd->vdev_obsolete_sm);
|
||||
|
||||
if ((error = space_map_open(&vd->vdev_obsolete_sm, mos,
|
||||
obsolete_sm_object, 0, vd->vdev_asize, 0))) {
|
||||
@ -4521,7 +4521,7 @@ top:
|
||||
/*
|
||||
* Prevent any future allocations.
|
||||
*/
|
||||
ASSERT3P(tvd->vdev_log_mg, ==, NULL);
|
||||
ASSERT0P(tvd->vdev_log_mg);
|
||||
metaslab_group_passivate(mg);
|
||||
(void) spa_vdev_state_exit(spa, vd, 0);
|
||||
|
||||
|
||||
@ -477,7 +477,7 @@ vdev_draid_generate_perms(const draid_map_t *map, uint8_t **permsp)
|
||||
VERIFY3U(map->dm_children, <=, VDEV_DRAID_MAX_CHILDREN);
|
||||
VERIFY3U(map->dm_seed, !=, 0);
|
||||
VERIFY3U(map->dm_nperms, !=, 0);
|
||||
VERIFY3P(map->dm_perms, ==, NULL);
|
||||
VERIFY0P(map->dm_perms);
|
||||
|
||||
#ifdef _KERNEL
|
||||
/*
|
||||
@ -704,7 +704,7 @@ vdev_draid_map_alloc_scrub(zio_t *zio, uint64_t abd_offset, raidz_row_t *rr)
|
||||
uint64_t skip_off = 0;
|
||||
|
||||
ASSERT3U(zio->io_type, ==, ZIO_TYPE_READ);
|
||||
ASSERT3P(rr->rr_abd_empty, ==, NULL);
|
||||
ASSERT0P(rr->rr_abd_empty);
|
||||
|
||||
if (rr->rr_nempty > 0) {
|
||||
rr->rr_abd_empty = abd_alloc_linear(rr->rr_nempty * skip_size,
|
||||
@ -793,7 +793,7 @@ vdev_draid_map_alloc_empty(zio_t *zio, raidz_row_t *rr)
|
||||
uint64_t skip_off = 0;
|
||||
|
||||
ASSERT3U(zio->io_type, ==, ZIO_TYPE_READ);
|
||||
ASSERT3P(rr->rr_abd_empty, ==, NULL);
|
||||
ASSERT0P(rr->rr_abd_empty);
|
||||
|
||||
if (rr->rr_nempty > 0) {
|
||||
rr->rr_abd_empty = abd_alloc_linear(rr->rr_nempty * skip_size,
|
||||
@ -807,7 +807,7 @@ vdev_draid_map_alloc_empty(zio_t *zio, raidz_row_t *rr)
|
||||
/* empty data column (small read), add a skip sector */
|
||||
ASSERT3U(skip_size, ==, parity_size);
|
||||
ASSERT3U(rr->rr_nempty, !=, 0);
|
||||
ASSERT3P(rc->rc_abd, ==, NULL);
|
||||
ASSERT0P(rc->rc_abd);
|
||||
rc->rc_abd = abd_get_offset_size(rr->rr_abd_empty,
|
||||
skip_off, skip_size);
|
||||
skip_off += skip_size;
|
||||
|
||||
@ -792,7 +792,7 @@ spa_condense_indirect_start_sync(vdev_t *vd, dmu_tx_t *tx)
|
||||
DMU_POOL_CONDENSING_INDIRECT, sizeof (uint64_t),
|
||||
sizeof (*scip) / sizeof (uint64_t), scip, tx));
|
||||
|
||||
ASSERT3P(spa->spa_condensing_indirect, ==, NULL);
|
||||
ASSERT0P(spa->spa_condensing_indirect);
|
||||
spa->spa_condensing_indirect = spa_condensing_indirect_create(spa);
|
||||
|
||||
zfs_dbgmsg("starting condense of vdev %llu in txg %llu: "
|
||||
@ -882,7 +882,7 @@ spa_condense_fini(spa_t *spa)
|
||||
void
|
||||
spa_start_indirect_condensing_thread(spa_t *spa)
|
||||
{
|
||||
ASSERT3P(spa->spa_condense_zthr, ==, NULL);
|
||||
ASSERT0P(spa->spa_condense_zthr);
|
||||
spa->spa_condense_zthr = zthr_create("z_indirect_condense",
|
||||
spa_condense_indirect_thread_check,
|
||||
spa_condense_indirect_thread, spa, minclsyspri);
|
||||
@ -1504,7 +1504,7 @@ vdev_indirect_splits_checksum_validate(indirect_vsd_t *iv, zio_t *zio)
|
||||
is != NULL; is = list_next(&iv->iv_splits, is)) {
|
||||
|
||||
ASSERT3P(is->is_good_child->ic_data, !=, NULL);
|
||||
ASSERT3P(is->is_good_child->ic_duplicate, ==, NULL);
|
||||
ASSERT0P(is->is_good_child->ic_duplicate);
|
||||
|
||||
abd_copy_off(zio->io_abd, is->is_good_child->ic_data,
|
||||
is->is_split_offset, 0, is->is_size);
|
||||
|
||||
@ -632,7 +632,7 @@ vdev_initialize(vdev_t *vd)
|
||||
ASSERT(MUTEX_HELD(&vd->vdev_initialize_lock));
|
||||
ASSERT(vd->vdev_ops->vdev_op_leaf);
|
||||
ASSERT(vdev_is_concrete(vd));
|
||||
ASSERT3P(vd->vdev_initialize_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_initialize_thread);
|
||||
ASSERT(!vd->vdev_detached);
|
||||
ASSERT(!vd->vdev_initialize_exit_wanted);
|
||||
ASSERT(!vd->vdev_top->vdev_removing);
|
||||
@ -653,7 +653,7 @@ vdev_uninitialize(vdev_t *vd)
|
||||
ASSERT(MUTEX_HELD(&vd->vdev_initialize_lock));
|
||||
ASSERT(vd->vdev_ops->vdev_op_leaf);
|
||||
ASSERT(vdev_is_concrete(vd));
|
||||
ASSERT3P(vd->vdev_initialize_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_initialize_thread);
|
||||
ASSERT(!vd->vdev_detached);
|
||||
ASSERT(!vd->vdev_initialize_exit_wanted);
|
||||
ASSERT(!vd->vdev_top->vdev_removing);
|
||||
@ -672,7 +672,7 @@ vdev_initialize_stop_wait_impl(vdev_t *vd)
|
||||
while (vd->vdev_initialize_thread != NULL)
|
||||
cv_wait(&vd->vdev_initialize_cv, &vd->vdev_initialize_lock);
|
||||
|
||||
ASSERT3P(vd->vdev_initialize_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_initialize_thread);
|
||||
vd->vdev_initialize_exit_wanted = B_FALSE;
|
||||
}
|
||||
|
||||
|
||||
@ -780,7 +780,7 @@ vdev_queue_aggregate(vdev_queue_t *vq, zio_t *zio)
|
||||
if (dio->io_flags & ZIO_FLAG_NODATA) {
|
||||
/* allocate a buffer for a write gap */
|
||||
ASSERT3U(dio->io_type, ==, ZIO_TYPE_WRITE);
|
||||
ASSERT3P(dio->io_abd, ==, NULL);
|
||||
ASSERT0P(dio->io_abd);
|
||||
abd_gang_add(aio->io_abd,
|
||||
abd_get_zeros(dio->io_size), B_TRUE);
|
||||
} else {
|
||||
|
||||
@ -412,7 +412,7 @@ vdev_raidz_map_free(raidz_map_t *rm)
|
||||
rm->rm_nphys_cols);
|
||||
}
|
||||
|
||||
ASSERT3P(rm->rm_lr, ==, NULL);
|
||||
ASSERT0P(rm->rm_lr);
|
||||
kmem_free(rm, offsetof(raidz_map_t, rm_row[rm->rm_nrows]));
|
||||
}
|
||||
|
||||
@ -2431,7 +2431,7 @@ raidz_start_skip_writes(zio_t *zio)
|
||||
vdev_t *cvd = vd->vdev_child[rc->rc_devidx];
|
||||
if (rc->rc_size != 0)
|
||||
continue;
|
||||
ASSERT3P(rc->rc_abd, ==, NULL);
|
||||
ASSERT0P(rc->rc_abd);
|
||||
|
||||
ASSERT3U(rc->rc_offset, <,
|
||||
cvd->vdev_psize - VDEV_LABEL_END_SIZE);
|
||||
@ -4743,7 +4743,7 @@ spa_raidz_expand_thread(void *arg, zthr_t *zthr)
|
||||
void
|
||||
spa_start_raidz_expansion_thread(spa_t *spa)
|
||||
{
|
||||
ASSERT3P(spa->spa_raidz_expand_zthr, ==, NULL);
|
||||
ASSERT0P(spa->spa_raidz_expand_zthr);
|
||||
spa->spa_raidz_expand_zthr = zthr_create("raidz_expand",
|
||||
spa_raidz_expand_thread_check, spa_raidz_expand_thread,
|
||||
spa, defclsyspri);
|
||||
|
||||
@ -256,7 +256,7 @@ vdev_rebuild_initiate_sync(void *arg, dmu_tx_t *tx)
|
||||
"vdev_id=%llu vdev_guid=%llu started",
|
||||
(u_longlong_t)vd->vdev_id, (u_longlong_t)vd->vdev_guid);
|
||||
|
||||
ASSERT3P(vd->vdev_rebuild_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_rebuild_thread);
|
||||
vd->vdev_rebuild_thread = thread_create(NULL, 0,
|
||||
vdev_rebuild_thread, vd, 0, &p0, TS_RUN, maxclsyspri);
|
||||
|
||||
@ -413,7 +413,7 @@ vdev_rebuild_reset_sync(void *arg, dmu_tx_t *tx)
|
||||
mutex_enter(&vd->vdev_rebuild_lock);
|
||||
|
||||
ASSERT(vrp->vrp_rebuild_state == VDEV_REBUILD_ACTIVE);
|
||||
ASSERT3P(vd->vdev_rebuild_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_rebuild_thread);
|
||||
|
||||
vrp->vrp_last_offset = 0;
|
||||
vrp->vrp_min_txg = 0;
|
||||
|
||||
@ -423,7 +423,7 @@ vdev_remove_initiate_sync(void *arg, dmu_tx_t *tx)
|
||||
svr = spa_vdev_removal_create(vd);
|
||||
|
||||
ASSERT(vd->vdev_removing);
|
||||
ASSERT3P(vd->vdev_indirect_mapping, ==, NULL);
|
||||
ASSERT0P(vd->vdev_indirect_mapping);
|
||||
|
||||
spa_feature_incr(spa, SPA_FEATURE_DEVICE_REMOVAL, tx);
|
||||
if (spa_feature_is_enabled(spa, SPA_FEATURE_OBSOLETE_COUNTS)) {
|
||||
@ -529,7 +529,7 @@ vdev_remove_initiate_sync(void *arg, dmu_tx_t *tx)
|
||||
* but in any case only when there are outstanding free i/os, which
|
||||
* there are not).
|
||||
*/
|
||||
ASSERT3P(spa->spa_vdev_removal, ==, NULL);
|
||||
ASSERT0P(spa->spa_vdev_removal);
|
||||
spa->spa_vdev_removal = svr;
|
||||
svr->svr_thread = thread_create(NULL, 0,
|
||||
spa_vdev_remove_thread, spa, 0, &p0, TS_RUN, minclsyspri);
|
||||
@ -1362,11 +1362,11 @@ vdev_remove_complete(spa_t *spa)
|
||||
txg_wait_synced(spa->spa_dsl_pool, 0);
|
||||
txg = spa_vdev_enter(spa);
|
||||
vdev_t *vd = vdev_lookup_top(spa, spa->spa_vdev_removal->svr_vdev_id);
|
||||
ASSERT3P(vd->vdev_initialize_thread, ==, NULL);
|
||||
ASSERT3P(vd->vdev_trim_thread, ==, NULL);
|
||||
ASSERT3P(vd->vdev_autotrim_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_initialize_thread);
|
||||
ASSERT0P(vd->vdev_trim_thread);
|
||||
ASSERT0P(vd->vdev_autotrim_thread);
|
||||
vdev_rebuild_stop_wait(vd);
|
||||
ASSERT3P(vd->vdev_rebuild_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_rebuild_thread);
|
||||
|
||||
sysevent_t *ev = spa_event_create(spa, vd, NULL,
|
||||
ESC_ZFS_VDEV_REMOVE_DEV);
|
||||
@ -1868,7 +1868,7 @@ spa_vdev_remove_cancel_sync(void *arg, dmu_tx_t *tx)
|
||||
vdev_indirect_mapping_t *vim = vd->vdev_indirect_mapping;
|
||||
objset_t *mos = spa->spa_meta_objset;
|
||||
|
||||
ASSERT3P(svr->svr_thread, ==, NULL);
|
||||
ASSERT0P(svr->svr_thread);
|
||||
|
||||
spa_feature_decr(spa, SPA_FEATURE_DEVICE_REMOVAL, tx);
|
||||
|
||||
@ -2076,7 +2076,7 @@ spa_vdev_remove_log(vdev_t *vd, uint64_t *txg)
|
||||
|
||||
ASSERT(vd->vdev_islog);
|
||||
ASSERT(vd == vd->vdev_top);
|
||||
ASSERT3P(vd->vdev_log_mg, ==, NULL);
|
||||
ASSERT0P(vd->vdev_log_mg);
|
||||
ASSERT(MUTEX_HELD(&spa_namespace_lock));
|
||||
|
||||
/*
|
||||
@ -2112,7 +2112,7 @@ spa_vdev_remove_log(vdev_t *vd, uint64_t *txg)
|
||||
|
||||
if (error != 0) {
|
||||
metaslab_group_activate(mg);
|
||||
ASSERT3P(vd->vdev_log_mg, ==, NULL);
|
||||
ASSERT0P(vd->vdev_log_mg);
|
||||
return (error);
|
||||
}
|
||||
ASSERT0(vd->vdev_stat.vs_alloc);
|
||||
|
||||
@ -1010,7 +1010,7 @@ vdev_trim(vdev_t *vd, uint64_t rate, boolean_t partial, boolean_t secure)
|
||||
ASSERT(MUTEX_HELD(&vd->vdev_trim_lock));
|
||||
ASSERT(vd->vdev_ops->vdev_op_leaf);
|
||||
ASSERT(vdev_is_concrete(vd));
|
||||
ASSERT3P(vd->vdev_trim_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_trim_thread);
|
||||
ASSERT(!vd->vdev_detached);
|
||||
ASSERT(!vd->vdev_trim_exit_wanted);
|
||||
ASSERT(!vd->vdev_top->vdev_removing);
|
||||
@ -1032,7 +1032,7 @@ vdev_trim_stop_wait_impl(vdev_t *vd)
|
||||
while (vd->vdev_trim_thread != NULL)
|
||||
cv_wait(&vd->vdev_trim_cv, &vd->vdev_trim_lock);
|
||||
|
||||
ASSERT3P(vd->vdev_trim_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_trim_thread);
|
||||
vd->vdev_trim_exit_wanted = B_FALSE;
|
||||
}
|
||||
|
||||
@ -1539,7 +1539,7 @@ vdev_autotrim_stop_wait(vdev_t *tvd)
|
||||
cv_wait(&tvd->vdev_autotrim_cv,
|
||||
&tvd->vdev_autotrim_lock);
|
||||
|
||||
ASSERT3P(tvd->vdev_autotrim_thread, ==, NULL);
|
||||
ASSERT0P(tvd->vdev_autotrim_thread);
|
||||
tvd->vdev_autotrim_exit_wanted = B_FALSE;
|
||||
}
|
||||
mutex_exit(&tvd->vdev_autotrim_lock);
|
||||
@ -1712,7 +1712,7 @@ vdev_trim_l2arc(spa_t *spa)
|
||||
mutex_enter(&vd->vdev_trim_lock);
|
||||
ASSERT(vd->vdev_ops->vdev_op_leaf);
|
||||
ASSERT(vdev_is_concrete(vd));
|
||||
ASSERT3P(vd->vdev_trim_thread, ==, NULL);
|
||||
ASSERT0P(vd->vdev_trim_thread);
|
||||
ASSERT(!vd->vdev_detached);
|
||||
ASSERT(!vd->vdev_trim_exit_wanted);
|
||||
ASSERT(!vd->vdev_top->vdev_removing);
|
||||
|
||||
@ -765,7 +765,7 @@ zcp_lua_alloc(void *ud, void *ptr, size_t osize, size_t nsize)
|
||||
return (NULL);
|
||||
}
|
||||
(void) memcpy(luabuf, ptr, osize);
|
||||
VERIFY3P(zcp_lua_alloc(ud, ptr, osize, 0), ==, NULL);
|
||||
VERIFY0P(zcp_lua_alloc(ud, ptr, osize, 0));
|
||||
return (luabuf);
|
||||
}
|
||||
}
|
||||
|
||||
@ -7337,8 +7337,8 @@ zfs_ioctl_register_legacy(zfs_ioc_t ioc, zfs_ioc_legacy_func_t *func,
|
||||
|
||||
ASSERT3U(ioc, >=, ZFS_IOC_FIRST);
|
||||
ASSERT3U(ioc, <, ZFS_IOC_LAST);
|
||||
ASSERT3P(vec->zvec_legacy_func, ==, NULL);
|
||||
ASSERT3P(vec->zvec_func, ==, NULL);
|
||||
ASSERT0P(vec->zvec_legacy_func);
|
||||
ASSERT0P(vec->zvec_func);
|
||||
|
||||
vec->zvec_legacy_func = func;
|
||||
vec->zvec_secpolicy = secpolicy;
|
||||
@ -7361,8 +7361,8 @@ zfs_ioctl_register(const char *name, zfs_ioc_t ioc, zfs_ioc_func_t *func,
|
||||
|
||||
ASSERT3U(ioc, >=, ZFS_IOC_FIRST);
|
||||
ASSERT3U(ioc, <, ZFS_IOC_LAST);
|
||||
ASSERT3P(vec->zvec_legacy_func, ==, NULL);
|
||||
ASSERT3P(vec->zvec_func, ==, NULL);
|
||||
ASSERT0P(vec->zvec_legacy_func);
|
||||
ASSERT0P(vec->zvec_func);
|
||||
|
||||
/* if we are logging, the name must be valid */
|
||||
ASSERT(!allow_log || namecheck != NO_NAME);
|
||||
|
||||
@ -864,9 +864,9 @@ zil_free_lwb(zilog_t *zilog, lwb_t *lwb)
|
||||
ASSERT(MUTEX_HELD(&zilog->zl_lock));
|
||||
ASSERT(lwb->lwb_state == LWB_STATE_NEW ||
|
||||
lwb->lwb_state == LWB_STATE_FLUSH_DONE);
|
||||
ASSERT3P(lwb->lwb_child_zio, ==, NULL);
|
||||
ASSERT3P(lwb->lwb_write_zio, ==, NULL);
|
||||
ASSERT3P(lwb->lwb_root_zio, ==, NULL);
|
||||
ASSERT0P(lwb->lwb_child_zio);
|
||||
ASSERT0P(lwb->lwb_write_zio);
|
||||
ASSERT0P(lwb->lwb_root_zio);
|
||||
ASSERT3U(lwb->lwb_alloc_txg, <=, spa_syncing_txg(zilog->zl_spa));
|
||||
ASSERT3U(lwb->lwb_max_txg, <=, spa_syncing_txg(zilog->zl_spa));
|
||||
VERIFY(list_is_empty(&lwb->lwb_itxs));
|
||||
@ -1351,7 +1351,7 @@ zil_commit_waiter_link_lwb(zil_commit_waiter_t *zcw, lwb_t *lwb)
|
||||
|
||||
ASSERT(!list_link_active(&zcw->zcw_node));
|
||||
list_insert_tail(&lwb->lwb_waiters, zcw);
|
||||
ASSERT3P(zcw->zcw_lwb, ==, NULL);
|
||||
ASSERT0P(zcw->zcw_lwb);
|
||||
zcw->zcw_lwb = lwb;
|
||||
}
|
||||
|
||||
@ -1365,7 +1365,7 @@ zil_commit_waiter_link_nolwb(zil_commit_waiter_t *zcw, list_t *nolwb)
|
||||
{
|
||||
ASSERT(!list_link_active(&zcw->zcw_node));
|
||||
list_insert_tail(nolwb, zcw);
|
||||
ASSERT3P(zcw->zcw_lwb, ==, NULL);
|
||||
ASSERT0P(zcw->zcw_lwb);
|
||||
}
|
||||
|
||||
void
|
||||
@ -3489,7 +3489,7 @@ static void
|
||||
zil_free_commit_waiter(zil_commit_waiter_t *zcw)
|
||||
{
|
||||
ASSERT(!list_link_active(&zcw->zcw_node));
|
||||
ASSERT3P(zcw->zcw_lwb, ==, NULL);
|
||||
ASSERT0P(zcw->zcw_lwb);
|
||||
ASSERT3B(zcw->zcw_done, ==, B_TRUE);
|
||||
mutex_destroy(&zcw->zcw_lock);
|
||||
cv_destroy(&zcw->zcw_cv);
|
||||
@ -3670,9 +3670,9 @@ zil_commit(zilog_t *zilog, uint64_t foid)
|
||||
* verifying that truth before we return to the caller.
|
||||
*/
|
||||
ASSERT(list_is_empty(&zilog->zl_lwb_list));
|
||||
ASSERT3P(zilog->zl_last_lwb_opened, ==, NULL);
|
||||
ASSERT0P(zilog->zl_last_lwb_opened);
|
||||
for (int i = 0; i < TXG_SIZE; i++)
|
||||
ASSERT3P(zilog->zl_itxg[i].itxg_itxs, ==, NULL);
|
||||
ASSERT0P(zilog->zl_itxg[i].itxg_itxs);
|
||||
return;
|
||||
}
|
||||
|
||||
@ -4005,8 +4005,8 @@ zil_open(objset_t *os, zil_get_data_t *get_data, zil_sums_t *zil_sums)
|
||||
{
|
||||
zilog_t *zilog = dmu_objset_zil(os);
|
||||
|
||||
ASSERT3P(zilog->zl_get_data, ==, NULL);
|
||||
ASSERT3P(zilog->zl_last_lwb_opened, ==, NULL);
|
||||
ASSERT0P(zilog->zl_get_data);
|
||||
ASSERT0P(zilog->zl_last_lwb_opened);
|
||||
ASSERT(list_is_empty(&zilog->zl_lwb_list));
|
||||
|
||||
zilog->zl_get_data = get_data;
|
||||
|
||||
@ -339,8 +339,8 @@ zio_fini(void)
|
||||
}
|
||||
|
||||
for (size_t i = 0; i < n; i++) {
|
||||
VERIFY3P(zio_buf_cache[i], ==, NULL);
|
||||
VERIFY3P(zio_data_buf_cache[i], ==, NULL);
|
||||
VERIFY0P(zio_buf_cache[i]);
|
||||
VERIFY0P(zio_data_buf_cache[i]);
|
||||
}
|
||||
|
||||
if (zio_ksp != NULL) {
|
||||
@ -1451,7 +1451,7 @@ zio_free(spa_t *spa, uint64_t txg, const blkptr_t *bp)
|
||||
metaslab_check_free(spa, bp);
|
||||
bplist_append(&spa->spa_free_bplist[txg & TXG_MASK], bp);
|
||||
} else {
|
||||
VERIFY3P(zio_free_sync(NULL, spa, txg, bp, 0), ==, NULL);
|
||||
VERIFY0P(zio_free_sync(NULL, spa, txg, bp, 0));
|
||||
}
|
||||
}
|
||||
|
||||
@ -1747,7 +1747,7 @@ zio_flush(zio_t *pio, vdev_t *vd)
|
||||
void
|
||||
zio_shrink(zio_t *zio, uint64_t size)
|
||||
{
|
||||
ASSERT3P(zio->io_executor, ==, NULL);
|
||||
ASSERT0P(zio->io_executor);
|
||||
ASSERT3U(zio->io_orig_size, ==, zio->io_size);
|
||||
ASSERT3U(size, <=, zio->io_size);
|
||||
|
||||
@ -2509,7 +2509,7 @@ zio_wait(zio_t *zio)
|
||||
int error;
|
||||
|
||||
ASSERT3S(zio->io_stage, ==, ZIO_STAGE_OPEN);
|
||||
ASSERT3P(zio->io_executor, ==, NULL);
|
||||
ASSERT0P(zio->io_executor);
|
||||
|
||||
zio->io_waiter = curthread;
|
||||
ASSERT0(zio->io_queued_timestamp);
|
||||
@ -2551,7 +2551,7 @@ zio_nowait(zio_t *zio)
|
||||
if (zio == NULL)
|
||||
return;
|
||||
|
||||
ASSERT3P(zio->io_executor, ==, NULL);
|
||||
ASSERT0P(zio->io_executor);
|
||||
|
||||
if (zio->io_child_type == ZIO_CHILD_LOGICAL &&
|
||||
list_is_empty(&zio->io_parent_list)) {
|
||||
|
||||
@ -1119,7 +1119,7 @@ zio_clear_fault(int id)
|
||||
kmem_free(handler->zi_lanes, sizeof (*handler->zi_lanes) *
|
||||
handler->zi_record.zi_nlanes);
|
||||
} else {
|
||||
ASSERT3P(handler->zi_lanes, ==, NULL);
|
||||
ASSERT0P(handler->zi_lanes);
|
||||
}
|
||||
|
||||
if (handler->zi_spa_name != NULL)
|
||||
|
||||
@ -129,7 +129,7 @@ zrl_tryenter(zrlock_t *zrl)
|
||||
(uint32_t *)&zrl->zr_refcount, 0, ZRL_LOCKED);
|
||||
if (cas == 0) {
|
||||
#ifdef ZFS_DEBUG
|
||||
ASSERT3P(zrl->zr_owner, ==, NULL);
|
||||
ASSERT0P(zrl->zr_owner);
|
||||
zrl->zr_owner = curthread;
|
||||
#endif
|
||||
return (1);
|
||||
|
||||
@ -316,7 +316,7 @@ zthr_destroy(zthr_t *t)
|
||||
{
|
||||
ASSERT(!MUTEX_HELD(&t->zthr_state_lock));
|
||||
ASSERT(!MUTEX_HELD(&t->zthr_request_lock));
|
||||
VERIFY3P(t->zthr_thread, ==, NULL);
|
||||
VERIFY0P(t->zthr_thread);
|
||||
mutex_destroy(&t->zthr_request_lock);
|
||||
mutex_destroy(&t->zthr_state_lock);
|
||||
cv_destroy(&t->zthr_cv);
|
||||
|
||||
Loading…
Reference in New Issue
Block a user