Change KM_PUSHPAGE -> KM_SLEEP

By marking DMU transaction processing contexts with PF_FSTRANS
we can revert the KM_PUSHPAGE -> KM_SLEEP changes.  This brings
us back in line with upstream.  In some cases this means simply
swapping the flags back.  For others fnvlist_alloc() was replaced
by nvlist_alloc(..., KM_PUSHPAGE) and must be reverted back to
fnvlist_alloc() which assumes KM_SLEEP.

The one place KM_PUSHPAGE is kept is when allocating ARC buffers
which allows us to dip in to reserved memory.  This is again the
same as upstream.

Signed-off-by: Brian Behlendorf <behlendorf1@llnl.gov>
This commit is contained in:
Brian Behlendorf 2014-11-20 19:09:39 -05:00
parent efcd79a883
commit 79c76d5b65
65 changed files with 269 additions and 281 deletions

View File

@ -344,7 +344,7 @@ _NOTE(CONSTCOND) } while (0)
#define dprintf_dbuf_bp(db, bp, fmt, ...) do { \ #define dprintf_dbuf_bp(db, bp, fmt, ...) do { \
if (zfs_flags & ZFS_DEBUG_DPRINTF) { \ if (zfs_flags & ZFS_DEBUG_DPRINTF) { \
char *__blkbuf = kmem_alloc(BP_SPRINTF_LEN, KM_PUSHPAGE); \ char *__blkbuf = kmem_alloc(BP_SPRINTF_LEN, KM_SLEEP); \
snprintf_blkptr(__blkbuf, BP_SPRINTF_LEN, bp); \ snprintf_blkptr(__blkbuf, BP_SPRINTF_LEN, bp); \
dprintf_dbuf(db, fmt " %s\n", __VA_ARGS__, __blkbuf); \ dprintf_dbuf(db, fmt " %s\n", __VA_ARGS__, __blkbuf); \
kmem_free(__blkbuf, BP_SPRINTF_LEN); \ kmem_free(__blkbuf, BP_SPRINTF_LEN); \

View File

@ -285,7 +285,7 @@ int dsl_dataset_rollback(const char *fsname, void *owner, nvlist_t *result);
#ifdef ZFS_DEBUG #ifdef ZFS_DEBUG
#define dprintf_ds(ds, fmt, ...) do { \ #define dprintf_ds(ds, fmt, ...) do { \
if (zfs_flags & ZFS_DEBUG_DPRINTF) { \ if (zfs_flags & ZFS_DEBUG_DPRINTF) { \
char *__ds_name = kmem_alloc(MAXNAMELEN, KM_PUSHPAGE); \ char *__ds_name = kmem_alloc(MAXNAMELEN, KM_SLEEP); \
dsl_dataset_name(ds, __ds_name); \ dsl_dataset_name(ds, __ds_name); \
dprintf("ds=%s " fmt, __ds_name, __VA_ARGS__); \ dprintf("ds=%s " fmt, __ds_name, __VA_ARGS__); \
kmem_free(__ds_name, MAXNAMELEN); \ kmem_free(__ds_name, MAXNAMELEN); \

View File

@ -150,7 +150,7 @@ void dsl_dir_set_reservation_sync_impl(dsl_dir_t *dd, uint64_t value,
#define dprintf_dd(dd, fmt, ...) do { \ #define dprintf_dd(dd, fmt, ...) do { \
if (zfs_flags & ZFS_DEBUG_DPRINTF) { \ if (zfs_flags & ZFS_DEBUG_DPRINTF) { \
char *__ds_name = kmem_alloc(MAXNAMELEN + strlen(MOS_DIR_NAME) + 1, \ char *__ds_name = kmem_alloc(MAXNAMELEN + strlen(MOS_DIR_NAME) + 1, \
KM_PUSHPAGE); \ KM_SLEEP); \
dsl_dir_name(dd, __ds_name); \ dsl_dir_name(dd, __ds_name); \
dprintf("dd=%s " fmt, __ds_name, __VA_ARGS__); \ dprintf("dd=%s " fmt, __ds_name, __VA_ARGS__); \
kmem_free(__ds_name, MAXNAMELEN + strlen(MOS_DIR_NAME) + 1); \ kmem_free(__ds_name, MAXNAMELEN + strlen(MOS_DIR_NAME) + 1); \

View File

@ -879,7 +879,7 @@ extern void spa_event_notify(spa_t *spa, vdev_t *vdev, const char *name);
#ifdef ZFS_DEBUG #ifdef ZFS_DEBUG
#define dprintf_bp(bp, fmt, ...) do { \ #define dprintf_bp(bp, fmt, ...) do { \
if (zfs_flags & ZFS_DEBUG_DPRINTF) { \ if (zfs_flags & ZFS_DEBUG_DPRINTF) { \
char *__blkbuf = kmem_alloc(BP_SPRINTF_LEN, KM_PUSHPAGE); \ char *__blkbuf = kmem_alloc(BP_SPRINTF_LEN, KM_SLEEP); \
snprintf_blkptr(__blkbuf, BP_SPRINTF_LEN, (bp)); \ snprintf_blkptr(__blkbuf, BP_SPRINTF_LEN, (bp)); \
dprintf(fmt " %s\n", __VA_ARGS__, __blkbuf); \ dprintf(fmt " %s\n", __VA_ARGS__, __blkbuf); \
kmem_free(__blkbuf, BP_SPRINTF_LEN); \ kmem_free(__blkbuf, BP_SPRINTF_LEN); \

View File

@ -447,7 +447,6 @@ typedef struct taskq_ent {
#define TQ_SLEEP KM_SLEEP /* Can block for memory */ #define TQ_SLEEP KM_SLEEP /* Can block for memory */
#define TQ_NOSLEEP KM_NOSLEEP /* cannot block for memory; may fail */ #define TQ_NOSLEEP KM_NOSLEEP /* cannot block for memory; may fail */
#define TQ_PUSHPAGE KM_PUSHPAGE /* Cannot perform I/O */
#define TQ_NOQUEUE 0x02 /* Do not enqueue if can't dispatch */ #define TQ_NOQUEUE 0x02 /* Do not enqueue if can't dispatch */
#define TQ_FRONT 0x08 /* Queue in front */ #define TQ_FRONT 0x08 /* Queue in front */

View File

@ -175,7 +175,7 @@ zprop_iter_common(zprop_func func, void *cb, boolean_t show_all,
size = num_props * sizeof (zprop_desc_t *); size = num_props * sizeof (zprop_desc_t *);
#if defined(_KERNEL) #if defined(_KERNEL)
order = kmem_alloc(size, KM_PUSHPAGE); order = kmem_alloc(size, KM_SLEEP);
#else #else
if ((order = malloc(size)) == NULL) if ((order = malloc(size)) == NULL)
return (ZPROP_CONT); return (ZPROP_CONT);

View File

@ -978,7 +978,7 @@ arc_cksum_compute(arc_buf_t *buf, boolean_t force)
return; return;
} }
buf->b_hdr->b_freeze_cksum = kmem_alloc(sizeof (zio_cksum_t), buf->b_hdr->b_freeze_cksum = kmem_alloc(sizeof (zio_cksum_t),
KM_PUSHPAGE); KM_SLEEP);
fletcher_2_native(buf->b_data, buf->b_hdr->b_size, fletcher_2_native(buf->b_data, buf->b_hdr->b_size,
buf->b_hdr->b_freeze_cksum); buf->b_hdr->b_freeze_cksum);
mutex_exit(&buf->b_hdr->b_freeze_lock); mutex_exit(&buf->b_hdr->b_freeze_lock);
@ -1477,7 +1477,7 @@ arc_buf_data_free(arc_buf_t *buf, void (*free_func)(void *, size_t))
if (HDR_L2_WRITING(hdr)) { if (HDR_L2_WRITING(hdr)) {
l2arc_data_free_t *df; l2arc_data_free_t *df;
df = kmem_alloc(sizeof (l2arc_data_free_t), KM_PUSHPAGE); df = kmem_alloc(sizeof (l2arc_data_free_t), KM_SLEEP);
df->l2df_data = buf->b_data; df->l2df_data = buf->b_data;
df->l2df_size = hdr->b_size; df->l2df_size = hdr->b_size;
df->l2df_func = free_func; df->l2df_func = free_func;
@ -3142,7 +3142,7 @@ top:
arc_callback_t *acb = NULL; arc_callback_t *acb = NULL;
acb = kmem_zalloc(sizeof (arc_callback_t), acb = kmem_zalloc(sizeof (arc_callback_t),
KM_PUSHPAGE); KM_SLEEP);
acb->acb_done = done; acb->acb_done = done;
acb->acb_private = private; acb->acb_private = private;
if (pio != NULL) if (pio != NULL)
@ -3284,7 +3284,7 @@ top:
ASSERT(!GHOST_STATE(hdr->b_state)); ASSERT(!GHOST_STATE(hdr->b_state));
acb = kmem_zalloc(sizeof (arc_callback_t), KM_PUSHPAGE); acb = kmem_zalloc(sizeof (arc_callback_t), KM_SLEEP);
acb->acb_done = done; acb->acb_done = done;
acb->acb_private = private; acb->acb_private = private;
@ -3341,7 +3341,7 @@ top:
atomic_inc_32(&hdr->b_l2hdr->b_hits); atomic_inc_32(&hdr->b_l2hdr->b_hits);
cb = kmem_zalloc(sizeof (l2arc_read_callback_t), cb = kmem_zalloc(sizeof (l2arc_read_callback_t),
KM_PUSHPAGE); KM_SLEEP);
cb->l2rcb_buf = buf; cb->l2rcb_buf = buf;
cb->l2rcb_spa = spa; cb->l2rcb_spa = spa;
cb->l2rcb_bp = *bp; cb->l2rcb_bp = *bp;
@ -3871,7 +3871,7 @@ arc_write(zio_t *pio, spa_t *spa, uint64_t txg,
hdr->b_flags |= ARC_L2CACHE; hdr->b_flags |= ARC_L2CACHE;
if (l2arc_compress) if (l2arc_compress)
hdr->b_flags |= ARC_L2COMPRESS; hdr->b_flags |= ARC_L2COMPRESS;
callback = kmem_zalloc(sizeof (arc_write_callback_t), KM_PUSHPAGE); callback = kmem_zalloc(sizeof (arc_write_callback_t), KM_SLEEP);
callback->awcb_ready = ready; callback->awcb_ready = ready;
callback->awcb_physdone = physdone; callback->awcb_physdone = physdone;
callback->awcb_done = done; callback->awcb_done = done;
@ -4979,7 +4979,7 @@ l2arc_write_buffers(spa_t *spa, l2arc_dev_t *dev, uint64_t target_sz,
list_insert_head(dev->l2ad_buflist, head); list_insert_head(dev->l2ad_buflist, head);
cb = kmem_alloc(sizeof (l2arc_write_callback_t), cb = kmem_alloc(sizeof (l2arc_write_callback_t),
KM_PUSHPAGE); KM_SLEEP);
cb->l2wcb_dev = dev; cb->l2wcb_dev = dev;
cb->l2wcb_head = head; cb->l2wcb_head = head;
pio = zio_root(spa, l2arc_write_done, cb, pio = zio_root(spa, l2arc_write_done, cb,
@ -4989,7 +4989,7 @@ l2arc_write_buffers(spa_t *spa, l2arc_dev_t *dev, uint64_t target_sz,
/* /*
* Create and add a new L2ARC header. * Create and add a new L2ARC header.
*/ */
l2hdr = kmem_cache_alloc(l2arc_hdr_cache, KM_PUSHPAGE); l2hdr = kmem_cache_alloc(l2arc_hdr_cache, KM_SLEEP);
l2hdr->b_dev = dev; l2hdr->b_dev = dev;
l2hdr->b_daddr = 0; l2hdr->b_daddr = 0;
arc_space_consume(L2HDR_SIZE, ARC_SPACE_L2HDRS); arc_space_consume(L2HDR_SIZE, ARC_SPACE_L2HDRS);

View File

@ -45,7 +45,7 @@ bplist_destroy(bplist_t *bpl)
void void
bplist_append(bplist_t *bpl, const blkptr_t *bp) bplist_append(bplist_t *bpl, const blkptr_t *bp)
{ {
bplist_entry_t *bpe = kmem_alloc(sizeof (*bpe), KM_PUSHPAGE); bplist_entry_t *bpe = kmem_alloc(sizeof (*bpe), KM_SLEEP);
mutex_enter(&bpl->bpl_lock); mutex_enter(&bpl->bpl_lock);
bpe->bpe_blk = *bp; bpe->bpe_blk = *bp;

View File

@ -134,7 +134,7 @@ bptree_add(objset_t *os, uint64_t obj, blkptr_t *bp, uint64_t birth_txg,
VERIFY3U(0, ==, dmu_bonus_hold(os, obj, FTAG, &db)); VERIFY3U(0, ==, dmu_bonus_hold(os, obj, FTAG, &db));
bt = db->db_data; bt = db->db_data;
bte = kmem_zalloc(sizeof (*bte), KM_PUSHPAGE); bte = kmem_zalloc(sizeof (*bte), KM_SLEEP);
bte->be_birth_txg = birth_txg; bte->be_birth_txg = birth_txg;
bte->be_bp = *bp; bte->be_bp = *bp;
dmu_write(os, obj, bt->bt_end * sizeof (*bte), sizeof (*bte), bte, tx); dmu_write(os, obj, bt->bt_end * sizeof (*bte), sizeof (*bte), bte, tx);

View File

@ -315,7 +315,7 @@ retry:
* Large allocations which do not require contiguous pages * Large allocations which do not require contiguous pages
* should be using vmem_alloc() in the linux kernel * should be using vmem_alloc() in the linux kernel
*/ */
h->hash_table = vmem_zalloc(hsize * sizeof (void *), KM_PUSHPAGE); h->hash_table = vmem_zalloc(hsize * sizeof (void *), KM_SLEEP);
#else #else
h->hash_table = kmem_zalloc(hsize * sizeof (void *), KM_NOSLEEP); h->hash_table = kmem_zalloc(hsize * sizeof (void *), KM_NOSLEEP);
#endif #endif
@ -1121,7 +1121,7 @@ dbuf_dirty(dmu_buf_impl_t *db, dmu_tx_t *tx)
dn->dn_dirtyctx = dn->dn_dirtyctx =
(dmu_tx_is_syncing(tx) ? DN_DIRTY_SYNC : DN_DIRTY_OPEN); (dmu_tx_is_syncing(tx) ? DN_DIRTY_SYNC : DN_DIRTY_OPEN);
ASSERT(dn->dn_dirtyctx_firstset == NULL); ASSERT(dn->dn_dirtyctx_firstset == NULL);
dn->dn_dirtyctx_firstset = kmem_alloc(1, KM_PUSHPAGE); dn->dn_dirtyctx_firstset = kmem_alloc(1, KM_SLEEP);
} }
mutex_exit(&dn->dn_mtx); mutex_exit(&dn->dn_mtx);
@ -1198,7 +1198,7 @@ dbuf_dirty(dmu_buf_impl_t *db, dmu_tx_t *tx)
* to make a copy of it so that the changes we make in this * to make a copy of it so that the changes we make in this
* transaction group won't leak out when we sync the older txg. * transaction group won't leak out when we sync the older txg.
*/ */
dr = kmem_zalloc(sizeof (dbuf_dirty_record_t), KM_PUSHPAGE); dr = kmem_zalloc(sizeof (dbuf_dirty_record_t), KM_SLEEP);
list_link_init(&dr->dr_dirty_node); list_link_init(&dr->dr_dirty_node);
if (db->db_level == 0) { if (db->db_level == 0) {
void *data_old = db->db_buf; void *data_old = db->db_buf;
@ -1764,7 +1764,7 @@ dbuf_create(dnode_t *dn, uint8_t level, uint64_t blkid,
ASSERT(RW_LOCK_HELD(&dn->dn_struct_rwlock)); ASSERT(RW_LOCK_HELD(&dn->dn_struct_rwlock));
ASSERT(dn->dn_type != DMU_OT_NONE); ASSERT(dn->dn_type != DMU_OT_NONE);
db = kmem_cache_alloc(dbuf_cache, KM_PUSHPAGE); db = kmem_cache_alloc(dbuf_cache, KM_SLEEP);
db->db_objset = os; db->db_objset = os;
db->db.db_object = dn->dn_object; db->db.db_object = dn->dn_object;
@ -2059,7 +2059,7 @@ dbuf_hold_impl(dnode_t *dn, uint8_t level, uint64_t blkid, int fail_sparse,
int error; int error;
dh = kmem_zalloc(sizeof (struct dbuf_hold_impl_data) * dh = kmem_zalloc(sizeof (struct dbuf_hold_impl_data) *
DBUF_HOLD_IMPL_MAX_DEPTH, KM_PUSHPAGE); DBUF_HOLD_IMPL_MAX_DEPTH, KM_SLEEP);
__dbuf_hold_impl_init(dh, dn, level, blkid, fail_sparse, tag, dbp, 0); __dbuf_hold_impl_init(dh, dn, level, blkid, fail_sparse, tag, dbp, 0);
error = __dbuf_hold_impl(dh); error = __dbuf_hold_impl(dh);

View File

@ -517,7 +517,7 @@ ddt_get_dedup_stats(spa_t *spa, ddt_stat_t *dds_total)
{ {
ddt_histogram_t *ddh_total; ddt_histogram_t *ddh_total;
ddh_total = kmem_zalloc(sizeof (ddt_histogram_t), KM_PUSHPAGE); ddh_total = kmem_zalloc(sizeof (ddt_histogram_t), KM_SLEEP);
ddt_get_dedup_histogram(spa, ddh_total); ddt_get_dedup_histogram(spa, ddh_total);
ddt_histogram_stat(dds_total, ddh_total); ddt_histogram_stat(dds_total, ddh_total);
kmem_free(ddh_total, sizeof (ddt_histogram_t)); kmem_free(ddh_total, sizeof (ddt_histogram_t));
@ -685,7 +685,7 @@ ddt_alloc(const ddt_key_t *ddk)
{ {
ddt_entry_t *dde; ddt_entry_t *dde;
dde = kmem_cache_alloc(ddt_entry_cache, KM_PUSHPAGE); dde = kmem_cache_alloc(ddt_entry_cache, KM_SLEEP);
bzero(dde, sizeof (ddt_entry_t)); bzero(dde, sizeof (ddt_entry_t));
cv_init(&dde->dde_cv, NULL, CV_DEFAULT, NULL); cv_init(&dde->dde_cv, NULL, CV_DEFAULT, NULL);
@ -834,7 +834,7 @@ ddt_table_alloc(spa_t *spa, enum zio_checksum c)
{ {
ddt_t *ddt; ddt_t *ddt;
ddt = kmem_cache_alloc(ddt_cache, KM_PUSHPAGE); ddt = kmem_cache_alloc(ddt_cache, KM_SLEEP);
bzero(ddt, sizeof (ddt_t)); bzero(ddt, sizeof (ddt_t));
mutex_init(&ddt->ddt_lock, NULL, MUTEX_DEFAULT, NULL); mutex_init(&ddt->ddt_lock, NULL, MUTEX_DEFAULT, NULL);
@ -937,7 +937,7 @@ ddt_class_contains(spa_t *spa, enum ddt_class max_class, const blkptr_t *bp)
return (B_TRUE); return (B_TRUE);
ddt = spa->spa_ddt[BP_GET_CHECKSUM(bp)]; ddt = spa->spa_ddt[BP_GET_CHECKSUM(bp)];
dde = kmem_cache_alloc(ddt_entry_cache, KM_PUSHPAGE); dde = kmem_cache_alloc(ddt_entry_cache, KM_SLEEP);
ddt_key_fill(&(dde->dde_key), bp); ddt_key_fill(&(dde->dde_key), bp);

View File

@ -62,7 +62,7 @@ ddt_zap_lookup(objset_t *os, uint64_t object, ddt_entry_t *dde)
uint64_t one, csize; uint64_t one, csize;
int error; int error;
cbuf = kmem_alloc(sizeof (dde->dde_phys) + 1, KM_PUSHPAGE); cbuf = kmem_alloc(sizeof (dde->dde_phys) + 1, KM_SLEEP);
error = zap_length_uint64(os, object, (uint64_t *)&dde->dde_key, error = zap_length_uint64(os, object, (uint64_t *)&dde->dde_key,
DDT_KEY_WORDS, &one, &csize); DDT_KEY_WORDS, &one, &csize);

View File

@ -417,7 +417,7 @@ dmu_buf_hold_array_by_dnode(dnode_t *dn, uint64_t offset, uint64_t length,
} }
nblks = 1; nblks = 1;
} }
dbp = kmem_zalloc(sizeof (dmu_buf_t *) * nblks, KM_PUSHPAGE); dbp = kmem_zalloc(sizeof (dmu_buf_t *) * nblks, KM_SLEEP);
zio = zio_root(dn->dn_objset->os_spa, NULL, NULL, ZIO_FLAG_CANFAIL); zio = zio_root(dn->dn_objset->os_spa, NULL, NULL, ZIO_FLAG_CANFAIL);
blkid = dbuf_whichblock(dn, offset); blkid = dbuf_whichblock(dn, offset);
@ -924,11 +924,11 @@ dmu_xuio_init(xuio_t *xuio, int nblk)
uio_t *uio = &xuio->xu_uio; uio_t *uio = &xuio->xu_uio;
uio->uio_iovcnt = nblk; uio->uio_iovcnt = nblk;
uio->uio_iov = kmem_zalloc(nblk * sizeof (iovec_t), KM_PUSHPAGE); uio->uio_iov = kmem_zalloc(nblk * sizeof (iovec_t), KM_SLEEP);
priv = kmem_zalloc(sizeof (dmu_xuio_t), KM_PUSHPAGE); priv = kmem_zalloc(sizeof (dmu_xuio_t), KM_SLEEP);
priv->cnt = nblk; priv->cnt = nblk;
priv->bufs = kmem_zalloc(nblk * sizeof (arc_buf_t *), KM_PUSHPAGE); priv->bufs = kmem_zalloc(nblk * sizeof (arc_buf_t *), KM_SLEEP);
priv->iovp = uio->uio_iov; priv->iovp = uio->uio_iov;
XUIO_XUZC_PRIV(xuio) = priv; XUIO_XUZC_PRIV(xuio) = priv;
@ -1530,7 +1530,7 @@ dmu_sync_late_arrival(zio_t *pio, objset_t *os, dmu_sync_cb_t *done, zgd_t *zgd,
return (SET_ERROR(EIO)); return (SET_ERROR(EIO));
} }
dsa = kmem_alloc(sizeof (dmu_sync_arg_t), KM_PUSHPAGE); dsa = kmem_alloc(sizeof (dmu_sync_arg_t), KM_SLEEP);
dsa->dsa_dr = NULL; dsa->dsa_dr = NULL;
dsa->dsa_done = done; dsa->dsa_done = done;
dsa->dsa_zgd = zgd; dsa->dsa_zgd = zgd;
@ -1670,7 +1670,7 @@ dmu_sync(zio_t *pio, uint64_t txg, dmu_sync_cb_t *done, zgd_t *zgd)
dr->dt.dl.dr_override_state = DR_IN_DMU_SYNC; dr->dt.dl.dr_override_state = DR_IN_DMU_SYNC;
mutex_exit(&db->db_mtx); mutex_exit(&db->db_mtx);
dsa = kmem_alloc(sizeof (dmu_sync_arg_t), KM_PUSHPAGE); dsa = kmem_alloc(sizeof (dmu_sync_arg_t), KM_SLEEP);
dsa->dsa_dr = dr; dsa->dsa_dr = dr;
dsa->dsa_done = done; dsa->dsa_done = done;
dsa->dsa_zgd = zgd; dsa->dsa_zgd = zgd;

View File

@ -279,7 +279,7 @@ dmu_objset_open_impl(spa_t *spa, dsl_dataset_t *ds, blkptr_t *bp,
ASSERT(ds == NULL || MUTEX_HELD(&ds->ds_opening_lock)); ASSERT(ds == NULL || MUTEX_HELD(&ds->ds_opening_lock));
os = kmem_zalloc(sizeof (objset_t), KM_PUSHPAGE); os = kmem_zalloc(sizeof (objset_t), KM_SLEEP);
os->os_dsl_dataset = ds; os->os_dsl_dataset = ds;
os->os_spa = spa; os->os_spa = spa;
os->os_rootbp = bp; os->os_rootbp = bp;
@ -1573,7 +1573,7 @@ dmu_objset_find_dp(dsl_pool_t *dp, uint64_t ddobj,
} }
thisobj = dd->dd_phys->dd_head_dataset_obj; thisobj = dd->dd_phys->dd_head_dataset_obj;
attr = kmem_alloc(sizeof (zap_attribute_t), KM_PUSHPAGE); attr = kmem_alloc(sizeof (zap_attribute_t), KM_SLEEP);
/* /*
* Iterate over all children. * Iterate over all children.
@ -1685,7 +1685,7 @@ dmu_objset_find_impl(spa_t *spa, const char *name,
} }
thisobj = dd->dd_phys->dd_head_dataset_obj; thisobj = dd->dd_phys->dd_head_dataset_obj;
attr = kmem_alloc(sizeof (zap_attribute_t), KM_PUSHPAGE); attr = kmem_alloc(sizeof (zap_attribute_t), KM_SLEEP);
/* /*
* Iterate over all children. * Iterate over all children.

View File

@ -288,7 +288,7 @@ traverse_visitbp(traverse_data_t *td, const dnode_phys_t *dnp,
if (err != 0) if (err != 0)
goto post; goto post;
czb = kmem_alloc(sizeof (zbookmark_phys_t), KM_PUSHPAGE); czb = kmem_alloc(sizeof (zbookmark_phys_t), KM_SLEEP);
for (i = 0; i < epb; i++) { for (i = 0; i < epb; i++) {
SET_BOOKMARK(czb, zb->zb_objset, zb->zb_object, SET_BOOKMARK(czb, zb->zb_objset, zb->zb_object,
@ -517,9 +517,9 @@ traverse_impl(spa_t *spa, dsl_dataset_t *ds, uint64_t objset, blkptr_t *rootbp,
*/ */
ASSERT(resume == NULL || !(flags & TRAVERSE_PREFETCH_DATA)); ASSERT(resume == NULL || !(flags & TRAVERSE_PREFETCH_DATA));
td = kmem_alloc(sizeof (traverse_data_t), KM_PUSHPAGE); td = kmem_alloc(sizeof (traverse_data_t), KM_SLEEP);
pd = kmem_zalloc(sizeof (prefetch_data_t), KM_PUSHPAGE); pd = kmem_zalloc(sizeof (prefetch_data_t), KM_SLEEP);
czb = kmem_alloc(sizeof (zbookmark_phys_t), KM_PUSHPAGE); czb = kmem_alloc(sizeof (zbookmark_phys_t), KM_SLEEP);
td->td_spa = spa; td->td_spa = spa;
td->td_objset = objset; td->td_objset = objset;

View File

@ -62,7 +62,7 @@ static kstat_t *dmu_tx_ksp;
dmu_tx_t * dmu_tx_t *
dmu_tx_create_dd(dsl_dir_t *dd) dmu_tx_create_dd(dsl_dir_t *dd)
{ {
dmu_tx_t *tx = kmem_zalloc(sizeof (dmu_tx_t), KM_PUSHPAGE); dmu_tx_t *tx = kmem_zalloc(sizeof (dmu_tx_t), KM_SLEEP);
tx->tx_dir = dd; tx->tx_dir = dd;
if (dd != NULL) if (dd != NULL)
tx->tx_pool = dd->dd_pool; tx->tx_pool = dd->dd_pool;
@ -141,7 +141,7 @@ dmu_tx_hold_object_impl(dmu_tx_t *tx, objset_t *os, uint64_t object,
} }
} }
txh = kmem_zalloc(sizeof (dmu_tx_hold_t), KM_PUSHPAGE); txh = kmem_zalloc(sizeof (dmu_tx_hold_t), KM_SLEEP);
txh->txh_tx = tx; txh->txh_tx = tx;
txh->txh_dnode = dn; txh->txh_dnode = dn;
#ifdef DEBUG_DMU_TX #ifdef DEBUG_DMU_TX
@ -1467,7 +1467,7 @@ dmu_tx_callback_register(dmu_tx_t *tx, dmu_tx_callback_func_t *func, void *data)
{ {
dmu_tx_callback_t *dcb; dmu_tx_callback_t *dcb;
dcb = kmem_alloc(sizeof (dmu_tx_callback_t), KM_PUSHPAGE); dcb = kmem_alloc(sizeof (dmu_tx_callback_t), KM_SLEEP);
dcb->dcb_func = func; dcb->dcb_func = func;
dcb->dcb_data = data; dcb->dcb_data = data;

View File

@ -706,7 +706,7 @@ dmu_zfetch(zfetch_t *zf, uint64_t offset, uint64_t size, int prefetched)
return; return;
} }
newstream = newstream =
kmem_zalloc(sizeof (zstream_t), KM_PUSHPAGE); kmem_zalloc(sizeof (zstream_t), KM_SLEEP);
} }
newstream->zst_offset = zst.zst_offset; newstream->zst_offset = zst.zst_offset;

View File

@ -366,7 +366,7 @@ static dnode_t *
dnode_create(objset_t *os, dnode_phys_t *dnp, dmu_buf_impl_t *db, dnode_create(objset_t *os, dnode_phys_t *dnp, dmu_buf_impl_t *db,
uint64_t object, dnode_handle_t *dnh) uint64_t object, dnode_handle_t *dnh)
{ {
dnode_t *dn = kmem_cache_alloc(dnode_cache, KM_PUSHPAGE); dnode_t *dn = kmem_cache_alloc(dnode_cache, KM_SLEEP);
ASSERT(!POINTER_IS_VALID(dn->dn_objset)); ASSERT(!POINTER_IS_VALID(dn->dn_objset));
dn->dn_moved = 0; dn->dn_moved = 0;
@ -1070,7 +1070,7 @@ dnode_hold_impl(objset_t *os, uint64_t object, int flag,
int i; int i;
dnode_children_t *winner; dnode_children_t *winner;
children_dnodes = kmem_alloc(sizeof (dnode_children_t) + children_dnodes = kmem_alloc(sizeof (dnode_children_t) +
(epb - 1) * sizeof (dnode_handle_t), KM_PUSHPAGE); (epb - 1) * sizeof (dnode_handle_t), KM_SLEEP);
children_dnodes->dnc_count = epb; children_dnodes->dnc_count = epb;
dnh = &children_dnodes->dnc_children[0]; dnh = &children_dnodes->dnc_children[0];
for (i = 0; i < epb; i++) { for (i = 0; i < epb; i++) {

View File

@ -365,7 +365,7 @@ dsl_dataset_hold_obj(dsl_pool_t *dp, uint64_t dsobj, void *tag,
if (ds == NULL) { if (ds == NULL) {
dsl_dataset_t *winner = NULL; dsl_dataset_t *winner = NULL;
ds = kmem_zalloc(sizeof (dsl_dataset_t), KM_PUSHPAGE); ds = kmem_zalloc(sizeof (dsl_dataset_t), KM_SLEEP);
ds->ds_dbuf = dbuf; ds->ds_dbuf = dbuf;
ds->ds_object = dsobj; ds->ds_object = dsobj;
ds->ds_phys = dbuf->db_data; ds->ds_phys = dbuf->db_data;
@ -1772,9 +1772,9 @@ dsl_dataset_rollback_check(void *arg, dmu_tx_t *tx)
} }
/* must not have any bookmarks after the most recent snapshot */ /* must not have any bookmarks after the most recent snapshot */
VERIFY0(nvlist_alloc(&proprequest, NV_UNIQUE_NAME, KM_PUSHPAGE)); proprequest = fnvlist_alloc();
fnvlist_add_boolean(proprequest, zfs_prop_to_name(ZFS_PROP_CREATETXG)); fnvlist_add_boolean(proprequest, zfs_prop_to_name(ZFS_PROP_CREATETXG));
VERIFY0(nvlist_alloc(&bookmarks, NV_UNIQUE_NAME, KM_PUSHPAGE)); bookmarks = fnvlist_alloc();
error = dsl_get_bookmarks_impl(ds, proprequest, bookmarks); error = dsl_get_bookmarks_impl(ds, proprequest, bookmarks);
fnvlist_free(proprequest); fnvlist_free(proprequest);
if (error != 0) if (error != 0)
@ -2262,7 +2262,7 @@ snaplist_make(dsl_pool_t *dp,
if (first_obj == 0) if (first_obj == 0)
first_obj = ds->ds_dir->dd_phys->dd_origin_obj; first_obj = ds->ds_dir->dd_phys->dd_origin_obj;
snap = kmem_alloc(sizeof (*snap), KM_PUSHPAGE); snap = kmem_alloc(sizeof (*snap), KM_SLEEP);
snap->ds = ds; snap->ds = ds;
list_insert_tail(l, snap); list_insert_tail(l, snap);
obj = ds->ds_phys->ds_prev_snap_obj; obj = ds->ds_phys->ds_prev_snap_obj;

View File

@ -82,7 +82,7 @@ dsl_deadlist_load_tree(dsl_deadlist_t *dl)
zap_cursor_advance(&zc)) { zap_cursor_advance(&zc)) {
dsl_deadlist_entry_t *dle; dsl_deadlist_entry_t *dle;
dle = kmem_alloc(sizeof (*dle), KM_PUSHPAGE); dle = kmem_alloc(sizeof (*dle), KM_SLEEP);
dle->dle_mintxg = strtonum(za.za_name, NULL); dle->dle_mintxg = strtonum(za.za_name, NULL);
VERIFY3U(0, ==, bpobj_open(&dle->dle_bpobj, dl->dl_os, VERIFY3U(0, ==, bpobj_open(&dle->dle_bpobj, dl->dl_os,
za.za_first_integer)); za.za_first_integer));
@ -254,7 +254,7 @@ dsl_deadlist_add_key(dsl_deadlist_t *dl, uint64_t mintxg, dmu_tx_t *tx)
dsl_deadlist_load_tree(dl); dsl_deadlist_load_tree(dl);
dle = kmem_alloc(sizeof (*dle), KM_PUSHPAGE); dle = kmem_alloc(sizeof (*dle), KM_SLEEP);
dle->dle_mintxg = mintxg; dle->dle_mintxg = mintxg;
obj = bpobj_alloc_empty(dl->dl_os, SPA_MAXBLOCKSIZE, tx); obj = bpobj_alloc_empty(dl->dl_os, SPA_MAXBLOCKSIZE, tx);
VERIFY3U(0, ==, bpobj_open(&dle->dle_bpobj, dl->dl_os, obj)); VERIFY3U(0, ==, bpobj_open(&dle->dle_bpobj, dl->dl_os, obj));

View File

@ -214,8 +214,8 @@ dsl_dataset_remove_clones_key(dsl_dataset_t *ds, uint64_t mintxg, dmu_tx_t *tx)
if (ds->ds_dir->dd_phys->dd_clones == 0) if (ds->ds_dir->dd_phys->dd_clones == 0)
return; return;
zc = kmem_alloc(sizeof (zap_cursor_t), KM_PUSHPAGE); zc = kmem_alloc(sizeof (zap_cursor_t), KM_SLEEP);
za = kmem_alloc(sizeof (zap_attribute_t), KM_PUSHPAGE); za = kmem_alloc(sizeof (zap_attribute_t), KM_SLEEP);
for (zap_cursor_init(zc, mos, ds->ds_dir->dd_phys->dd_clones); for (zap_cursor_init(zc, mos, ds->ds_dir->dd_phys->dd_clones);
zap_cursor_retrieve(zc, za) == 0; zap_cursor_retrieve(zc, za) == 0;
@ -504,7 +504,7 @@ dsl_destroy_snapshots_nvl(nvlist_t *snaps, boolean_t defer,
dsda.dsda_snaps = snaps; dsda.dsda_snaps = snaps;
VERIFY0(nvlist_alloc(&dsda.dsda_successful_snaps, VERIFY0(nvlist_alloc(&dsda.dsda_successful_snaps,
NV_UNIQUE_NAME, KM_PUSHPAGE)); NV_UNIQUE_NAME, KM_SLEEP));
dsda.dsda_defer = defer; dsda.dsda_defer = defer;
dsda.dsda_errlist = errlist; dsda.dsda_errlist = errlist;
@ -520,11 +520,8 @@ int
dsl_destroy_snapshot(const char *name, boolean_t defer) dsl_destroy_snapshot(const char *name, boolean_t defer)
{ {
int error; int error;
nvlist_t *nvl; nvlist_t *nvl = fnvlist_alloc();
nvlist_t *errlist; nvlist_t *errlist = fnvlist_alloc();
VERIFY0(nvlist_alloc(&nvl, NV_UNIQUE_NAME, KM_PUSHPAGE));
VERIFY0(nvlist_alloc(&errlist, NV_UNIQUE_NAME, KM_PUSHPAGE));
fnvlist_add_boolean(nvl, name); fnvlist_add_boolean(nvl, name);
error = dsl_destroy_snapshots_nvl(nvl, defer, errlist); error = dsl_destroy_snapshots_nvl(nvl, defer, errlist);

View File

@ -97,7 +97,7 @@ dsl_dir_hold_obj(dsl_pool_t *dp, uint64_t ddobj,
if (dd == NULL) { if (dd == NULL) {
dsl_dir_t *winner; dsl_dir_t *winner;
dd = kmem_zalloc(sizeof (dsl_dir_t), KM_PUSHPAGE); dd = kmem_zalloc(sizeof (dsl_dir_t), KM_SLEEP);
dd->dd_object = ddobj; dd->dd_object = ddobj;
dd->dd_dbuf = dbuf; dd->dd_dbuf = dbuf;
dd->dd_pool = dp; dd->dd_pool = dp;
@ -313,7 +313,7 @@ dsl_dir_hold(dsl_pool_t *dp, const char *name, void *tag,
dsl_dir_t *dd; dsl_dir_t *dd;
uint64_t ddobj; uint64_t ddobj;
buf = kmem_alloc(MAXNAMELEN, KM_PUSHPAGE); buf = kmem_alloc(MAXNAMELEN, KM_SLEEP);
err = getcomponent(name, buf, &next); err = getcomponent(name, buf, &next);
if (err != 0) if (err != 0)
goto error; goto error;
@ -696,7 +696,7 @@ dsl_dir_tempreserve_impl(dsl_dir_t *dd, uint64_t asize, boolean_t netfree,
asize - ref_rsrv); asize - ref_rsrv);
mutex_exit(&dd->dd_lock); mutex_exit(&dd->dd_lock);
tr = kmem_zalloc(sizeof (struct tempreserve), KM_PUSHPAGE); tr = kmem_zalloc(sizeof (struct tempreserve), KM_SLEEP);
tr->tr_ds = dd; tr->tr_ds = dd;
tr->tr_size = asize; tr->tr_size = asize;
list_insert_tail(tr_list, tr); list_insert_tail(tr_list, tr);
@ -730,7 +730,7 @@ dsl_dir_tempreserve_space(dsl_dir_t *dd, uint64_t lsize, uint64_t asize,
return (0); return (0);
} }
tr_list = kmem_alloc(sizeof (list_t), KM_PUSHPAGE); tr_list = kmem_alloc(sizeof (list_t), KM_SLEEP);
list_create(tr_list, sizeof (struct tempreserve), list_create(tr_list, sizeof (struct tempreserve),
offsetof(struct tempreserve, tr_node)); offsetof(struct tempreserve, tr_node));
ASSERT3S(asize, >, 0); ASSERT3S(asize, >, 0);
@ -740,7 +740,7 @@ dsl_dir_tempreserve_space(dsl_dir_t *dd, uint64_t lsize, uint64_t asize,
if (err == 0) { if (err == 0) {
struct tempreserve *tr; struct tempreserve *tr;
tr = kmem_zalloc(sizeof (struct tempreserve), KM_PUSHPAGE); tr = kmem_zalloc(sizeof (struct tempreserve), KM_SLEEP);
tr->tr_size = lsize; tr->tr_size = lsize;
list_insert_tail(tr_list, tr); list_insert_tail(tr_list, tr);
} else { } else {

View File

@ -323,7 +323,7 @@ dsl_pool_close(dsl_pool_t *dp)
mutex_destroy(&dp->dp_lock); mutex_destroy(&dp->dp_lock);
taskq_destroy(dp->dp_iput_taskq); taskq_destroy(dp->dp_iput_taskq);
if (dp->dp_blkstats) if (dp->dp_blkstats)
kmem_free(dp->dp_blkstats, sizeof (zfs_all_blkstats_t)); vmem_free(dp->dp_blkstats, sizeof (zfs_all_blkstats_t));
kmem_free(dp, sizeof (dsl_pool_t)); kmem_free(dp, sizeof (dsl_pool_t));
} }

View File

@ -241,9 +241,9 @@ dsl_prop_register(dsl_dataset_t *ds, const char *propname,
if (err != 0) if (err != 0)
return (err); return (err);
cbr = kmem_alloc(sizeof (dsl_prop_cb_record_t), KM_PUSHPAGE); cbr = kmem_alloc(sizeof (dsl_prop_cb_record_t), KM_SLEEP);
cbr->cbr_ds = ds; cbr->cbr_ds = ds;
cbr->cbr_propname = kmem_alloc(strlen(propname)+1, KM_PUSHPAGE); cbr->cbr_propname = kmem_alloc(strlen(propname)+1, KM_SLEEP);
(void) strcpy((char *)cbr->cbr_propname, propname); (void) strcpy((char *)cbr->cbr_propname, propname);
cbr->cbr_func = callback; cbr->cbr_func = callback;
cbr->cbr_arg = cbarg; cbr->cbr_arg = cbarg;
@ -513,7 +513,7 @@ dsl_prop_changed_notify(dsl_pool_t *dp, uint64_t ddobj,
} }
mutex_exit(&dd->dd_lock); mutex_exit(&dd->dd_lock);
za = kmem_alloc(sizeof (zap_attribute_t), KM_PUSHPAGE); za = kmem_alloc(sizeof (zap_attribute_t), KM_SLEEP);
for (zap_cursor_init(&zc, mos, for (zap_cursor_init(&zc, mos,
dd->dd_phys->dd_child_dir_zapobj); dd->dd_phys->dd_child_dir_zapobj);
zap_cursor_retrieve(&zc, za) == 0; zap_cursor_retrieve(&zc, za) == 0;
@ -668,7 +668,7 @@ dsl_prop_set_sync_impl(dsl_dataset_t *ds, const char *propname,
if (source == ZPROP_SRC_LOCAL) { if (source == ZPROP_SRC_LOCAL) {
valstr = value; valstr = value;
} else { } else {
tbuf = kmem_alloc(ZAP_MAXVALUELEN, KM_PUSHPAGE); tbuf = kmem_alloc(ZAP_MAXVALUELEN, KM_SLEEP);
if (dsl_prop_get_ds(ds, propname, 1, if (dsl_prop_get_ds(ds, propname, 1,
ZAP_MAXVALUELEN, tbuf, NULL) == 0) ZAP_MAXVALUELEN, tbuf, NULL) == 0)
valstr = tbuf; valstr = tbuf;

View File

@ -261,8 +261,8 @@ dsl_scan_setup_sync(void *arg, dmu_tx_t *tx)
/* back to the generic stuff */ /* back to the generic stuff */
if (dp->dp_blkstats == NULL) { if (dp->dp_blkstats == NULL) {
dp->dp_blkstats = kmem_alloc(sizeof (zfs_all_blkstats_t), dp->dp_blkstats =
KM_PUSHPAGE); vmem_alloc(sizeof (zfs_all_blkstats_t), KM_SLEEP);
} }
bzero(dp->dp_blkstats, sizeof (zfs_all_blkstats_t)); bzero(dp->dp_blkstats, sizeof (zfs_all_blkstats_t));
@ -762,7 +762,7 @@ dsl_scan_visitbp(blkptr_t *bp, const zbookmark_phys_t *zb,
dsl_pool_t *dp = scn->scn_dp; dsl_pool_t *dp = scn->scn_dp;
blkptr_t *bp_toread; blkptr_t *bp_toread;
bp_toread = kmem_alloc(sizeof (blkptr_t), KM_PUSHPAGE); bp_toread = kmem_alloc(sizeof (blkptr_t), KM_SLEEP);
*bp_toread = *bp; *bp_toread = *bp;
/* ASSERT(pbuf == NULL || arc_released(pbuf)); */ /* ASSERT(pbuf == NULL || arc_released(pbuf)); */
@ -1059,7 +1059,7 @@ dsl_scan_visitds(dsl_scan_t *scn, uint64_t dsobj, dmu_tx_t *tx)
dmu_buf_will_dirty(ds->ds_dbuf, tx); dmu_buf_will_dirty(ds->ds_dbuf, tx);
dsl_scan_visit_rootbp(scn, ds, &ds->ds_phys->ds_bp, tx); dsl_scan_visit_rootbp(scn, ds, &ds->ds_phys->ds_bp, tx);
dsname = kmem_alloc(ZFS_MAXNAMELEN, KM_PUSHPAGE); dsname = kmem_alloc(ZFS_MAXNAMELEN, KM_SLEEP);
dsl_dataset_name(ds, dsname); dsl_dataset_name(ds, dsname);
zfs_dbgmsg("scanned dataset %llu (%s) with min=%llu max=%llu; " zfs_dbgmsg("scanned dataset %llu (%s) with min=%llu max=%llu; "
"pausing=%u", "pausing=%u",
@ -1325,8 +1325,8 @@ dsl_scan_visit(dsl_scan_t *scn, dmu_tx_t *tx)
* bookmark so we don't think that we're still trying to resume. * bookmark so we don't think that we're still trying to resume.
*/ */
bzero(&scn->scn_phys.scn_bookmark, sizeof (zbookmark_phys_t)); bzero(&scn->scn_phys.scn_bookmark, sizeof (zbookmark_phys_t));
zc = kmem_alloc(sizeof (zap_cursor_t), KM_PUSHPAGE); zc = kmem_alloc(sizeof (zap_cursor_t), KM_SLEEP);
za = kmem_alloc(sizeof (zap_attribute_t), KM_PUSHPAGE); za = kmem_alloc(sizeof (zap_attribute_t), KM_SLEEP);
/* keep pulling things out of the zap-object-as-queue */ /* keep pulling things out of the zap-object-as-queue */
while (zap_cursor_init(zc, dp->dp_meta_objset, while (zap_cursor_init(zc, dp->dp_meta_objset,

View File

@ -166,8 +166,7 @@ dsl_dataset_user_hold_sync_one_impl(nvlist_t *tmpholds, dsl_dataset_t *ds,
(u_longlong_t)ds->ds_object); (u_longlong_t)ds->ds_object);
if (nvlist_lookup_nvlist(tmpholds, name, &tags) != 0) { if (nvlist_lookup_nvlist(tmpholds, name, &tags) != 0) {
VERIFY0(nvlist_alloc(&tags, NV_UNIQUE_NAME, tags = fnvlist_alloc();
KM_PUSHPAGE));
fnvlist_add_boolean(tags, htag); fnvlist_add_boolean(tags, htag);
fnvlist_add_nvlist(tmpholds, name, tags); fnvlist_add_nvlist(tmpholds, name, tags);
fnvlist_free(tags); fnvlist_free(tags);
@ -226,7 +225,7 @@ dsl_onexit_hold_cleanup(spa_t *spa, nvlist_t *holds, minor_t minor)
} }
ASSERT(spa != NULL); ASSERT(spa != NULL);
ca = kmem_alloc(sizeof (*ca), KM_PUSHPAGE); ca = kmem_alloc(sizeof (*ca), KM_SLEEP);
(void) strlcpy(ca->zhca_spaname, spa_name(spa), (void) strlcpy(ca->zhca_spaname, spa_name(spa),
sizeof (ca->zhca_spaname)); sizeof (ca->zhca_spaname));
@ -243,7 +242,7 @@ dsl_dataset_user_hold_sync_one(dsl_dataset_t *ds, const char *htag,
nvlist_t *tmpholds; nvlist_t *tmpholds;
if (minor != 0) if (minor != 0)
VERIFY0(nvlist_alloc(&tmpholds, NV_UNIQUE_NAME, KM_PUSHPAGE)); tmpholds = fnvlist_alloc();
else else
tmpholds = NULL; tmpholds = NULL;
dsl_dataset_user_hold_sync_one_impl(tmpholds, ds, htag, minor, now, tx); dsl_dataset_user_hold_sync_one_impl(tmpholds, ds, htag, minor, now, tx);
@ -260,7 +259,7 @@ dsl_dataset_user_hold_sync(void *arg, dmu_tx_t *tx)
uint64_t now = gethrestime_sec(); uint64_t now = gethrestime_sec();
if (dduha->dduha_minor != 0) if (dduha->dduha_minor != 0)
VERIFY0(nvlist_alloc(&tmpholds, NV_UNIQUE_NAME, KM_PUSHPAGE)); tmpholds = fnvlist_alloc();
else else
tmpholds = NULL; tmpholds = NULL;
for (pair = nvlist_next_nvpair(dduha->dduha_chkholds, NULL); for (pair = nvlist_next_nvpair(dduha->dduha_chkholds, NULL);
@ -315,8 +314,7 @@ dsl_dataset_user_hold(nvlist_t *holds, minor_t cleanup_minor, nvlist_t *errlist)
return (0); return (0);
dduha.dduha_holds = holds; dduha.dduha_holds = holds;
VERIFY0(nvlist_alloc(&dduha.dduha_chkholds, NV_UNIQUE_NAME, dduha.dduha_chkholds = fnvlist_alloc();
KM_PUSHPAGE));
dduha.dduha_errlist = errlist; dduha.dduha_errlist = errlist;
dduha.dduha_minor = cleanup_minor; dduha.dduha_minor = cleanup_minor;
@ -365,7 +363,7 @@ dsl_dataset_user_release_check_one(dsl_dataset_user_release_arg_t *ddura,
numholds = 0; numholds = 0;
mos = ds->ds_dir->dd_pool->dp_meta_objset; mos = ds->ds_dir->dd_pool->dp_meta_objset;
zapobj = ds->ds_phys->ds_userrefs_obj; zapobj = ds->ds_phys->ds_userrefs_obj;
VERIFY0(nvlist_alloc(&holds_found, NV_UNIQUE_NAME, KM_PUSHPAGE)); VERIFY0(nvlist_alloc(&holds_found, NV_UNIQUE_NAME, KM_SLEEP));
for (pair = nvlist_next_nvpair(holds, NULL); pair != NULL; for (pair = nvlist_next_nvpair(holds, NULL); pair != NULL;
pair = nvlist_next_nvpair(holds, pair)) { pair = nvlist_next_nvpair(holds, pair)) {
@ -605,9 +603,9 @@ dsl_dataset_user_release_impl(nvlist_t *holds, nvlist_t *errlist,
ddura.ddura_holds = holds; ddura.ddura_holds = holds;
ddura.ddura_errlist = errlist; ddura.ddura_errlist = errlist;
VERIFY0(nvlist_alloc(&ddura.ddura_todelete, NV_UNIQUE_NAME, VERIFY0(nvlist_alloc(&ddura.ddura_todelete, NV_UNIQUE_NAME,
KM_PUSHPAGE)); KM_SLEEP));
VERIFY0(nvlist_alloc(&ddura.ddura_chkholds, NV_UNIQUE_NAME, VERIFY0(nvlist_alloc(&ddura.ddura_chkholds, NV_UNIQUE_NAME,
KM_PUSHPAGE)); KM_SLEEP));
error = dsl_sync_task(pool, dsl_dataset_user_release_check, error = dsl_sync_task(pool, dsl_dataset_user_release_check,
dsl_dataset_user_release_sync, &ddura, 0); dsl_dataset_user_release_sync, &ddura, 0);
@ -657,7 +655,7 @@ dsl_dataset_get_holds(const char *dsname, nvlist_t *nvl)
zap_attribute_t *za; zap_attribute_t *za;
zap_cursor_t zc; zap_cursor_t zc;
za = kmem_alloc(sizeof (zap_attribute_t), KM_PUSHPAGE); za = kmem_alloc(sizeof (zap_attribute_t), KM_SLEEP);
for (zap_cursor_init(&zc, ds->ds_dir->dd_pool->dp_meta_objset, for (zap_cursor_init(&zc, ds->ds_dir->dd_pool->dp_meta_objset,
ds->ds_phys->ds_userrefs_obj); ds->ds_phys->ds_userrefs_obj);
zap_cursor_retrieve(&zc, za) == 0; zap_cursor_retrieve(&zc, za) == 0;

View File

@ -426,7 +426,7 @@ zfs_zevent_alloc(void)
{ {
zevent_t *ev; zevent_t *ev;
ev = kmem_zalloc(sizeof (zevent_t), KM_PUSHPAGE); ev = kmem_zalloc(sizeof (zevent_t), KM_SLEEP);
if (ev == NULL) if (ev == NULL)
return (NULL); return (NULL);
@ -776,7 +776,7 @@ zfs_zevent_destroy(zfs_zevent_t *ze)
static void * static void *
i_fm_alloc(nv_alloc_t *nva, size_t size) i_fm_alloc(nv_alloc_t *nva, size_t size)
{ {
return (kmem_zalloc(size, KM_PUSHPAGE)); return (kmem_zalloc(size, KM_SLEEP));
} }
/* ARGSUSED */ /* ARGSUSED */
@ -844,7 +844,7 @@ fm_nvlist_create(nv_alloc_t *nva)
nv_alloc_t *nvhdl; nv_alloc_t *nvhdl;
if (nva == NULL) { if (nva == NULL) {
nvhdl = kmem_zalloc(sizeof (nv_alloc_t), KM_PUSHPAGE); nvhdl = kmem_zalloc(sizeof (nv_alloc_t), KM_SLEEP);
if (nv_alloc_init(nvhdl, &fm_mem_alloc_ops, NULL, 0) != 0) { if (nv_alloc_init(nvhdl, &fm_mem_alloc_ops, NULL, 0) != 0) {
kmem_free(nvhdl, sizeof (nv_alloc_t)); kmem_free(nvhdl, sizeof (nv_alloc_t));

View File

@ -838,7 +838,7 @@ real_LZ4_compress(const char *source, char *dest, int isize, int osize)
int result; int result;
ASSERT(lz4_cache != NULL); ASSERT(lz4_cache != NULL);
ctx = kmem_cache_alloc(lz4_cache, KM_PUSHPAGE); ctx = kmem_cache_alloc(lz4_cache, KM_SLEEP);
/* /*
* out of kernel memory, gently fall through - this will disable * out of kernel memory, gently fall through - this will disable

View File

@ -57,7 +57,7 @@ lzjb_compress(void *s_start, void *d_start, size_t s_len, size_t d_len, int n)
uint16_t *hp; uint16_t *hp;
uint16_t *lempel; uint16_t *lempel;
lempel = kmem_zalloc(LEMPEL_SIZE * sizeof (uint16_t), KM_PUSHPAGE); lempel = kmem_zalloc(LEMPEL_SIZE * sizeof (uint16_t), KM_SLEEP);
while (src < (uchar_t *)s_start + s_len) { while (src < (uchar_t *)s_start + s_len) {
if ((copymask <<= 1) == (1 << NBBY)) { if ((copymask <<= 1) == (1 << NBBY)) {
if (dst >= (uchar_t *)d_start + d_len - 1 - 2 * NBBY) { if (dst >= (uchar_t *)d_start + d_len - 1 - 2 * NBBY) {

View File

@ -192,7 +192,7 @@ metaslab_class_create(spa_t *spa, metaslab_ops_t *ops)
{ {
metaslab_class_t *mc; metaslab_class_t *mc;
mc = kmem_zalloc(sizeof (metaslab_class_t), KM_PUSHPAGE); mc = kmem_zalloc(sizeof (metaslab_class_t), KM_SLEEP);
mc->mc_spa = spa; mc->mc_spa = spa;
mc->mc_rotor = NULL; mc->mc_rotor = NULL;
@ -286,7 +286,7 @@ metaslab_class_histogram_verify(metaslab_class_t *mc)
return; return;
mc_hist = kmem_zalloc(sizeof (uint64_t) * RANGE_TREE_HISTOGRAM_SIZE, mc_hist = kmem_zalloc(sizeof (uint64_t) * RANGE_TREE_HISTOGRAM_SIZE,
KM_PUSHPAGE); KM_SLEEP);
for (c = 0; c < rvd->vdev_children; c++) { for (c = 0; c < rvd->vdev_children; c++) {
vdev_t *tvd = rvd->vdev_child[c]; vdev_t *tvd = rvd->vdev_child[c];
@ -482,7 +482,7 @@ metaslab_group_create(metaslab_class_t *mc, vdev_t *vd)
{ {
metaslab_group_t *mg; metaslab_group_t *mg;
mg = kmem_zalloc(sizeof (metaslab_group_t), KM_PUSHPAGE); mg = kmem_zalloc(sizeof (metaslab_group_t), KM_SLEEP);
mutex_init(&mg->mg_lock, NULL, MUTEX_DEFAULT, NULL); mutex_init(&mg->mg_lock, NULL, MUTEX_DEFAULT, NULL);
avl_create(&mg->mg_metaslab_tree, metaslab_compare, avl_create(&mg->mg_metaslab_tree, metaslab_compare,
sizeof (metaslab_t), offsetof(struct metaslab, ms_group_node)); sizeof (metaslab_t), offsetof(struct metaslab, ms_group_node));
@ -598,7 +598,7 @@ metaslab_group_histogram_verify(metaslab_group_t *mg)
return; return;
mg_hist = kmem_zalloc(sizeof (uint64_t) * RANGE_TREE_HISTOGRAM_SIZE, mg_hist = kmem_zalloc(sizeof (uint64_t) * RANGE_TREE_HISTOGRAM_SIZE,
KM_PUSHPAGE); KM_SLEEP);
ASSERT3U(RANGE_TREE_HISTOGRAM_SIZE, >=, ASSERT3U(RANGE_TREE_HISTOGRAM_SIZE, >=,
SPACE_MAP_HISTOGRAM_SIZE + ashift); SPACE_MAP_HISTOGRAM_SIZE + ashift);
@ -1246,7 +1246,7 @@ metaslab_init(metaslab_group_t *mg, uint64_t id, uint64_t object, uint64_t txg,
metaslab_t *ms; metaslab_t *ms;
int error; int error;
ms = kmem_zalloc(sizeof (metaslab_t), KM_PUSHPAGE); ms = kmem_zalloc(sizeof (metaslab_t), KM_SLEEP);
mutex_init(&ms->ms_lock, NULL, MUTEX_DEFAULT, NULL); mutex_init(&ms->ms_lock, NULL, MUTEX_DEFAULT, NULL);
cv_init(&ms->ms_load_cv, NULL, CV_DEFAULT, NULL); cv_init(&ms->ms_load_cv, NULL, CV_DEFAULT, NULL);
ms->ms_id = id; ms->ms_id = id;
@ -1639,7 +1639,7 @@ metaslab_group_preload(metaslab_group_t *mg)
*/ */
mutex_exit(&mg->mg_lock); mutex_exit(&mg->mg_lock);
VERIFY(taskq_dispatch(mg->mg_taskq, metaslab_preload, VERIFY(taskq_dispatch(mg->mg_taskq, metaslab_preload,
msp, TQ_PUSHPAGE) != 0); msp, TQ_SLEEP) != 0);
mutex_enter(&mg->mg_lock); mutex_enter(&mg->mg_lock);
msp = msp_next; msp = msp_next;
} }

View File

@ -132,7 +132,7 @@ range_tree_create(range_tree_ops_t *ops, void *arg, kmutex_t *lp)
{ {
range_tree_t *rt; range_tree_t *rt;
rt = kmem_zalloc(sizeof (range_tree_t), KM_PUSHPAGE); rt = kmem_zalloc(sizeof (range_tree_t), KM_SLEEP);
avl_create(&rt->rt_root, range_tree_seg_compare, avl_create(&rt->rt_root, range_tree_seg_compare,
sizeof (range_seg_t), offsetof(range_seg_t, rs_node)); sizeof (range_seg_t), offsetof(range_seg_t, rs_node));
@ -221,7 +221,7 @@ range_tree_add(void *arg, uint64_t start, uint64_t size)
rs_after->rs_start = start; rs_after->rs_start = start;
rs = rs_after; rs = rs_after;
} else { } else {
rs = kmem_cache_alloc(range_seg_cache, KM_PUSHPAGE); rs = kmem_cache_alloc(range_seg_cache, KM_SLEEP);
rs->rs_start = start; rs->rs_start = start;
rs->rs_end = end; rs->rs_end = end;
avl_insert(&rt->rt_root, rs, where); avl_insert(&rt->rt_root, rs, where);
@ -270,7 +270,7 @@ range_tree_remove(void *arg, uint64_t start, uint64_t size)
rt->rt_ops->rtop_remove(rt, rs, rt->rt_arg); rt->rt_ops->rtop_remove(rt, rs, rt->rt_arg);
if (left_over && right_over) { if (left_over && right_over) {
newseg = kmem_cache_alloc(range_seg_cache, KM_PUSHPAGE); newseg = kmem_cache_alloc(range_seg_cache, KM_SLEEP);
newseg->rs_start = end; newseg->rs_start = end;
newseg->rs_end = rs->rs_end; newseg->rs_end = rs->rs_end;
range_tree_stat_incr(rt, newseg); range_tree_stat_incr(rt, newseg);

View File

@ -121,7 +121,7 @@ refcount_add_many(refcount_t *rc, uint64_t number, void *holder)
int64_t count; int64_t count;
if (rc->rc_tracked) { if (rc->rc_tracked) {
ref = kmem_cache_alloc(reference_cache, KM_PUSHPAGE); ref = kmem_cache_alloc(reference_cache, KM_SLEEP);
ref->ref_holder = holder; ref->ref_holder = holder;
ref->ref_number = number; ref->ref_number = number;
} }
@ -165,7 +165,7 @@ refcount_remove_many(refcount_t *rc, uint64_t number, void *holder)
if (reference_history > 0) { if (reference_history > 0) {
ref->ref_removed = ref->ref_removed =
kmem_cache_alloc(reference_history_cache, kmem_cache_alloc(reference_history_cache,
KM_PUSHPAGE); KM_SLEEP);
list_insert_head(&rc->rc_removed, ref); list_insert_head(&rc->rc_removed, ref);
rc->rc_removed_count++; rc->rc_removed_count++;
if (rc->rc_removed_count > reference_history) { if (rc->rc_removed_count > reference_history) {

View File

@ -103,7 +103,7 @@ rrn_add(rrwlock_t *rrl, void *tag)
{ {
rrw_node_t *rn; rrw_node_t *rn;
rn = kmem_alloc(sizeof (*rn), KM_PUSHPAGE); rn = kmem_alloc(sizeof (*rn), KM_SLEEP);
rn->rn_rrl = rrl; rn->rn_rrl = rrl;
rn->rn_next = tsd_get(rrw_tsd_key); rn->rn_next = tsd_get(rrw_tsd_key);
rn->rn_tag = tag; rn->rn_tag = tag;

View File

@ -434,10 +434,10 @@ sa_add_layout_entry(objset_t *os, sa_attr_type_t *attrs, int attr_count,
avl_index_t loc; avl_index_t loc;
ASSERT(MUTEX_HELD(&sa->sa_lock)); ASSERT(MUTEX_HELD(&sa->sa_lock));
tb = kmem_zalloc(sizeof (sa_lot_t), KM_PUSHPAGE); tb = kmem_zalloc(sizeof (sa_lot_t), KM_SLEEP);
tb->lot_attr_count = attr_count; tb->lot_attr_count = attr_count;
tb->lot_attrs = kmem_alloc(sizeof (sa_attr_type_t) * attr_count, tb->lot_attrs = kmem_alloc(sizeof (sa_attr_type_t) * attr_count,
KM_PUSHPAGE); KM_SLEEP);
bcopy(attrs, tb->lot_attrs, sizeof (sa_attr_type_t) * attr_count); bcopy(attrs, tb->lot_attrs, sizeof (sa_attr_type_t) * attr_count);
tb->lot_num = lot_num; tb->lot_num = lot_num;
tb->lot_hash = hash; tb->lot_hash = hash;
@ -740,7 +740,7 @@ sa_build_layouts(sa_handle_t *hdl, sa_bulk_attr_t *attr_desc, int attr_count,
buf_space = hdl->sa_bonus->db_size - hdrsize; buf_space = hdl->sa_bonus->db_size - hdrsize;
attrs_start = attrs = kmem_alloc(sizeof (sa_attr_type_t) * attr_count, attrs_start = attrs = kmem_alloc(sizeof (sa_attr_type_t) * attr_count,
KM_PUSHPAGE); KM_SLEEP);
lot_count = 0; lot_count = 0;
for (i = 0, len_idx = 0, hash = -1ULL; i != attr_count; i++) { for (i = 0, len_idx = 0, hash = -1ULL; i != attr_count; i++) {
@ -864,7 +864,7 @@ sa_attr_table_setup(objset_t *os, sa_attr_reg_t *reg_attrs, int count)
dmu_objset_type_t ostype = dmu_objset_type(os); dmu_objset_type_t ostype = dmu_objset_type(os);
sa->sa_user_table = sa->sa_user_table =
kmem_zalloc(count * sizeof (sa_attr_type_t), KM_PUSHPAGE); kmem_zalloc(count * sizeof (sa_attr_type_t), KM_SLEEP);
sa->sa_user_table_sz = count * sizeof (sa_attr_type_t); sa->sa_user_table_sz = count * sizeof (sa_attr_type_t);
if (sa->sa_reg_attr_obj != 0) { if (sa->sa_reg_attr_obj != 0) {
@ -923,7 +923,7 @@ sa_attr_table_setup(objset_t *os, sa_attr_reg_t *reg_attrs, int count)
sa->sa_num_attrs = sa_attr_count; sa->sa_num_attrs = sa_attr_count;
tb = sa->sa_attr_table = tb = sa->sa_attr_table =
kmem_zalloc(sizeof (sa_attr_table_t) * sa_attr_count, KM_PUSHPAGE); kmem_zalloc(sizeof (sa_attr_table_t) * sa_attr_count, KM_SLEEP);
/* /*
* Attribute table is constructed from requested attribute list, * Attribute table is constructed from requested attribute list,
@ -948,7 +948,7 @@ sa_attr_table_setup(objset_t *os, sa_attr_reg_t *reg_attrs, int count)
continue; continue;
} }
tb[ATTR_NUM(value)].sa_name = tb[ATTR_NUM(value)].sa_name =
kmem_zalloc(strlen(za.za_name) +1, KM_PUSHPAGE); kmem_zalloc(strlen(za.za_name) +1, KM_SLEEP);
(void) strlcpy(tb[ATTR_NUM(value)].sa_name, za.za_name, (void) strlcpy(tb[ATTR_NUM(value)].sa_name, za.za_name,
strlen(za.za_name) +1); strlen(za.za_name) +1);
} }
@ -974,7 +974,7 @@ sa_attr_table_setup(objset_t *os, sa_attr_reg_t *reg_attrs, int count)
tb[i].sa_registered = B_FALSE; tb[i].sa_registered = B_FALSE;
tb[i].sa_name = tb[i].sa_name =
kmem_zalloc(strlen(sa_legacy_attrs[i].sa_name) +1, kmem_zalloc(strlen(sa_legacy_attrs[i].sa_name) +1,
KM_PUSHPAGE); KM_SLEEP);
(void) strlcpy(tb[i].sa_name, (void) strlcpy(tb[i].sa_name,
sa_legacy_attrs[i].sa_name, sa_legacy_attrs[i].sa_name,
strlen(sa_legacy_attrs[i].sa_name) + 1); strlen(sa_legacy_attrs[i].sa_name) + 1);
@ -992,7 +992,7 @@ sa_attr_table_setup(objset_t *os, sa_attr_reg_t *reg_attrs, int count)
tb[attr_id].sa_byteswap = reg_attrs[i].sa_byteswap; tb[attr_id].sa_byteswap = reg_attrs[i].sa_byteswap;
tb[attr_id].sa_attr = attr_id; tb[attr_id].sa_attr = attr_id;
tb[attr_id].sa_name = tb[attr_id].sa_name =
kmem_zalloc(strlen(reg_attrs[i].sa_name) + 1, KM_PUSHPAGE); kmem_zalloc(strlen(reg_attrs[i].sa_name) + 1, KM_SLEEP);
(void) strlcpy(tb[attr_id].sa_name, reg_attrs[i].sa_name, (void) strlcpy(tb[attr_id].sa_name, reg_attrs[i].sa_name,
strlen(reg_attrs[i].sa_name) + 1); strlen(reg_attrs[i].sa_name) + 1);
} }
@ -1029,7 +1029,7 @@ sa_setup(objset_t *os, uint64_t sa_obj, sa_attr_reg_t *reg_attrs, int count,
return (0); return (0);
} }
sa = kmem_zalloc(sizeof (sa_os_t), KM_PUSHPAGE); sa = kmem_zalloc(sizeof (sa_os_t), KM_SLEEP);
mutex_init(&sa->sa_lock, NULL, MUTEX_DEFAULT, NULL); mutex_init(&sa->sa_lock, NULL, MUTEX_DEFAULT, NULL);
sa->sa_master_obj = sa_obj; sa->sa_master_obj = sa_obj;
@ -1077,7 +1077,7 @@ sa_setup(objset_t *os, uint64_t sa_obj, sa_attr_reg_t *reg_attrs, int count,
uint64_t lot_num; uint64_t lot_num;
lot_attrs = kmem_zalloc(sizeof (sa_attr_type_t) * lot_attrs = kmem_zalloc(sizeof (sa_attr_type_t) *
za.za_num_integers, KM_PUSHPAGE); za.za_num_integers, KM_SLEEP);
if ((error = (zap_lookup(os, sa->sa_layout_attr_obj, if ((error = (zap_lookup(os, sa->sa_layout_attr_obj,
za.za_name, 2, za.za_num_integers, za.za_name, 2, za.za_num_integers,
@ -1563,14 +1563,14 @@ sa_find_idx_tab(objset_t *os, dmu_object_type_t bonustype, void *data)
} }
/* No such luck, create a new entry */ /* No such luck, create a new entry */
idx_tab = kmem_zalloc(sizeof (sa_idx_tab_t), KM_PUSHPAGE); idx_tab = kmem_zalloc(sizeof (sa_idx_tab_t), KM_SLEEP);
idx_tab->sa_idx_tab = idx_tab->sa_idx_tab =
kmem_zalloc(sizeof (uint32_t) * sa->sa_num_attrs, KM_PUSHPAGE); kmem_zalloc(sizeof (uint32_t) * sa->sa_num_attrs, KM_SLEEP);
idx_tab->sa_layout = tb; idx_tab->sa_layout = tb;
refcount_create(&idx_tab->sa_refcount); refcount_create(&idx_tab->sa_refcount);
if (tb->lot_var_sizes) if (tb->lot_var_sizes)
idx_tab->sa_variable_lengths = kmem_alloc(sizeof (uint16_t) * idx_tab->sa_variable_lengths = kmem_alloc(sizeof (uint16_t) *
tb->lot_var_sizes, KM_PUSHPAGE); tb->lot_var_sizes, KM_SLEEP);
sa_attr_iter(os, hdr, bonustype, sa_build_idx_tab, sa_attr_iter(os, hdr, bonustype, sa_build_idx_tab,
tb, idx_tab); tb, idx_tab);

View File

@ -170,7 +170,7 @@ spa_prop_add_list(nvlist_t *nvl, zpool_prop_t prop, char *strval,
const char *propname = zpool_prop_to_name(prop); const char *propname = zpool_prop_to_name(prop);
nvlist_t *propval; nvlist_t *propval;
VERIFY(nvlist_alloc(&propval, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); VERIFY(nvlist_alloc(&propval, NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_uint64(propval, ZPROP_SOURCE, src) == 0); VERIFY(nvlist_add_uint64(propval, ZPROP_SOURCE, src) == 0);
if (strval != NULL) if (strval != NULL)
@ -285,7 +285,7 @@ spa_prop_get(spa_t *spa, nvlist_t **nvp)
zap_attribute_t za; zap_attribute_t za;
int err; int err;
err = nvlist_alloc(nvp, NV_UNIQUE_NAME, KM_PUSHPAGE); err = nvlist_alloc(nvp, NV_UNIQUE_NAME, KM_SLEEP);
if (err) if (err)
return (err); return (err);
@ -337,7 +337,7 @@ spa_prop_get(spa_t *spa, nvlist_t **nvp)
strval = kmem_alloc( strval = kmem_alloc(
MAXNAMELEN + strlen(MOS_DIR_NAME) + 1, MAXNAMELEN + strlen(MOS_DIR_NAME) + 1,
KM_PUSHPAGE); KM_SLEEP);
dsl_dataset_name(ds, strval); dsl_dataset_name(ds, strval);
dsl_dataset_rele(ds, FTAG); dsl_dataset_rele(ds, FTAG);
dsl_pool_config_exit(dp, FTAG); dsl_pool_config_exit(dp, FTAG);
@ -356,7 +356,7 @@ spa_prop_get(spa_t *spa, nvlist_t **nvp)
case 1: case 1:
/* string property */ /* string property */
strval = kmem_alloc(za.za_num_integers, KM_PUSHPAGE); strval = kmem_alloc(za.za_num_integers, KM_SLEEP);
err = zap_lookup(mos, spa->spa_pool_props_object, err = zap_lookup(mos, spa->spa_pool_props_object,
za.za_name, 1, za.za_num_integers, strval); za.za_name, 1, za.za_num_integers, strval);
if (err) { if (err) {
@ -609,7 +609,7 @@ spa_configfile_set(spa_t *spa, nvlist_t *nvp, boolean_t need_sync)
return; return;
dp = kmem_alloc(sizeof (spa_config_dirent_t), dp = kmem_alloc(sizeof (spa_config_dirent_t),
KM_PUSHPAGE); KM_SLEEP);
if (cachefile[0] == '\0') if (cachefile[0] == '\0')
dp->scd_path = spa_strdup(spa_config_path); dp->scd_path = spa_strdup(spa_config_path);
@ -1380,7 +1380,7 @@ spa_load_spares(spa_t *spa)
* active configuration, then we also mark this vdev as an active spare. * active configuration, then we also mark this vdev as an active spare.
*/ */
spa->spa_spares.sav_vdevs = kmem_zalloc(nspares * sizeof (void *), spa->spa_spares.sav_vdevs = kmem_zalloc(nspares * sizeof (void *),
KM_PUSHPAGE); KM_SLEEP);
for (i = 0; i < spa->spa_spares.sav_count; i++) { for (i = 0; i < spa->spa_spares.sav_count; i++) {
VERIFY(spa_config_parse(spa, &vd, spares[i], NULL, 0, VERIFY(spa_config_parse(spa, &vd, spares[i], NULL, 0,
VDEV_ALLOC_SPARE) == 0); VDEV_ALLOC_SPARE) == 0);
@ -1428,7 +1428,7 @@ spa_load_spares(spa_t *spa)
DATA_TYPE_NVLIST_ARRAY) == 0); DATA_TYPE_NVLIST_ARRAY) == 0);
spares = kmem_alloc(spa->spa_spares.sav_count * sizeof (void *), spares = kmem_alloc(spa->spa_spares.sav_count * sizeof (void *),
KM_PUSHPAGE); KM_SLEEP);
for (i = 0; i < spa->spa_spares.sav_count; i++) for (i = 0; i < spa->spa_spares.sav_count; i++)
spares[i] = vdev_config_generate(spa, spares[i] = vdev_config_generate(spa,
spa->spa_spares.sav_vdevs[i], B_TRUE, VDEV_CONFIG_SPARE); spa->spa_spares.sav_vdevs[i], B_TRUE, VDEV_CONFIG_SPARE);
@ -1462,7 +1462,7 @@ spa_load_l2cache(spa_t *spa)
if (sav->sav_config != NULL) { if (sav->sav_config != NULL) {
VERIFY(nvlist_lookup_nvlist_array(sav->sav_config, VERIFY(nvlist_lookup_nvlist_array(sav->sav_config,
ZPOOL_CONFIG_L2CACHE, &l2cache, &nl2cache) == 0); ZPOOL_CONFIG_L2CACHE, &l2cache, &nl2cache) == 0);
newvdevs = kmem_alloc(nl2cache * sizeof (void *), KM_PUSHPAGE); newvdevs = kmem_alloc(nl2cache * sizeof (void *), KM_SLEEP);
} else { } else {
nl2cache = 0; nl2cache = 0;
newvdevs = NULL; newvdevs = NULL;
@ -1557,7 +1557,7 @@ spa_load_l2cache(spa_t *spa)
VERIFY(nvlist_remove(sav->sav_config, ZPOOL_CONFIG_L2CACHE, VERIFY(nvlist_remove(sav->sav_config, ZPOOL_CONFIG_L2CACHE,
DATA_TYPE_NVLIST_ARRAY) == 0); DATA_TYPE_NVLIST_ARRAY) == 0);
l2cache = kmem_alloc(sav->sav_count * sizeof (void *), KM_PUSHPAGE); l2cache = kmem_alloc(sav->sav_count * sizeof (void *), KM_SLEEP);
for (i = 0; i < sav->sav_count; i++) for (i = 0; i < sav->sav_count; i++)
l2cache[i] = vdev_config_generate(spa, l2cache[i] = vdev_config_generate(spa,
sav->sav_vdevs[i], B_TRUE, VDEV_CONFIG_L2CACHE); sav->sav_vdevs[i], B_TRUE, VDEV_CONFIG_L2CACHE);
@ -1586,7 +1586,7 @@ load_nvlist(spa_t *spa, uint64_t obj, nvlist_t **value)
nvsize = *(uint64_t *)db->db_data; nvsize = *(uint64_t *)db->db_data;
dmu_buf_rele(db, FTAG); dmu_buf_rele(db, FTAG);
packed = kmem_alloc(nvsize, KM_PUSHPAGE); packed = kmem_alloc(nvsize, KM_SLEEP);
error = dmu_read(spa->spa_meta_objset, obj, 0, nvsize, packed, error = dmu_read(spa->spa_meta_objset, obj, 0, nvsize, packed,
DMU_READ_PREFETCH); DMU_READ_PREFETCH);
if (error == 0) if (error == 0)
@ -1643,8 +1643,8 @@ spa_config_valid(spa_t *spa, nvlist_t *config)
uint64_t idx = 0; uint64_t idx = 0;
child = kmem_alloc(rvd->vdev_children * sizeof (nvlist_t **), child = kmem_alloc(rvd->vdev_children * sizeof (nvlist_t **),
KM_PUSHPAGE); KM_SLEEP);
VERIFY(nvlist_alloc(&nv, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); VERIFY(nvlist_alloc(&nv, NV_UNIQUE_NAME, KM_SLEEP) == 0);
for (c = 0; c < rvd->vdev_children; c++) { for (c = 0; c < rvd->vdev_children; c++) {
vdev_t *tvd = rvd->vdev_child[c]; vdev_t *tvd = rvd->vdev_child[c];
@ -2037,7 +2037,7 @@ spa_try_repair(spa_t *spa, nvlist_t *config)
&glist, &gcount) != 0) &glist, &gcount) != 0)
return; return;
vd = kmem_zalloc(gcount * sizeof (vdev_t *), KM_PUSHPAGE); vd = kmem_zalloc(gcount * sizeof (vdev_t *), KM_SLEEP);
/* attempt to online all the vdevs & validate */ /* attempt to online all the vdevs & validate */
attempt_reopen = B_TRUE; attempt_reopen = B_TRUE;
@ -2123,7 +2123,7 @@ spa_load(spa_t *spa, spa_load_state_t state, spa_import_type_t type,
if (nvlist_lookup_nvlist(config, ZPOOL_CONFIG_SPLIT, if (nvlist_lookup_nvlist(config, ZPOOL_CONFIG_SPLIT,
&nvl) == 0) { &nvl) == 0) {
VERIFY(nvlist_dup(nvl, &spa->spa_config_splitting, VERIFY(nvlist_dup(nvl, &spa->spa_config_splitting,
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
} }
nvlist_free(spa->spa_load_info); nvlist_free(spa->spa_load_info);
@ -3010,7 +3010,7 @@ spa_open_common(const char *pool, spa_t **spapp, void *tag, nvlist_t *nvpolicy,
*/ */
if (config != NULL && spa->spa_config) { if (config != NULL && spa->spa_config) {
VERIFY(nvlist_dup(spa->spa_config, config, VERIFY(nvlist_dup(spa->spa_config, config,
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
VERIFY(nvlist_add_nvlist(*config, VERIFY(nvlist_add_nvlist(*config,
ZPOOL_CONFIG_LOAD_INFO, ZPOOL_CONFIG_LOAD_INFO,
spa->spa_load_info) == 0); spa->spa_load_info) == 0);
@ -3434,13 +3434,13 @@ spa_set_aux_vdevs(spa_aux_vdev_t *sav, nvlist_t **devs, int ndevs,
&olddevs, &oldndevs) == 0); &olddevs, &oldndevs) == 0);
newdevs = kmem_alloc(sizeof (void *) * newdevs = kmem_alloc(sizeof (void *) *
(ndevs + oldndevs), KM_PUSHPAGE); (ndevs + oldndevs), KM_SLEEP);
for (i = 0; i < oldndevs; i++) for (i = 0; i < oldndevs; i++)
VERIFY(nvlist_dup(olddevs[i], &newdevs[i], VERIFY(nvlist_dup(olddevs[i], &newdevs[i],
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
for (i = 0; i < ndevs; i++) for (i = 0; i < ndevs; i++)
VERIFY(nvlist_dup(devs[i], &newdevs[i + oldndevs], VERIFY(nvlist_dup(devs[i], &newdevs[i + oldndevs],
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
VERIFY(nvlist_remove(sav->sav_config, config, VERIFY(nvlist_remove(sav->sav_config, config,
DATA_TYPE_NVLIST_ARRAY) == 0); DATA_TYPE_NVLIST_ARRAY) == 0);
@ -3455,7 +3455,7 @@ spa_set_aux_vdevs(spa_aux_vdev_t *sav, nvlist_t **devs, int ndevs,
* Generate a new dev list. * Generate a new dev list.
*/ */
VERIFY(nvlist_alloc(&sav->sav_config, NV_UNIQUE_NAME, VERIFY(nvlist_alloc(&sav->sav_config, NV_UNIQUE_NAME,
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
VERIFY(nvlist_add_nvlist_array(sav->sav_config, config, VERIFY(nvlist_add_nvlist_array(sav->sav_config, config,
devs, ndevs) == 0); devs, ndevs) == 0);
} }
@ -3610,7 +3610,7 @@ spa_create(const char *pool, nvlist_t *nvroot, nvlist_t *props,
if (nvlist_lookup_nvlist_array(nvroot, ZPOOL_CONFIG_SPARES, if (nvlist_lookup_nvlist_array(nvroot, ZPOOL_CONFIG_SPARES,
&spares, &nspares) == 0) { &spares, &nspares) == 0) {
VERIFY(nvlist_alloc(&spa->spa_spares.sav_config, NV_UNIQUE_NAME, VERIFY(nvlist_alloc(&spa->spa_spares.sav_config, NV_UNIQUE_NAME,
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
VERIFY(nvlist_add_nvlist_array(spa->spa_spares.sav_config, VERIFY(nvlist_add_nvlist_array(spa->spa_spares.sav_config,
ZPOOL_CONFIG_SPARES, spares, nspares) == 0); ZPOOL_CONFIG_SPARES, spares, nspares) == 0);
spa_config_enter(spa, SCL_ALL, FTAG, RW_WRITER); spa_config_enter(spa, SCL_ALL, FTAG, RW_WRITER);
@ -3625,7 +3625,7 @@ spa_create(const char *pool, nvlist_t *nvroot, nvlist_t *props,
if (nvlist_lookup_nvlist_array(nvroot, ZPOOL_CONFIG_L2CACHE, if (nvlist_lookup_nvlist_array(nvroot, ZPOOL_CONFIG_L2CACHE,
&l2cache, &nl2cache) == 0) { &l2cache, &nl2cache) == 0) {
VERIFY(nvlist_alloc(&spa->spa_l2cache.sav_config, VERIFY(nvlist_alloc(&spa->spa_l2cache.sav_config,
NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_nvlist_array(spa->spa_l2cache.sav_config, VERIFY(nvlist_add_nvlist_array(spa->spa_l2cache.sav_config,
ZPOOL_CONFIG_L2CACHE, l2cache, nl2cache) == 0); ZPOOL_CONFIG_L2CACHE, l2cache, nl2cache) == 0);
spa_config_enter(spa, SCL_ALL, FTAG, RW_WRITER); spa_config_enter(spa, SCL_ALL, FTAG, RW_WRITER);
@ -3766,7 +3766,7 @@ spa_generate_rootconf(char *devpath, char *devid, uint64_t *guid)
/* /*
* Put this pool's top-level vdevs into a root vdev. * Put this pool's top-level vdevs into a root vdev.
*/ */
VERIFY(nvlist_alloc(&nvroot, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); VERIFY(nvlist_alloc(&nvroot, NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_string(nvroot, ZPOOL_CONFIG_TYPE, VERIFY(nvlist_add_string(nvroot, ZPOOL_CONFIG_TYPE,
VDEV_TYPE_ROOT) == 0); VDEV_TYPE_ROOT) == 0);
VERIFY(nvlist_add_uint64(nvroot, ZPOOL_CONFIG_ID, 0ULL) == 0); VERIFY(nvlist_add_uint64(nvroot, ZPOOL_CONFIG_ID, 0ULL) == 0);
@ -4074,7 +4074,7 @@ spa_import(char *pool, nvlist_t *config, nvlist_t *props, uint64_t flags)
ZPOOL_CONFIG_SPARES, DATA_TYPE_NVLIST_ARRAY) == 0); ZPOOL_CONFIG_SPARES, DATA_TYPE_NVLIST_ARRAY) == 0);
else else
VERIFY(nvlist_alloc(&spa->spa_spares.sav_config, VERIFY(nvlist_alloc(&spa->spa_spares.sav_config,
NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_nvlist_array(spa->spa_spares.sav_config, VERIFY(nvlist_add_nvlist_array(spa->spa_spares.sav_config,
ZPOOL_CONFIG_SPARES, spares, nspares) == 0); ZPOOL_CONFIG_SPARES, spares, nspares) == 0);
spa_config_enter(spa, SCL_ALL, FTAG, RW_WRITER); spa_config_enter(spa, SCL_ALL, FTAG, RW_WRITER);
@ -4089,7 +4089,7 @@ spa_import(char *pool, nvlist_t *config, nvlist_t *props, uint64_t flags)
ZPOOL_CONFIG_L2CACHE, DATA_TYPE_NVLIST_ARRAY) == 0); ZPOOL_CONFIG_L2CACHE, DATA_TYPE_NVLIST_ARRAY) == 0);
else else
VERIFY(nvlist_alloc(&spa->spa_l2cache.sav_config, VERIFY(nvlist_alloc(&spa->spa_l2cache.sav_config,
NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_nvlist_array(spa->spa_l2cache.sav_config, VERIFY(nvlist_add_nvlist_array(spa->spa_l2cache.sav_config,
ZPOOL_CONFIG_L2CACHE, l2cache, nl2cache) == 0); ZPOOL_CONFIG_L2CACHE, l2cache, nl2cache) == 0);
spa_config_enter(spa, SCL_ALL, FTAG, RW_WRITER); spa_config_enter(spa, SCL_ALL, FTAG, RW_WRITER);
@ -4180,7 +4180,7 @@ spa_tryimport(nvlist_t *tryconfig)
* pools are bootable. * pools are bootable.
*/ */
if ((!error || error == EEXIST) && spa->spa_bootfs) { if ((!error || error == EEXIST) && spa->spa_bootfs) {
char *tmpname = kmem_alloc(MAXPATHLEN, KM_PUSHPAGE); char *tmpname = kmem_alloc(MAXPATHLEN, KM_SLEEP);
/* /*
* We have to play games with the name since the * We have to play games with the name since the
@ -4191,7 +4191,7 @@ spa_tryimport(nvlist_t *tryconfig)
char *cp; char *cp;
char *dsname; char *dsname;
dsname = kmem_alloc(MAXPATHLEN, KM_PUSHPAGE); dsname = kmem_alloc(MAXPATHLEN, KM_SLEEP);
cp = strchr(tmpname, '/'); cp = strchr(tmpname, '/');
if (cp == NULL) { if (cp == NULL) {
@ -4596,7 +4596,7 @@ spa_vdev_attach(spa_t *spa, uint64_t guid, nvlist_t *nvroot, int replacing)
if (strcmp(oldvd->vdev_path, newvd->vdev_path) == 0) { if (strcmp(oldvd->vdev_path, newvd->vdev_path) == 0) {
spa_strfree(oldvd->vdev_path); spa_strfree(oldvd->vdev_path);
oldvd->vdev_path = kmem_alloc(strlen(newvd->vdev_path) + 5, oldvd->vdev_path = kmem_alloc(strlen(newvd->vdev_path) + 5,
KM_PUSHPAGE); KM_SLEEP);
(void) sprintf(oldvd->vdev_path, "%s/%s", (void) sprintf(oldvd->vdev_path, "%s/%s",
newvd->vdev_path, "old"); newvd->vdev_path, "old");
if (oldvd->vdev_devid != NULL) { if (oldvd->vdev_devid != NULL) {
@ -4992,8 +4992,8 @@ spa_vdev_split_mirror(spa_t *spa, char *newname, nvlist_t *config,
nvlist_lookup_nvlist(nvl, ZPOOL_CONFIG_L2CACHE, &tmp) == 0) nvlist_lookup_nvlist(nvl, ZPOOL_CONFIG_L2CACHE, &tmp) == 0)
return (spa_vdev_exit(spa, NULL, txg, EINVAL)); return (spa_vdev_exit(spa, NULL, txg, EINVAL));
vml = kmem_zalloc(children * sizeof (vdev_t *), KM_PUSHPAGE); vml = kmem_zalloc(children * sizeof (vdev_t *), KM_SLEEP);
glist = kmem_zalloc(children * sizeof (uint64_t), KM_PUSHPAGE); glist = kmem_zalloc(children * sizeof (uint64_t), KM_SLEEP);
/* then, loop over each vdev and validate it */ /* then, loop over each vdev and validate it */
for (c = 0; c < children; c++) { for (c = 0; c < children; c++) {
@ -5073,7 +5073,7 @@ spa_vdev_split_mirror(spa_t *spa, char *newname, nvlist_t *config,
* Temporarily record the splitting vdevs in the spa config. This * Temporarily record the splitting vdevs in the spa config. This
* will disappear once the config is regenerated. * will disappear once the config is regenerated.
*/ */
VERIFY(nvlist_alloc(&nvl, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); VERIFY(nvlist_alloc(&nvl, NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_uint64_array(nvl, ZPOOL_CONFIG_SPLIT_LIST, VERIFY(nvlist_add_uint64_array(nvl, ZPOOL_CONFIG_SPLIT_LIST,
glist, children) == 0); glist, children) == 0);
kmem_free(glist, children * sizeof (uint64_t)); kmem_free(glist, children * sizeof (uint64_t));
@ -5120,7 +5120,7 @@ spa_vdev_split_mirror(spa_t *spa, char *newname, nvlist_t *config,
/* if that worked, generate a real config for the new pool */ /* if that worked, generate a real config for the new pool */
if (newspa->spa_root_vdev != NULL) { if (newspa->spa_root_vdev != NULL) {
VERIFY(nvlist_alloc(&newspa->spa_config_splitting, VERIFY(nvlist_alloc(&newspa->spa_config_splitting,
NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_uint64(newspa->spa_config_splitting, VERIFY(nvlist_add_uint64(newspa->spa_config_splitting,
ZPOOL_CONFIG_SPLIT_GUID, spa_guid(spa)) == 0); ZPOOL_CONFIG_SPLIT_GUID, spa_guid(spa)) == 0);
spa_config_set(newspa, spa_config_generate(newspa, NULL, -1ULL, spa_config_set(newspa, spa_config_generate(newspa, NULL, -1ULL,
@ -5231,12 +5231,12 @@ spa_vdev_remove_aux(nvlist_t *config, char *name, nvlist_t **dev, int count,
int i, j; int i, j;
if (count > 1) if (count > 1)
newdev = kmem_alloc((count - 1) * sizeof (void *), KM_PUSHPAGE); newdev = kmem_alloc((count - 1) * sizeof (void *), KM_SLEEP);
for (i = 0, j = 0; i < count; i++) { for (i = 0, j = 0; i < count; i++) {
if (dev[i] == dev_to_remove) if (dev[i] == dev_to_remove)
continue; continue;
VERIFY(nvlist_dup(dev[i], &newdev[j++], KM_PUSHPAGE) == 0); VERIFY(nvlist_dup(dev[i], &newdev[j++], KM_SLEEP) == 0);
} }
VERIFY(nvlist_remove(config, name, DATA_TYPE_NVLIST_ARRAY) == 0); VERIFY(nvlist_remove(config, name, DATA_TYPE_NVLIST_ARRAY) == 0);
@ -5910,10 +5910,10 @@ spa_sync_nvlist(spa_t *spa, uint64_t obj, nvlist_t *nv, dmu_tx_t *tx)
* saves us a pre-read to get data we don't actually care about. * saves us a pre-read to get data we don't actually care about.
*/ */
bufsize = P2ROUNDUP((uint64_t)nvsize, SPA_CONFIG_BLOCKSIZE); bufsize = P2ROUNDUP((uint64_t)nvsize, SPA_CONFIG_BLOCKSIZE);
packed = vmem_alloc(bufsize, KM_PUSHPAGE); packed = vmem_alloc(bufsize, KM_SLEEP);
VERIFY(nvlist_pack(nv, &packed, &nvsize, NV_ENCODE_XDR, VERIFY(nvlist_pack(nv, &packed, &nvsize, NV_ENCODE_XDR,
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
bzero(packed + nvsize, bufsize - nvsize); bzero(packed + nvsize, bufsize - nvsize);
dmu_write(spa->spa_meta_objset, obj, 0, bufsize, packed, tx); dmu_write(spa->spa_meta_objset, obj, 0, bufsize, packed, tx);
@ -5951,11 +5951,11 @@ spa_sync_aux_dev(spa_t *spa, spa_aux_vdev_t *sav, dmu_tx_t *tx,
&sav->sav_object, tx) == 0); &sav->sav_object, tx) == 0);
} }
VERIFY(nvlist_alloc(&nvroot, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); VERIFY(nvlist_alloc(&nvroot, NV_UNIQUE_NAME, KM_SLEEP) == 0);
if (sav->sav_count == 0) { if (sav->sav_count == 0) {
VERIFY(nvlist_add_nvlist_array(nvroot, config, NULL, 0) == 0); VERIFY(nvlist_add_nvlist_array(nvroot, config, NULL, 0) == 0);
} else { } else {
list = kmem_alloc(sav->sav_count*sizeof (void *), KM_PUSHPAGE); list = kmem_alloc(sav->sav_count*sizeof (void *), KM_SLEEP);
for (i = 0; i < sav->sav_count; i++) for (i = 0; i < sav->sav_count; i++)
list[i] = vdev_config_generate(spa, sav->sav_vdevs[i], list[i] = vdev_config_generate(spa, sav->sav_vdevs[i],
B_FALSE, VDEV_CONFIG_L2CACHE); B_FALSE, VDEV_CONFIG_L2CACHE);
@ -6283,7 +6283,7 @@ spa_sync(spa_t *spa, uint64_t txg)
spa->spa_sync_starttime = gethrtime(); spa->spa_sync_starttime = gethrtime();
taskq_cancel_id(system_taskq, spa->spa_deadman_tqid); taskq_cancel_id(system_taskq, spa->spa_deadman_tqid);
spa->spa_deadman_tqid = taskq_dispatch_delay(system_taskq, spa->spa_deadman_tqid = taskq_dispatch_delay(system_taskq,
spa_deadman, spa, TQ_PUSHPAGE, ddi_get_lbolt() + spa_deadman, spa, TQ_SLEEP, ddi_get_lbolt() +
NSEC_TO_TICK(spa->spa_deadman_synctime)); NSEC_TO_TICK(spa->spa_deadman_synctime));
/* /*

View File

@ -87,7 +87,7 @@ spa_config_load(void)
/* /*
* Open the configuration file. * Open the configuration file.
*/ */
pathname = kmem_alloc(MAXPATHLEN, KM_PUSHPAGE); pathname = kmem_alloc(MAXPATHLEN, KM_SLEEP);
(void) snprintf(pathname, MAXPATHLEN, "%s%s", (void) snprintf(pathname, MAXPATHLEN, "%s%s",
(rootdir != NULL) ? "./" : "", spa_config_path); (rootdir != NULL) ? "./" : "", spa_config_path);
@ -102,7 +102,7 @@ spa_config_load(void)
if (kobj_get_filesize(file, &fsize) != 0) if (kobj_get_filesize(file, &fsize) != 0)
goto out; goto out;
buf = kmem_alloc(fsize, KM_PUSHPAGE); buf = kmem_alloc(fsize, KM_SLEEP);
/* /*
* Read the nvlist from the file. * Read the nvlist from the file.
@ -113,7 +113,7 @@ spa_config_load(void)
/* /*
* Unpack the nvlist. * Unpack the nvlist.
*/ */
if (nvlist_unpack(buf, fsize, &nvlist, KM_PUSHPAGE) != 0) if (nvlist_unpack(buf, fsize, &nvlist, KM_SLEEP) != 0)
goto out; goto out;
/* /*
@ -165,11 +165,11 @@ spa_config_write(spa_config_dirent_t *dp, nvlist_t *nvl)
*/ */
VERIFY(nvlist_size(nvl, &buflen, NV_ENCODE_XDR) == 0); VERIFY(nvlist_size(nvl, &buflen, NV_ENCODE_XDR) == 0);
buf = kmem_alloc(buflen, KM_PUSHPAGE); buf = kmem_alloc(buflen, KM_SLEEP);
temp = kmem_zalloc(MAXPATHLEN, KM_PUSHPAGE); temp = kmem_zalloc(MAXPATHLEN, KM_SLEEP);
VERIFY(nvlist_pack(nvl, &buf, &buflen, NV_ENCODE_XDR, VERIFY(nvlist_pack(nvl, &buf, &buflen, NV_ENCODE_XDR,
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
/* /*
* Write the configuration to disk. We need to do the traditional * Write the configuration to disk. We need to do the traditional
@ -252,7 +252,7 @@ spa_config_sync(spa_t *target, boolean_t removing, boolean_t postsysevent)
if (nvl == NULL) if (nvl == NULL)
VERIFY(nvlist_alloc(&nvl, NV_UNIQUE_NAME, VERIFY(nvlist_alloc(&nvl, NV_UNIQUE_NAME,
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
if (spa->spa_import_flags & ZFS_IMPORT_TEMP_NAME) { if (spa->spa_import_flags & ZFS_IMPORT_TEMP_NAME) {
VERIFY0(nvlist_lookup_string(spa->spa_config, VERIFY0(nvlist_lookup_string(spa->spa_config,
@ -301,7 +301,7 @@ spa_all_configs(uint64_t *generation)
if (*generation == spa_config_generation) if (*generation == spa_config_generation)
return (NULL); return (NULL);
VERIFY(nvlist_alloc(&pools, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); VERIFY(nvlist_alloc(&pools, NV_UNIQUE_NAME, KM_SLEEP) == 0);
mutex_enter(&spa_namespace_lock); mutex_enter(&spa_namespace_lock);
while ((spa = spa_next(spa)) != NULL) { while ((spa = spa_next(spa)) != NULL) {
@ -376,7 +376,7 @@ spa_config_generate(spa_t *spa, vdev_t *vd, uint64_t txg, int getstats)
} else } else
pool_name = spa_name(spa); pool_name = spa_name(spa);
VERIFY(nvlist_alloc(&config, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); VERIFY(nvlist_alloc(&config, NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_uint64(config, ZPOOL_CONFIG_VERSION, VERIFY(nvlist_add_uint64(config, ZPOOL_CONFIG_VERSION,
spa_version(spa)) == 0); spa_version(spa)) == 0);
@ -463,21 +463,21 @@ spa_config_generate(spa_t *spa, vdev_t *vd, uint64_t txg, int getstats)
ddt_stat_t *dds; ddt_stat_t *dds;
ddt_object_t *ddo; ddt_object_t *ddo;
ddh = kmem_zalloc(sizeof (ddt_histogram_t), KM_PUSHPAGE); ddh = kmem_zalloc(sizeof (ddt_histogram_t), KM_SLEEP);
ddt_get_dedup_histogram(spa, ddh); ddt_get_dedup_histogram(spa, ddh);
VERIFY(nvlist_add_uint64_array(config, VERIFY(nvlist_add_uint64_array(config,
ZPOOL_CONFIG_DDT_HISTOGRAM, ZPOOL_CONFIG_DDT_HISTOGRAM,
(uint64_t *)ddh, sizeof (*ddh) / sizeof (uint64_t)) == 0); (uint64_t *)ddh, sizeof (*ddh) / sizeof (uint64_t)) == 0);
kmem_free(ddh, sizeof (ddt_histogram_t)); kmem_free(ddh, sizeof (ddt_histogram_t));
ddo = kmem_zalloc(sizeof (ddt_object_t), KM_PUSHPAGE); ddo = kmem_zalloc(sizeof (ddt_object_t), KM_SLEEP);
ddt_get_dedup_object_stats(spa, ddo); ddt_get_dedup_object_stats(spa, ddo);
VERIFY(nvlist_add_uint64_array(config, VERIFY(nvlist_add_uint64_array(config,
ZPOOL_CONFIG_DDT_OBJ_STATS, ZPOOL_CONFIG_DDT_OBJ_STATS,
(uint64_t *)ddo, sizeof (*ddo) / sizeof (uint64_t)) == 0); (uint64_t *)ddo, sizeof (*ddo) / sizeof (uint64_t)) == 0);
kmem_free(ddo, sizeof (ddt_object_t)); kmem_free(ddo, sizeof (ddt_object_t));
dds = kmem_zalloc(sizeof (ddt_stat_t), KM_PUSHPAGE); dds = kmem_zalloc(sizeof (ddt_stat_t), KM_SLEEP);
ddt_get_dedup_stats(spa, dds); ddt_get_dedup_stats(spa, dds);
VERIFY(nvlist_add_uint64_array(config, VERIFY(nvlist_add_uint64_array(config,
ZPOOL_CONFIG_DDT_STATS, ZPOOL_CONFIG_DDT_STATS,

View File

@ -260,7 +260,7 @@ spa_history_log_sync(void *arg, dmu_tx_t *tx)
} }
VERIFY3U(nvlist_pack(nvl, &record_packed, &reclen, NV_ENCODE_NATIVE, VERIFY3U(nvlist_pack(nvl, &record_packed, &reclen, NV_ENCODE_NATIVE,
KM_PUSHPAGE), ==, 0); KM_SLEEP), ==, 0);
mutex_enter(&spa->spa_history_lock); mutex_enter(&spa->spa_history_lock);
@ -289,9 +289,7 @@ int
spa_history_log(spa_t *spa, const char *msg) spa_history_log(spa_t *spa, const char *msg)
{ {
int err; int err;
nvlist_t *nvl; nvlist_t *nvl = fnvlist_alloc();
VERIFY0(nvlist_alloc(&nvl, NV_UNIQUE_NAME, KM_PUSHPAGE));
fnvlist_add_string(nvl, ZPOOL_HIST_CMD, msg); fnvlist_add_string(nvl, ZPOOL_HIST_CMD, msg);
err = spa_history_log_nvl(spa, nvl); err = spa_history_log_nvl(spa, nvl);
@ -316,7 +314,7 @@ spa_history_log_nvl(spa_t *spa, nvlist_t *nvl)
return (err); return (err);
} }
VERIFY0(nvlist_dup(nvl, &nvarg, KM_PUSHPAGE)); VERIFY0(nvlist_dup(nvl, &nvarg, KM_SLEEP));
if (spa_history_zone() != NULL) { if (spa_history_zone() != NULL) {
fnvlist_add_string(nvarg, ZPOOL_HIST_ZONE, fnvlist_add_string(nvarg, ZPOOL_HIST_ZONE,
spa_history_zone()); spa_history_zone());
@ -453,7 +451,7 @@ log_internal(nvlist_t *nvl, const char *operation, spa_t *spa,
va_copy(adx1, adx); va_copy(adx1, adx);
size = vsnprintf(NULL, 0, fmt, adx1) + 1; size = vsnprintf(NULL, 0, fmt, adx1) + 1;
msg = kmem_alloc(size, KM_PUSHPAGE); msg = kmem_alloc(size, KM_SLEEP);
va_end(adx1); va_end(adx1);
va_copy(adx1, adx); va_copy(adx1, adx);
(void) vsprintf(msg, fmt, adx1); (void) vsprintf(msg, fmt, adx1);
@ -479,7 +477,6 @@ spa_history_log_internal(spa_t *spa, const char *operation,
{ {
dmu_tx_t *htx = tx; dmu_tx_t *htx = tx;
va_list adx; va_list adx;
nvlist_t *nvl;
/* create a tx if we didn't get one */ /* create a tx if we didn't get one */
if (tx == NULL) { if (tx == NULL) {
@ -491,8 +488,7 @@ spa_history_log_internal(spa_t *spa, const char *operation,
} }
va_start(adx, fmt); va_start(adx, fmt);
VERIFY0(nvlist_alloc(&nvl, NV_UNIQUE_NAME, KM_PUSHPAGE)); log_internal(fnvlist_alloc(), operation, spa, htx, fmt, adx);
log_internal(nvl, operation, spa, htx, fmt, adx);
va_end(adx); va_end(adx);
/* if we didn't get a tx from the caller, commit the one we made */ /* if we didn't get a tx from the caller, commit the one we made */
@ -506,12 +502,11 @@ spa_history_log_internal_ds(dsl_dataset_t *ds, const char *operation,
{ {
va_list adx; va_list adx;
char namebuf[MAXNAMELEN]; char namebuf[MAXNAMELEN];
nvlist_t *nvl; nvlist_t *nvl = fnvlist_alloc();
ASSERT(tx != NULL); ASSERT(tx != NULL);
dsl_dataset_name(ds, namebuf); dsl_dataset_name(ds, namebuf);
VERIFY0(nvlist_alloc(&nvl, NV_UNIQUE_NAME, KM_PUSHPAGE));
fnvlist_add_string(nvl, ZPOOL_HIST_DSNAME, namebuf); fnvlist_add_string(nvl, ZPOOL_HIST_DSNAME, namebuf);
fnvlist_add_uint64(nvl, ZPOOL_HIST_DSID, ds->ds_object); fnvlist_add_uint64(nvl, ZPOOL_HIST_DSID, ds->ds_object);
@ -526,12 +521,11 @@ spa_history_log_internal_dd(dsl_dir_t *dd, const char *operation,
{ {
va_list adx; va_list adx;
char namebuf[MAXNAMELEN]; char namebuf[MAXNAMELEN];
nvlist_t *nvl; nvlist_t *nvl = fnvlist_alloc();
ASSERT(tx != NULL); ASSERT(tx != NULL);
dsl_dir_name(dd, namebuf); dsl_dir_name(dd, namebuf);
VERIFY0(nvlist_alloc(&nvl, NV_UNIQUE_NAME, KM_PUSHPAGE));
fnvlist_add_string(nvl, ZPOOL_HIST_DSNAME, namebuf); fnvlist_add_string(nvl, ZPOOL_HIST_DSNAME, namebuf);
fnvlist_add_uint64(nvl, ZPOOL_HIST_DSID, fnvlist_add_uint64(nvl, ZPOOL_HIST_DSID,
dd->dd_phys->dd_head_dataset_obj); dd->dd_phys->dd_head_dataset_obj);

View File

@ -501,7 +501,7 @@ spa_deadman(void *arg)
vdev_deadman(spa->spa_root_vdev); vdev_deadman(spa->spa_root_vdev);
spa->spa_deadman_tqid = taskq_dispatch_delay(system_taskq, spa->spa_deadman_tqid = taskq_dispatch_delay(system_taskq,
spa_deadman, spa, TQ_PUSHPAGE, ddi_get_lbolt() + spa_deadman, spa, KM_SLEEP, ddi_get_lbolt() +
NSEC_TO_TICK(spa->spa_deadman_synctime)); NSEC_TO_TICK(spa->spa_deadman_synctime));
} }
@ -520,7 +520,7 @@ spa_add(const char *name, nvlist_t *config, const char *altroot)
ASSERT(MUTEX_HELD(&spa_namespace_lock)); ASSERT(MUTEX_HELD(&spa_namespace_lock));
spa = kmem_zalloc(sizeof (spa_t), KM_PUSHPAGE); spa = kmem_zalloc(sizeof (spa_t), KM_SLEEP);
mutex_init(&spa->spa_async_lock, NULL, MUTEX_DEFAULT, NULL); mutex_init(&spa->spa_async_lock, NULL, MUTEX_DEFAULT, NULL);
mutex_init(&spa->spa_errlist_lock, NULL, MUTEX_DEFAULT, NULL); mutex_init(&spa->spa_errlist_lock, NULL, MUTEX_DEFAULT, NULL);
@ -570,12 +570,12 @@ spa_add(const char *name, nvlist_t *config, const char *altroot)
list_create(&spa->spa_config_list, sizeof (spa_config_dirent_t), list_create(&spa->spa_config_list, sizeof (spa_config_dirent_t),
offsetof(spa_config_dirent_t, scd_link)); offsetof(spa_config_dirent_t, scd_link));
dp = kmem_zalloc(sizeof (spa_config_dirent_t), KM_PUSHPAGE); dp = kmem_zalloc(sizeof (spa_config_dirent_t), KM_SLEEP);
dp->scd_path = altroot ? NULL : spa_strdup(spa_config_path); dp->scd_path = altroot ? NULL : spa_strdup(spa_config_path);
list_insert_head(&spa->spa_config_list, dp); list_insert_head(&spa->spa_config_list, dp);
VERIFY(nvlist_alloc(&spa->spa_load_info, NV_UNIQUE_NAME, VERIFY(nvlist_alloc(&spa->spa_load_info, NV_UNIQUE_NAME,
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
if (config != NULL) { if (config != NULL) {
nvlist_t *features; nvlist_t *features;
@ -591,7 +591,7 @@ spa_add(const char *name, nvlist_t *config, const char *altroot)
if (spa->spa_label_features == NULL) { if (spa->spa_label_features == NULL) {
VERIFY(nvlist_alloc(&spa->spa_label_features, NV_UNIQUE_NAME, VERIFY(nvlist_alloc(&spa->spa_label_features, NV_UNIQUE_NAME,
KM_PUSHPAGE) == 0); KM_SLEEP) == 0);
} }
spa->spa_debug = ((zfs_flags & ZFS_DEBUG_SPA) != 0); spa->spa_debug = ((zfs_flags & ZFS_DEBUG_SPA) != 0);
@ -772,7 +772,7 @@ spa_aux_add(vdev_t *vd, avl_tree_t *avl)
if ((aux = avl_find(avl, &search, &where)) != NULL) { if ((aux = avl_find(avl, &search, &where)) != NULL) {
aux->aux_count++; aux->aux_count++;
} else { } else {
aux = kmem_zalloc(sizeof (spa_aux_t), KM_PUSHPAGE); aux = kmem_zalloc(sizeof (spa_aux_t), KM_SLEEP);
aux->aux_guid = vd->vdev_guid; aux->aux_guid = vd->vdev_guid;
aux->aux_count = 1; aux->aux_count = 1;
avl_insert(avl, aux, where); avl_insert(avl, aux, where);
@ -1280,7 +1280,7 @@ spa_strdup(const char *s)
char *new; char *new;
len = strlen(s); len = strlen(s);
new = kmem_alloc(len + 1, KM_PUSHPAGE); new = kmem_alloc(len + 1, KM_SLEEP);
bcopy(s, new, len); bcopy(s, new, len);
new[len] = '\0'; new[len] = '\0';

View File

@ -203,7 +203,7 @@ spa_read_history_add(spa_t *spa, const zbookmark_phys_t *zb, uint32_t aflags)
if (zfs_read_history_hits == 0 && (aflags & ARC_CACHED)) if (zfs_read_history_hits == 0 && (aflags & ARC_CACHED))
return; return;
srh = kmem_zalloc(sizeof (spa_read_history_t), KM_PUSHPAGE); srh = kmem_zalloc(sizeof (spa_read_history_t), KM_SLEEP);
strlcpy(srh->comm, getcomm(), sizeof (srh->comm)); strlcpy(srh->comm, getcomm(), sizeof (srh->comm));
srh->start = gethrtime(); srh->start = gethrtime();
srh->objset = zb->zb_objset; srh->objset = zb->zb_objset;
@ -423,7 +423,7 @@ spa_txg_history_add(spa_t *spa, uint64_t txg, hrtime_t birth_time)
if (zfs_txg_history == 0 && ssh->size == 0) if (zfs_txg_history == 0 && ssh->size == 0)
return; return;
sth = kmem_zalloc(sizeof (spa_txg_history_t), KM_PUSHPAGE); sth = kmem_zalloc(sizeof (spa_txg_history_t), KM_SLEEP);
sth->txg = txg; sth->txg = txg;
sth->state = TXG_STATE_OPEN; sth->state = TXG_STATE_OPEN;
sth->times[TXG_STATE_BIRTH] = birth_time; sth->times[TXG_STATE_BIRTH] = birth_time;

View File

@ -364,7 +364,7 @@ space_map_open(space_map_t **smp, objset_t *os, uint64_t object,
ASSERT(os != NULL); ASSERT(os != NULL);
ASSERT(object != 0); ASSERT(object != 0);
sm = kmem_alloc(sizeof (space_map_t), KM_PUSHPAGE); sm = kmem_alloc(sizeof (space_map_t), KM_SLEEP);
sm->sm_start = start; sm->sm_start = start;
sm->sm_size = size; sm->sm_size = size;

View File

@ -94,7 +94,7 @@ space_reftree_add_node(avl_tree_t *t, uint64_t offset, int64_t refcnt)
{ {
space_ref_t *sr; space_ref_t *sr;
sr = kmem_alloc(sizeof (*sr), KM_PUSHPAGE); sr = kmem_alloc(sizeof (*sr), KM_SLEEP);
sr->sr_offset = offset; sr->sr_offset = offset;
sr->sr_refcnt = refcnt; sr->sr_refcnt = refcnt;

View File

@ -449,7 +449,7 @@ txg_dispatch_callbacks(dsl_pool_t *dp, uint64_t txg)
TASKQ_THREADS_CPU_PCT | TASKQ_PREPOPULATE); TASKQ_THREADS_CPU_PCT | TASKQ_PREPOPULATE);
} }
cb_list = kmem_alloc(sizeof (list_t), KM_PUSHPAGE); cb_list = kmem_alloc(sizeof (list_t), KM_SLEEP);
list_create(cb_list, sizeof (dmu_tx_callback_t), list_create(cb_list, sizeof (dmu_tx_callback_t),
offsetof(dmu_tx_callback_t, dcb_node)); offsetof(dmu_tx_callback_t, dcb_node));
@ -486,8 +486,8 @@ txg_sync_thread(dsl_pool_t *dp)
(void) spl_fstrans_mark(); (void) spl_fstrans_mark();
txg_thread_enter(tx, &cpr); txg_thread_enter(tx, &cpr);
vs1 = kmem_alloc(sizeof (vdev_stat_t), KM_PUSHPAGE); vs1 = kmem_alloc(sizeof (vdev_stat_t), KM_SLEEP);
vs2 = kmem_alloc(sizeof (vdev_stat_t), KM_PUSHPAGE); vs2 = kmem_alloc(sizeof (vdev_stat_t), KM_SLEEP);
start = delta = 0; start = delta = 0;
for (;;) { for (;;) {

View File

@ -79,7 +79,7 @@ uint64_t
unique_insert(uint64_t value) unique_insert(uint64_t value)
{ {
avl_index_t idx; avl_index_t idx;
unique_t *un = kmem_alloc(sizeof (unique_t), KM_PUSHPAGE); unique_t *un = kmem_alloc(sizeof (unique_t), KM_SLEEP);
un->un_value = value; un->un_value = value;

View File

@ -200,7 +200,7 @@ vdev_add_child(vdev_t *pvd, vdev_t *cvd)
pvd->vdev_children = MAX(pvd->vdev_children, id + 1); pvd->vdev_children = MAX(pvd->vdev_children, id + 1);
newsize = pvd->vdev_children * sizeof (vdev_t *); newsize = pvd->vdev_children * sizeof (vdev_t *);
newchild = kmem_alloc(newsize, KM_PUSHPAGE); newchild = kmem_alloc(newsize, KM_SLEEP);
if (pvd->vdev_child != NULL) { if (pvd->vdev_child != NULL) {
bcopy(pvd->vdev_child, newchild, oldsize); bcopy(pvd->vdev_child, newchild, oldsize);
kmem_free(pvd->vdev_child, oldsize); kmem_free(pvd->vdev_child, oldsize);
@ -270,7 +270,7 @@ vdev_compact_children(vdev_t *pvd)
if (pvd->vdev_child[c]) if (pvd->vdev_child[c])
newc++; newc++;
newchild = kmem_zalloc(newc * sizeof (vdev_t *), KM_PUSHPAGE); newchild = kmem_zalloc(newc * sizeof (vdev_t *), KM_SLEEP);
for (c = newc = 0; c < oldc; c++) { for (c = newc = 0; c < oldc; c++) {
if ((cvd = pvd->vdev_child[c]) != NULL) { if ((cvd = pvd->vdev_child[c]) != NULL) {
@ -293,7 +293,7 @@ vdev_alloc_common(spa_t *spa, uint_t id, uint64_t guid, vdev_ops_t *ops)
vdev_t *vd; vdev_t *vd;
int t; int t;
vd = kmem_zalloc(sizeof (vdev_t), KM_PUSHPAGE); vd = kmem_zalloc(sizeof (vdev_t), KM_SLEEP);
if (spa->spa_root_vdev == NULL) { if (spa->spa_root_vdev == NULL) {
ASSERT(ops == &vdev_root_ops); ASSERT(ops == &vdev_root_ops);
@ -856,7 +856,7 @@ vdev_metaslab_init(vdev_t *vd, uint64_t txg)
ASSERT(oldc <= newc); ASSERT(oldc <= newc);
mspp = kmem_zalloc(newc * sizeof (*mspp), KM_PUSHPAGE); mspp = kmem_zalloc(newc * sizeof (*mspp), KM_SLEEP);
if (oldc != 0) { if (oldc != 0) {
bcopy(vd->vdev_ms, mspp, oldc * sizeof (*mspp)); bcopy(vd->vdev_ms, mspp, oldc * sizeof (*mspp));
@ -1011,7 +1011,7 @@ vdev_probe(vdev_t *vd, zio_t *zio)
mutex_enter(&vd->vdev_probe_lock); mutex_enter(&vd->vdev_probe_lock);
if ((pio = vd->vdev_probe_zio) == NULL) { if ((pio = vd->vdev_probe_zio) == NULL) {
vps = kmem_zalloc(sizeof (*vps), KM_PUSHPAGE); vps = kmem_zalloc(sizeof (*vps), KM_SLEEP);
vps->vps_flags = ZIO_FLAG_CANFAIL | ZIO_FLAG_PROBE | vps->vps_flags = ZIO_FLAG_CANFAIL | ZIO_FLAG_PROBE |
ZIO_FLAG_DONT_CACHE | ZIO_FLAG_DONT_AGGREGATE | ZIO_FLAG_DONT_CACHE | ZIO_FLAG_DONT_AGGREGATE |

View File

@ -180,7 +180,7 @@ vdev_cache_allocate(zio_t *zio)
vdev_cache_evict(vc, ve); vdev_cache_evict(vc, ve);
} }
ve = kmem_zalloc(sizeof (vdev_cache_entry_t), KM_PUSHPAGE); ve = kmem_zalloc(sizeof (vdev_cache_entry_t), KM_SLEEP);
ve->ve_offset = offset; ve->ve_offset = offset;
ve->ve_lastused = ddi_get_lbolt(); ve->ve_lastused = ddi_get_lbolt();
ve->ve_data = zio_buf_alloc(VCBS); ve->ve_data = zio_buf_alloc(VCBS);
@ -277,7 +277,7 @@ vdev_cache_read(zio_t *zio)
mutex_enter(&vc->vc_lock); mutex_enter(&vc->vc_lock);
ve_search = kmem_alloc(sizeof (vdev_cache_entry_t), KM_PUSHPAGE); ve_search = kmem_alloc(sizeof (vdev_cache_entry_t), KM_SLEEP);
ve_search->ve_offset = cache_offset; ve_search->ve_offset = cache_offset;
ve = avl_find(&vc->vc_offset_tree, ve_search, NULL); ve = avl_find(&vc->vc_offset_tree, ve_search, NULL);
kmem_free(ve_search, sizeof (vdev_cache_entry_t)); kmem_free(ve_search, sizeof (vdev_cache_entry_t));

View File

@ -262,7 +262,7 @@ vdev_disk_open(vdev_t *v, uint64_t *psize, uint64_t *max_psize,
goto skip_open; goto skip_open;
} }
vd = kmem_zalloc(sizeof (vdev_disk_t), KM_PUSHPAGE); vd = kmem_zalloc(sizeof (vdev_disk_t), KM_SLEEP);
if (vd == NULL) if (vd == NULL)
return (ENOMEM); return (ENOMEM);
@ -339,7 +339,7 @@ vdev_disk_dio_alloc(int bio_count)
int i; int i;
dr = kmem_zalloc(sizeof (dio_request_t) + dr = kmem_zalloc(sizeof (dio_request_t) +
sizeof (struct bio *) * bio_count, KM_PUSHPAGE); sizeof (struct bio *) * bio_count, KM_SLEEP);
if (dr) { if (dr) {
init_completion(&dr->dr_comp); init_completion(&dr->dr_comp);
atomic_set(&dr->dr_ref, 0); atomic_set(&dr->dr_ref, 0);
@ -807,7 +807,7 @@ vdev_disk_read_rootlabel(char *devpath, char *devid, nvlist_t **config)
} }
size = P2ALIGN_TYPED(s, sizeof (vdev_label_t), uint64_t); size = P2ALIGN_TYPED(s, sizeof (vdev_label_t), uint64_t);
label = vmem_alloc(sizeof (vdev_label_t), KM_PUSHPAGE); label = vmem_alloc(sizeof (vdev_label_t), KM_SLEEP);
for (i = 0; i < VDEV_LABELS; i++) { for (i = 0; i < VDEV_LABELS; i++) {
uint64_t offset, state, txg = 0; uint64_t offset, state, txg = 0;

View File

@ -77,7 +77,7 @@ vdev_file_open(vdev_t *vd, uint64_t *psize, uint64_t *max_psize,
goto skip_open; goto skip_open;
} }
vf = vd->vdev_tsd = kmem_zalloc(sizeof (vdev_file_t), KM_PUSHPAGE); vf = vd->vdev_tsd = kmem_zalloc(sizeof (vdev_file_t), KM_SLEEP);
/* /*
* We always open the files from the root of the global zone, even if * We always open the files from the root of the global zone, even if
@ -215,7 +215,7 @@ vdev_file_io_start(zio_t *zio)
} }
VERIFY3U(taskq_dispatch(vdev_file_taskq, vdev_file_io_strategy, zio, VERIFY3U(taskq_dispatch(vdev_file_taskq, vdev_file_io_strategy, zio,
TQ_PUSHPAGE), !=, 0); TQ_SLEEP), !=, 0);
return (ZIO_PIPELINE_STOP); return (ZIO_PIPELINE_STOP);
} }

View File

@ -216,7 +216,7 @@ vdev_config_generate(spa_t *spa, vdev_t *vd, boolean_t getstats,
{ {
nvlist_t *nv = NULL; nvlist_t *nv = NULL;
VERIFY(nvlist_alloc(&nv, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); nv = fnvlist_alloc();
fnvlist_add_string(nv, ZPOOL_CONFIG_TYPE, vd->vdev_ops->vdev_op_type); fnvlist_add_string(nv, ZPOOL_CONFIG_TYPE, vd->vdev_ops->vdev_op_type);
if (!(flags & (VDEV_CONFIG_SPARE | VDEV_CONFIG_L2CACHE))) if (!(flags & (VDEV_CONFIG_SPARE | VDEV_CONFIG_L2CACHE)))
@ -314,7 +314,7 @@ vdev_config_generate(spa_t *spa, vdev_t *vd, boolean_t getstats,
ASSERT(!vd->vdev_ishole); ASSERT(!vd->vdev_ishole);
child = kmem_alloc(vd->vdev_children * sizeof (nvlist_t *), child = kmem_alloc(vd->vdev_children * sizeof (nvlist_t *),
KM_PUSHPAGE); KM_SLEEP);
for (c = 0, idx = 0; c < vd->vdev_children; c++) { for (c = 0, idx = 0; c < vd->vdev_children; c++) {
vdev_t *cvd = vd->vdev_child[c]; vdev_t *cvd = vd->vdev_child[c];
@ -396,7 +396,7 @@ vdev_top_config_generate(spa_t *spa, nvlist_t *config)
uint64_t *array; uint64_t *array;
uint_t c, idx; uint_t c, idx;
array = kmem_alloc(rvd->vdev_children * sizeof (uint64_t), KM_PUSHPAGE); array = kmem_alloc(rvd->vdev_children * sizeof (uint64_t), KM_SLEEP);
for (c = 0, idx = 0; c < rvd->vdev_children; c++) { for (c = 0, idx = 0; c < rvd->vdev_children; c++) {
vdev_t *tvd = rvd->vdev_child[c]; vdev_t *tvd = rvd->vdev_child[c];
@ -728,7 +728,7 @@ vdev_label_init(vdev_t *vd, uint64_t crtxg, vdev_labeltype_t reason)
* active hot spare (in which case we want to revert the * active hot spare (in which case we want to revert the
* labels). * labels).
*/ */
VERIFY(nvlist_alloc(&label, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); VERIFY(nvlist_alloc(&label, NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_uint64(label, ZPOOL_CONFIG_VERSION, VERIFY(nvlist_add_uint64(label, ZPOOL_CONFIG_VERSION,
spa_version(spa)) == 0); spa_version(spa)) == 0);
@ -741,7 +741,7 @@ vdev_label_init(vdev_t *vd, uint64_t crtxg, vdev_labeltype_t reason)
/* /*
* For level 2 ARC devices, add a special label. * For level 2 ARC devices, add a special label.
*/ */
VERIFY(nvlist_alloc(&label, NV_UNIQUE_NAME, KM_PUSHPAGE) == 0); VERIFY(nvlist_alloc(&label, NV_UNIQUE_NAME, KM_SLEEP) == 0);
VERIFY(nvlist_add_uint64(label, ZPOOL_CONFIG_VERSION, VERIFY(nvlist_add_uint64(label, ZPOOL_CONFIG_VERSION,
spa_version(spa)) == 0); spa_version(spa)) == 0);
@ -768,7 +768,7 @@ vdev_label_init(vdev_t *vd, uint64_t crtxg, vdev_labeltype_t reason)
buf = vp->vp_nvlist; buf = vp->vp_nvlist;
buflen = sizeof (vp->vp_nvlist); buflen = sizeof (vp->vp_nvlist);
error = nvlist_pack(label, &buf, &buflen, NV_ENCODE_XDR, KM_PUSHPAGE); error = nvlist_pack(label, &buf, &buflen, NV_ENCODE_XDR, KM_SLEEP);
if (error != 0) { if (error != 0) {
nvlist_free(label); nvlist_free(label);
zio_buf_free(vp, sizeof (vdev_phys_t)); zio_buf_free(vp, sizeof (vdev_phys_t));
@ -1118,7 +1118,7 @@ vdev_label_sync(zio_t *zio, vdev_t *vd, int l, uint64_t txg, int flags)
buf = vp->vp_nvlist; buf = vp->vp_nvlist;
buflen = sizeof (vp->vp_nvlist); buflen = sizeof (vp->vp_nvlist);
if (!nvlist_pack(label, &buf, &buflen, NV_ENCODE_XDR, KM_PUSHPAGE)) { if (!nvlist_pack(label, &buf, &buflen, NV_ENCODE_XDR, KM_SLEEP)) {
for (; l < VDEV_LABELS; l += 2) { for (; l < VDEV_LABELS; l += 2) {
vdev_label_write(zio, vd, l, vp, vdev_label_write(zio, vd, l, vp,
offsetof(vdev_label_t, vl_vdev_phys), offsetof(vdev_label_t, vl_vdev_phys),
@ -1151,7 +1151,7 @@ vdev_label_sync_list(spa_t *spa, int l, uint64_t txg, int flags)
ASSERT(!vd->vdev_ishole); ASSERT(!vd->vdev_ishole);
good_writes = kmem_zalloc(sizeof (uint64_t), KM_PUSHPAGE); good_writes = kmem_zalloc(sizeof (uint64_t), KM_SLEEP);
vio = zio_null(zio, spa, NULL, vio = zio_null(zio, spa, NULL,
(vd->vdev_islog || vd->vdev_aux != NULL) ? (vd->vdev_islog || vd->vdev_aux != NULL) ?
vdev_label_sync_ignore_done : vdev_label_sync_top_done, vdev_label_sync_ignore_done : vdev_label_sync_top_done,

View File

@ -111,7 +111,7 @@ vdev_mirror_map_alloc(zio_t *zio)
c = BP_GET_NDVAS(zio->io_bp); c = BP_GET_NDVAS(zio->io_bp);
mm = kmem_zalloc(offsetof(mirror_map_t, mm_child[c]), mm = kmem_zalloc(offsetof(mirror_map_t, mm_child[c]),
KM_PUSHPAGE); KM_SLEEP);
mm->mm_children = c; mm->mm_children = c;
mm->mm_replacing = B_FALSE; mm->mm_replacing = B_FALSE;
mm->mm_preferred = spa_get_random(c); mm->mm_preferred = spa_get_random(c);
@ -142,7 +142,7 @@ vdev_mirror_map_alloc(zio_t *zio)
c = vd->vdev_children; c = vd->vdev_children;
mm = kmem_zalloc(offsetof(mirror_map_t, mm_child[c]), mm = kmem_zalloc(offsetof(mirror_map_t, mm_child[c]),
KM_PUSHPAGE); KM_SLEEP);
mm->mm_children = c; mm->mm_children = c;
mm->mm_replacing = (vd->vdev_ops == &vdev_replacing_ops || mm->mm_replacing = (vd->vdev_ops == &vdev_replacing_ops ||
vd->vdev_ops == &vdev_spare_ops); vd->vdev_ops == &vdev_spare_ops);

View File

@ -486,7 +486,7 @@ vdev_raidz_map_alloc(zio_t *zio, uint64_t unit_shift, uint64_t dcols,
ASSERT3U(acols, <=, scols); ASSERT3U(acols, <=, scols);
rm = kmem_alloc(offsetof(raidz_map_t, rm_col[scols]), KM_PUSHPAGE); rm = kmem_alloc(offsetof(raidz_map_t, rm_col[scols]), KM_SLEEP);
rm->rm_cols = acols; rm->rm_cols = acols;
rm->rm_scols = scols; rm->rm_scols = scols;
@ -1227,7 +1227,7 @@ vdev_raidz_matrix_reconstruct(raidz_map_t *rm, int n, int nmissing,
size_t psize; size_t psize;
psize = sizeof (invlog[0][0]) * n * nmissing; psize = sizeof (invlog[0][0]) * n * nmissing;
p = kmem_alloc(psize, KM_PUSHPAGE); p = kmem_alloc(psize, KM_SLEEP);
for (pp = p, i = 0; i < nmissing; i++) { for (pp = p, i = 0; i < nmissing; i++) {
invlog[i] = pp; invlog[i] = pp;
@ -1344,7 +1344,7 @@ vdev_raidz_reconstruct_general(raidz_map_t *rm, int *tgts, int ntgts)
psize = (sizeof (rows[0][0]) + sizeof (invrows[0][0])) * psize = (sizeof (rows[0][0]) + sizeof (invrows[0][0])) *
nmissing_rows * n + sizeof (used[0]) * n; nmissing_rows * n + sizeof (used[0]) * n;
p = kmem_alloc(psize, KM_PUSHPAGE); p = kmem_alloc(psize, KM_SLEEP);
for (pp = p, i = 0; i < nmissing_rows; i++) { for (pp = p, i = 0; i < nmissing_rows; i++) {
rows[i] = pp; rows[i] = pp;

View File

@ -115,7 +115,7 @@ fzap_upgrade(zap_t *zap, dmu_tx_t *tx, zap_flags_t flags)
1<<FZAP_BLOCK_SHIFT(zap), FTAG, &db, DMU_READ_NO_PREFETCH)); 1<<FZAP_BLOCK_SHIFT(zap), FTAG, &db, DMU_READ_NO_PREFETCH));
dmu_buf_will_dirty(db, tx); dmu_buf_will_dirty(db, tx);
l = kmem_zalloc(sizeof (zap_leaf_t), KM_PUSHPAGE); l = kmem_zalloc(sizeof (zap_leaf_t), KM_SLEEP);
l->l_dbuf = db; l->l_dbuf = db;
l->l_phys = db->db_data; l->l_phys = db->db_data;
@ -392,7 +392,7 @@ static zap_leaf_t *
zap_create_leaf(zap_t *zap, dmu_tx_t *tx) zap_create_leaf(zap_t *zap, dmu_tx_t *tx)
{ {
void *winner; void *winner;
zap_leaf_t *l = kmem_alloc(sizeof (zap_leaf_t), KM_PUSHPAGE); zap_leaf_t *l = kmem_alloc(sizeof (zap_leaf_t), KM_SLEEP);
ASSERT(RW_WRITE_HELD(&zap->zap_rwlock)); ASSERT(RW_WRITE_HELD(&zap->zap_rwlock));
@ -454,7 +454,7 @@ zap_open_leaf(uint64_t blkid, dmu_buf_t *db)
ASSERT(blkid != 0); ASSERT(blkid != 0);
l = kmem_alloc(sizeof (zap_leaf_t), KM_PUSHPAGE); l = kmem_alloc(sizeof (zap_leaf_t), KM_SLEEP);
rw_init(&l->l_rwlock, NULL, RW_DEFAULT, NULL); rw_init(&l->l_rwlock, NULL, RW_DEFAULT, NULL);
rw_enter(&l->l_rwlock, RW_WRITER); rw_enter(&l->l_rwlock, RW_WRITER);
l->l_blkid = blkid; l->l_blkid = blkid;
@ -972,7 +972,7 @@ zap_value_search(objset_t *os, uint64_t zapobj, uint64_t value, uint64_t mask,
if (mask == 0) if (mask == 0)
mask = -1ULL; mask = -1ULL;
za = kmem_alloc(sizeof (zap_attribute_t), KM_PUSHPAGE); za = kmem_alloc(sizeof (zap_attribute_t), KM_SLEEP);
for (zap_cursor_init(&zc, os, zapobj); for (zap_cursor_init(&zc, os, zapobj);
(err = zap_cursor_retrieve(&zc, za)) == 0; (err = zap_cursor_retrieve(&zc, za)) == 0;
zap_cursor_advance(&zc)) { zap_cursor_advance(&zc)) {

View File

@ -342,7 +342,7 @@ zap_leaf_array_match(zap_leaf_t *l, zap_name_t *zn,
ASSERT(zn->zn_key_intlen == sizeof (*thiskey)); ASSERT(zn->zn_key_intlen == sizeof (*thiskey));
thiskey = kmem_alloc(array_numints * sizeof (*thiskey), thiskey = kmem_alloc(array_numints * sizeof (*thiskey),
KM_PUSHPAGE); KM_SLEEP);
zap_leaf_array_read(l, chunk, sizeof (*thiskey), array_numints, zap_leaf_array_read(l, chunk, sizeof (*thiskey), array_numints,
sizeof (*thiskey), array_numints, thiskey); sizeof (*thiskey), array_numints, thiskey);
@ -354,7 +354,7 @@ zap_leaf_array_match(zap_leaf_t *l, zap_name_t *zn,
ASSERT(zn->zn_key_intlen == 1); ASSERT(zn->zn_key_intlen == 1);
if (zn->zn_matchtype == MT_FIRST) { if (zn->zn_matchtype == MT_FIRST) {
char *thisname = kmem_alloc(array_numints, KM_PUSHPAGE); char *thisname = kmem_alloc(array_numints, KM_SLEEP);
boolean_t match; boolean_t match;
zap_leaf_array_read(l, chunk, sizeof (char), array_numints, zap_leaf_array_read(l, chunk, sizeof (char), array_numints,

View File

@ -172,7 +172,7 @@ zap_name_free(zap_name_t *zn)
zap_name_t * zap_name_t *
zap_name_alloc(zap_t *zap, const char *key, matchtype_t mt) zap_name_alloc(zap_t *zap, const char *key, matchtype_t mt)
{ {
zap_name_t *zn = kmem_alloc(sizeof (zap_name_t), KM_PUSHPAGE); zap_name_t *zn = kmem_alloc(sizeof (zap_name_t), KM_SLEEP);
zn->zn_zap = zap; zn->zn_zap = zap;
zn->zn_key_intlen = sizeof (*key); zn->zn_key_intlen = sizeof (*key);
@ -202,7 +202,7 @@ zap_name_alloc(zap_t *zap, const char *key, matchtype_t mt)
zap_name_t * zap_name_t *
zap_name_alloc_uint64(zap_t *zap, const uint64_t *key, int numints) zap_name_alloc_uint64(zap_t *zap, const uint64_t *key, int numints)
{ {
zap_name_t *zn = kmem_alloc(sizeof (zap_name_t), KM_PUSHPAGE); zap_name_t *zn = kmem_alloc(sizeof (zap_name_t), KM_SLEEP);
ASSERT(zap->zap_normflags == 0); ASSERT(zap->zap_normflags == 0);
zn->zn_zap = zap; zn->zn_zap = zap;
@ -271,7 +271,7 @@ mze_insert(zap_t *zap, int chunkid, uint64_t hash)
ASSERT(zap->zap_ismicro); ASSERT(zap->zap_ismicro);
ASSERT(RW_WRITE_HELD(&zap->zap_rwlock)); ASSERT(RW_WRITE_HELD(&zap->zap_rwlock));
mze = kmem_alloc(sizeof (mzap_ent_t), KM_PUSHPAGE); mze = kmem_alloc(sizeof (mzap_ent_t), KM_SLEEP);
mze->mze_chunkid = chunkid; mze->mze_chunkid = chunkid;
mze->mze_hash = hash; mze->mze_hash = hash;
mze->mze_cd = MZE_PHYS(zap, mze)->mze_cd; mze->mze_cd = MZE_PHYS(zap, mze)->mze_cd;
@ -365,7 +365,7 @@ mzap_open(objset_t *os, uint64_t obj, dmu_buf_t *db)
ASSERT3U(MZAP_ENT_LEN, ==, sizeof (mzap_ent_phys_t)); ASSERT3U(MZAP_ENT_LEN, ==, sizeof (mzap_ent_phys_t));
zap = kmem_zalloc(sizeof (zap_t), KM_PUSHPAGE); zap = kmem_zalloc(sizeof (zap_t), KM_SLEEP);
rw_init(&zap->zap_rwlock, NULL, RW_DEFAULT, NULL); rw_init(&zap->zap_rwlock, NULL, RW_DEFAULT, NULL);
rw_enter(&zap->zap_rwlock, RW_WRITER); rw_enter(&zap->zap_rwlock, RW_WRITER);
zap->zap_objset = os; zap->zap_objset = os;

View File

@ -454,7 +454,7 @@ zfs_acl_alloc(int vers)
{ {
zfs_acl_t *aclp; zfs_acl_t *aclp;
aclp = kmem_zalloc(sizeof (zfs_acl_t), KM_PUSHPAGE); aclp = kmem_zalloc(sizeof (zfs_acl_t), KM_SLEEP);
list_create(&aclp->z_acl, sizeof (zfs_acl_node_t), list_create(&aclp->z_acl, sizeof (zfs_acl_node_t),
offsetof(zfs_acl_node_t, z_next)); offsetof(zfs_acl_node_t, z_next));
aclp->z_version = vers; aclp->z_version = vers;
@ -470,9 +470,9 @@ zfs_acl_node_alloc(size_t bytes)
{ {
zfs_acl_node_t *aclnode; zfs_acl_node_t *aclnode;
aclnode = kmem_zalloc(sizeof (zfs_acl_node_t), KM_PUSHPAGE); aclnode = kmem_zalloc(sizeof (zfs_acl_node_t), KM_SLEEP);
if (bytes) { if (bytes) {
aclnode->z_acldata = kmem_alloc(bytes, KM_PUSHPAGE); aclnode->z_acldata = kmem_alloc(bytes, KM_SLEEP);
aclnode->z_allocdata = aclnode->z_acldata; aclnode->z_allocdata = aclnode->z_acldata;
aclnode->z_allocsize = bytes; aclnode->z_allocsize = bytes;
aclnode->z_size = bytes; aclnode->z_size = bytes;

View File

@ -78,7 +78,7 @@ zfs_dbgmsg(const char *fmt, ...)
* There is one byte of string in sizeof (zfs_dbgmsg_t), used * There is one byte of string in sizeof (zfs_dbgmsg_t), used
* for the terminating null. * for the terminating null.
*/ */
zdm = kmem_alloc(sizeof (zfs_dbgmsg_t) + size, KM_PUSHPAGE); zdm = kmem_alloc(sizeof (zfs_dbgmsg_t) + size, KM_SLEEP);
zdm->zdm_timestamp = gethrestime_sec(); zdm->zdm_timestamp = gethrestime_sec();
va_start(adx, fmt); va_start(adx, fmt);

View File

@ -321,9 +321,9 @@ zfs_ereport_start(nvlist_t **ereport_out, nvlist_t **detector_out,
spare_count = spa->spa_spares.sav_count; spare_count = spa->spa_spares.sav_count;
spare_paths = kmem_zalloc(sizeof (char *) * spare_count, spare_paths = kmem_zalloc(sizeof (char *) * spare_count,
KM_PUSHPAGE); KM_SLEEP);
spare_guids = kmem_zalloc(sizeof (uint64_t) * spare_count, spare_guids = kmem_zalloc(sizeof (uint64_t) * spare_count,
KM_PUSHPAGE); KM_SLEEP);
for (i = 0; i < spare_count; i++) { for (i = 0; i < spare_count; i++) {
spare_vd = spa->spa_spares.sav_vdevs[i]; spare_vd = spa->spa_spares.sav_vdevs[i];
@ -583,7 +583,7 @@ annotate_ecksum(nvlist_t *ereport, zio_bad_cksum_t *info,
size_t offset = 0; size_t offset = 0;
ssize_t start = -1; ssize_t start = -1;
zfs_ecksum_info_t *eip = kmem_zalloc(sizeof (*eip), KM_PUSHPAGE); zfs_ecksum_info_t *eip = kmem_zalloc(sizeof (*eip), KM_SLEEP);
/* don't do any annotation for injected checksum errors */ /* don't do any annotation for injected checksum errors */
if (info != NULL && info->zbc_injected) if (info != NULL && info->zbc_injected)
@ -752,7 +752,7 @@ zfs_ereport_start_checksum(spa_t *spa, vdev_t *vd,
struct zio *zio, uint64_t offset, uint64_t length, void *arg, struct zio *zio, uint64_t offset, uint64_t length, void *arg,
zio_bad_cksum_t *info) zio_bad_cksum_t *info)
{ {
zio_cksum_report_t *report = kmem_zalloc(sizeof (*report), KM_PUSHPAGE); zio_cksum_report_t *report = kmem_zalloc(sizeof (*report), KM_SLEEP);
if (zio->io_vsd != NULL) if (zio->io_vsd != NULL)
zio->io_vsd_ops->vsd_cksum_report(zio, report, arg); zio->io_vsd_ops->vsd_cksum_report(zio, report, arg);
@ -761,7 +761,7 @@ zfs_ereport_start_checksum(spa_t *spa, vdev_t *vd,
/* copy the checksum failure information if it was provided */ /* copy the checksum failure information if it was provided */
if (info != NULL) { if (info != NULL) {
report->zcr_ckinfo = kmem_zalloc(sizeof (*info), KM_PUSHPAGE); report->zcr_ckinfo = kmem_zalloc(sizeof (*info), KM_SLEEP);
bcopy(info, report->zcr_ckinfo, sizeof (*info)); bcopy(info, report->zcr_ckinfo, sizeof (*info));
} }

View File

@ -254,7 +254,7 @@ __dprintf(const char *file, const char *func, int line, const char *fmt, ...)
{ {
const char *newfile; const char *newfile;
size_t size = 4096; size_t size = 4096;
char *buf = kmem_alloc(size, KM_PUSHPAGE); char *buf = kmem_alloc(size, KM_SLEEP);
char *nl; char *nl;
va_list adx; va_list adx;
@ -5827,7 +5827,7 @@ zfsdev_ioctl(struct file *filp, unsigned cmd, unsigned long arg)
} }
} }
VERIFY0(nvlist_alloc(&outnvl, NV_UNIQUE_NAME, KM_PUSHPAGE)); outnvl = fnvlist_alloc();
error = vec->zvec_func(zc->zc_name, innvl, outnvl); error = vec->zvec_func(zc->zc_name, innvl, outnvl);
if (error == 0 && vec->zvec_allow_log && if (error == 0 && vec->zvec_allow_log &&

View File

@ -156,7 +156,7 @@ zfs_onexit_add_cb(minor_t minor, void (*func)(void *), void *data,
if (error) if (error)
return (error); return (error);
ap = kmem_alloc(sizeof (zfs_onexit_action_node_t), KM_PUSHPAGE); ap = kmem_alloc(sizeof (zfs_onexit_action_node_t), KM_SLEEP);
list_link_init(&ap->za_link); list_link_init(&ap->za_link);
ap->za_func = func; ap->za_func = func;
ap->za_data = data; ap->za_data = data;

View File

@ -202,7 +202,7 @@ zfs_range_proxify(avl_tree_t *tree, rl_t *rl)
rl->r_cnt = 0; rl->r_cnt = 0;
/* create a proxy range lock */ /* create a proxy range lock */
proxy = kmem_alloc(sizeof (rl_t), KM_PUSHPAGE); proxy = kmem_alloc(sizeof (rl_t), KM_SLEEP);
proxy->r_off = rl->r_off; proxy->r_off = rl->r_off;
proxy->r_len = rl->r_len; proxy->r_len = rl->r_len;
proxy->r_cnt = 1; proxy->r_cnt = 1;
@ -231,7 +231,7 @@ zfs_range_split(avl_tree_t *tree, rl_t *rl, uint64_t off)
ASSERT(rl->r_read_wanted == B_FALSE); ASSERT(rl->r_read_wanted == B_FALSE);
/* create the rear proxy range lock */ /* create the rear proxy range lock */
rear = kmem_alloc(sizeof (rl_t), KM_PUSHPAGE); rear = kmem_alloc(sizeof (rl_t), KM_SLEEP);
rear->r_off = off; rear->r_off = off;
rear->r_len = rl->r_off + rl->r_len - off; rear->r_len = rl->r_off + rl->r_len - off;
rear->r_cnt = rl->r_cnt; rear->r_cnt = rl->r_cnt;
@ -256,7 +256,7 @@ zfs_range_new_proxy(avl_tree_t *tree, uint64_t off, uint64_t len)
rl_t *rl; rl_t *rl;
ASSERT(len); ASSERT(len);
rl = kmem_alloc(sizeof (rl_t), KM_PUSHPAGE); rl = kmem_alloc(sizeof (rl_t), KM_SLEEP);
rl->r_off = off; rl->r_off = off;
rl->r_len = len; rl->r_len = len;
rl->r_cnt = 1; rl->r_cnt = 1;
@ -429,7 +429,7 @@ zfs_range_lock(znode_t *zp, uint64_t off, uint64_t len, rl_type_t type)
ASSERT(type == RL_READER || type == RL_WRITER || type == RL_APPEND); ASSERT(type == RL_READER || type == RL_WRITER || type == RL_APPEND);
new = kmem_alloc(sizeof (rl_t), KM_PUSHPAGE); new = kmem_alloc(sizeof (rl_t), KM_SLEEP);
new->r_zp = zp; new->r_zp = zp;
new->r_off = off; new->r_off = off;
if (len + off < off) /* overflow */ if (len + off < off) /* overflow */

View File

@ -922,7 +922,7 @@ zfs_iput_async(struct inode *ip)
if (atomic_read(&ip->i_count) == 1) if (atomic_read(&ip->i_count) == 1)
taskq_dispatch(dsl_pool_iput_taskq(dmu_objset_pool(os)), taskq_dispatch(dsl_pool_iput_taskq(dmu_objset_pool(os)),
(task_func_t *)iput, ip, TQ_PUSHPAGE); (task_func_t *)iput, ip, TQ_SLEEP);
else else
iput(ip); iput(ip);
} }
@ -987,7 +987,7 @@ zfs_get_data(void *arg, lr_write_t *lr, char *buf, zio_t *zio)
return (SET_ERROR(ENOENT)); return (SET_ERROR(ENOENT));
} }
zgd = (zgd_t *)kmem_zalloc(sizeof (zgd_t), KM_PUSHPAGE); zgd = (zgd_t *)kmem_zalloc(sizeof (zgd_t), KM_SLEEP);
zgd->zgd_zilog = zsb->z_log; zgd->zgd_zilog = zsb->z_log;
zgd->zgd_private = zp; zgd->zgd_private = zp;

View File

@ -182,7 +182,7 @@ zfs_create_share_dir(zfs_sb_t *zsb, dmu_tx_t *tx)
vattr.va_uid = crgetuid(kcred); vattr.va_uid = crgetuid(kcred);
vattr.va_gid = crgetgid(kcred); vattr.va_gid = crgetgid(kcred);
sharezp = kmem_cache_alloc(znode_cache, KM_PUSHPAGE); sharezp = kmem_cache_alloc(znode_cache, KM_SLEEP);
sharezp->z_moved = 0; sharezp->z_moved = 0;
sharezp->z_unlinked = 0; sharezp->z_unlinked = 0;
sharezp->z_atime_dirty = 0; sharezp->z_atime_dirty = 0;
@ -256,7 +256,7 @@ zfs_inode_alloc(struct super_block *sb, struct inode **ip)
{ {
znode_t *zp; znode_t *zp;
zp = kmem_cache_alloc(znode_cache, KM_PUSHPAGE); zp = kmem_cache_alloc(znode_cache, KM_SLEEP);
*ip = ZTOI(zp); *ip = ZTOI(zp);
return (0); return (0);
@ -682,7 +682,7 @@ zfs_mknode(znode_t *dzp, vattr_t *vap, dmu_tx_t *tx, cred_t *cr,
* order for DMU_OT_ZNODE is critical since it needs to be constructed * order for DMU_OT_ZNODE is critical since it needs to be constructed
* in the old znode_phys_t format. Don't change this ordering * in the old znode_phys_t format. Don't change this ordering
*/ */
sa_attrs = kmem_alloc(sizeof (sa_bulk_attr_t) * ZPL_END, KM_PUSHPAGE); sa_attrs = kmem_alloc(sizeof (sa_bulk_attr_t) * ZPL_END, KM_SLEEP);
if (obj_type == DMU_OT_ZNODE) { if (obj_type == DMU_OT_ZNODE) {
SA_ADD_BULK_ATTR(sa_attrs, cnt, SA_ZPL_ATIME(zsb), SA_ADD_BULK_ATTR(sa_attrs, cnt, SA_ZPL_ATIME(zsb),
@ -1696,13 +1696,13 @@ zfs_create_fs(objset_t *os, cred_t *cr, nvlist_t *zplprops, dmu_tx_t *tx)
vattr.va_uid = crgetuid(cr); vattr.va_uid = crgetuid(cr);
vattr.va_gid = crgetgid(cr); vattr.va_gid = crgetgid(cr);
rootzp = kmem_cache_alloc(znode_cache, KM_PUSHPAGE); rootzp = kmem_cache_alloc(znode_cache, KM_SLEEP);
rootzp->z_moved = 0; rootzp->z_moved = 0;
rootzp->z_unlinked = 0; rootzp->z_unlinked = 0;
rootzp->z_atime_dirty = 0; rootzp->z_atime_dirty = 0;
rootzp->z_is_sa = USE_SA(version, os); rootzp->z_is_sa = USE_SA(version, os);
zsb = kmem_zalloc(sizeof (zfs_sb_t), KM_PUSHPAGE); zsb = kmem_zalloc(sizeof (zfs_sb_t), KM_SLEEP);
zsb->z_os = os; zsb->z_os = os;
zsb->z_parent = zsb; zsb->z_parent = zsb;
zsb->z_version = version; zsb->z_version = version;
@ -1710,7 +1710,7 @@ zfs_create_fs(objset_t *os, cred_t *cr, nvlist_t *zplprops, dmu_tx_t *tx)
zsb->z_use_sa = USE_SA(version, os); zsb->z_use_sa = USE_SA(version, os);
zsb->z_norm = norm; zsb->z_norm = norm;
sb = kmem_zalloc(sizeof (struct super_block), KM_PUSHPAGE); sb = kmem_zalloc(sizeof (struct super_block), KM_SLEEP);
sb->s_fs_info = zsb; sb->s_fs_info = zsb;
ZTOI(rootzp)->i_sb = sb; ZTOI(rootzp)->i_sb = sb;

View File

@ -172,7 +172,7 @@ zil_bp_tree_add(zilog_t *zilog, const blkptr_t *bp)
if (avl_find(t, dva, &where) != NULL) if (avl_find(t, dva, &where) != NULL)
return (SET_ERROR(EEXIST)); return (SET_ERROR(EEXIST));
zn = kmem_alloc(sizeof (zil_bp_node_t), KM_PUSHPAGE); zn = kmem_alloc(sizeof (zil_bp_node_t), KM_SLEEP);
zn->zn_dva = *dva; zn->zn_dva = *dva;
avl_insert(t, zn, where); avl_insert(t, zn, where);
@ -464,7 +464,7 @@ zil_alloc_lwb(zilog_t *zilog, blkptr_t *bp, uint64_t txg, boolean_t fastwrite)
{ {
lwb_t *lwb; lwb_t *lwb;
lwb = kmem_cache_alloc(zil_lwb_cache, KM_PUSHPAGE); lwb = kmem_cache_alloc(zil_lwb_cache, KM_SLEEP);
lwb->lwb_zilog = zilog; lwb->lwb_zilog = zilog;
lwb->lwb_blk = *bp; lwb->lwb_blk = *bp;
lwb->lwb_fastwrite = fastwrite; lwb->lwb_fastwrite = fastwrite;
@ -815,7 +815,7 @@ zil_add_block(zilog_t *zilog, const blkptr_t *bp)
for (i = 0; i < ndvas; i++) { for (i = 0; i < ndvas; i++) {
zvsearch.zv_vdev = DVA_GET_VDEV(&bp->blk_dva[i]); zvsearch.zv_vdev = DVA_GET_VDEV(&bp->blk_dva[i]);
if (avl_find(t, &zvsearch, &where) == NULL) { if (avl_find(t, &zvsearch, &where) == NULL) {
zv = kmem_alloc(sizeof (*zv), KM_PUSHPAGE); zv = kmem_alloc(sizeof (*zv), KM_SLEEP);
zv->zv_vdev = zvsearch.zv_vdev; zv->zv_vdev = zvsearch.zv_vdev;
avl_insert(t, zv, where); avl_insert(t, zv, where);
} }
@ -1192,7 +1192,7 @@ zil_itx_create(uint64_t txtype, size_t lrsize)
lrsize = P2ROUNDUP_TYPED(lrsize, sizeof (uint64_t), size_t); lrsize = P2ROUNDUP_TYPED(lrsize, sizeof (uint64_t), size_t);
itx = vmem_alloc(offsetof(itx_t, itx_lr) + lrsize, KM_PUSHPAGE); itx = vmem_alloc(offsetof(itx_t, itx_lr) + lrsize, KM_SLEEP);
itx->itx_lr.lrc_txtype = txtype; itx->itx_lr.lrc_txtype = txtype;
itx->itx_lr.lrc_reclen = lrsize; itx->itx_lr.lrc_reclen = lrsize;
itx->itx_sod = lrsize; /* if write & WR_NEED_COPY will be increased */ itx->itx_sod = lrsize; /* if write & WR_NEED_COPY will be increased */
@ -1359,7 +1359,7 @@ zil_itx_assign(zilog_t *zilog, itx_t *itx, dmu_tx_t *tx)
ASSERT(itxg->itxg_sod == 0); ASSERT(itxg->itxg_sod == 0);
itxg->itxg_txg = txg; itxg->itxg_txg = txg;
itxs = itxg->itxg_itxs = kmem_zalloc(sizeof (itxs_t), itxs = itxg->itxg_itxs = kmem_zalloc(sizeof (itxs_t),
KM_PUSHPAGE); KM_SLEEP);
list_create(&itxs->i_sync_list, sizeof (itx_t), list_create(&itxs->i_sync_list, sizeof (itx_t),
offsetof(itx_t, itx_node)); offsetof(itx_t, itx_node));
@ -1380,7 +1380,7 @@ zil_itx_assign(zilog_t *zilog, itx_t *itx, dmu_tx_t *tx)
ian = avl_find(t, &foid, &where); ian = avl_find(t, &foid, &where);
if (ian == NULL) { if (ian == NULL) {
ian = kmem_alloc(sizeof (itx_async_node_t), ian = kmem_alloc(sizeof (itx_async_node_t),
KM_PUSHPAGE); KM_SLEEP);
list_create(&ian->ia_list, sizeof (itx_t), list_create(&ian->ia_list, sizeof (itx_t),
offsetof(itx_t, itx_node)); offsetof(itx_t, itx_node));
ian->ia_foid = foid; ian->ia_foid = foid;
@ -1798,7 +1798,7 @@ zil_alloc(objset_t *os, zil_header_t *zh_phys)
zilog_t *zilog; zilog_t *zilog;
int i; int i;
zilog = kmem_zalloc(sizeof (zilog_t), KM_PUSHPAGE); zilog = kmem_zalloc(sizeof (zilog_t), KM_SLEEP);
zilog->zl_header = zh_phys; zilog->zl_header = zh_phys;
zilog->zl_os = os; zilog->zl_os = os;
@ -2204,7 +2204,7 @@ zil_replay(objset_t *os, void *arg, zil_replay_func_t replay_func[TX_MAX_TYPE])
zr.zr_replay = replay_func; zr.zr_replay = replay_func;
zr.zr_arg = arg; zr.zr_arg = arg;
zr.zr_byteswap = BP_SHOULD_BYTESWAP(&zh->zh_log); zr.zr_byteswap = BP_SHOULD_BYTESWAP(&zh->zh_log);
zr.zr_lr = vmem_alloc(2 * SPA_MAXBLOCKSIZE, KM_PUSHPAGE); zr.zr_lr = vmem_alloc(2 * SPA_MAXBLOCKSIZE, KM_SLEEP);
/* /*
* Wait for in-progress removes to sync before starting replay. * Wait for in-progress removes to sync before starting replay.

View File

@ -312,7 +312,7 @@ static void
zio_push_transform(zio_t *zio, void *data, uint64_t size, uint64_t bufsize, zio_push_transform(zio_t *zio, void *data, uint64_t size, uint64_t bufsize,
zio_transform_func_t *transform) zio_transform_func_t *transform)
{ {
zio_transform_t *zt = kmem_alloc(sizeof (zio_transform_t), KM_PUSHPAGE); zio_transform_t *zt = kmem_alloc(sizeof (zio_transform_t), KM_SLEEP);
zt->zt_orig_data = zio->io_data; zt->zt_orig_data = zio->io_data;
zt->zt_orig_size = zio->io_size; zt->zt_orig_size = zio->io_size;
@ -427,7 +427,7 @@ zio_unique_parent(zio_t *cio)
void void
zio_add_child(zio_t *pio, zio_t *cio) zio_add_child(zio_t *pio, zio_t *cio)
{ {
zio_link_t *zl = kmem_cache_alloc(zio_link_cache, KM_PUSHPAGE); zio_link_t *zl = kmem_cache_alloc(zio_link_cache, KM_SLEEP);
int w; int w;
/* /*
@ -551,7 +551,7 @@ zio_create(zio_t *pio, spa_t *spa, uint64_t txg, const blkptr_t *bp,
ASSERT(!bp || !(flags & ZIO_FLAG_CONFIG_WRITER)); ASSERT(!bp || !(flags & ZIO_FLAG_CONFIG_WRITER));
ASSERT(vd || stage == ZIO_STAGE_OPEN); ASSERT(vd || stage == ZIO_STAGE_OPEN);
zio = kmem_cache_alloc(zio_cache, KM_PUSHPAGE); zio = kmem_cache_alloc(zio_cache, KM_SLEEP);
if (vd != NULL) if (vd != NULL)
zio->io_child_type = ZIO_CHILD_VDEV; zio->io_child_type = ZIO_CHILD_VDEV;
@ -1757,7 +1757,7 @@ zio_gang_node_alloc(zio_gang_node_t **gnpp)
ASSERT(*gnpp == NULL); ASSERT(*gnpp == NULL);
gn = kmem_zalloc(sizeof (*gn), KM_PUSHPAGE); gn = kmem_zalloc(sizeof (*gn), KM_SLEEP);
gn->gn_gbh = zio_buf_alloc(SPA_GANGBLOCKSIZE); gn->gn_gbh = zio_buf_alloc(SPA_GANGBLOCKSIZE);
*gnpp = gn; *gnpp = gn;

View File

@ -824,7 +824,7 @@ zpl_get_acl(struct inode *ip, int type)
size = zpl_xattr_get(ip, name, NULL, 0); size = zpl_xattr_get(ip, name, NULL, 0);
if (size > 0) { if (size > 0) {
value = kmem_alloc(size, KM_PUSHPAGE); value = kmem_alloc(size, KM_SLEEP);
size = zpl_xattr_get(ip, name, value, size); size = zpl_xattr_get(ip, name, value, size);
} }

View File

@ -825,7 +825,7 @@ zvol_get_data(void *arg, lr_write_t *lr, char *buf, zio_t *zio)
ASSERT(zio != NULL); ASSERT(zio != NULL);
ASSERT(size != 0); ASSERT(size != 0);
zgd = (zgd_t *)kmem_zalloc(sizeof (zgd_t), KM_PUSHPAGE); zgd = (zgd_t *)kmem_zalloc(sizeof (zgd_t), KM_SLEEP);
zgd->zgd_zilog = zv->zv_zilog; zgd->zgd_zilog = zv->zv_zilog;
zgd->zgd_rl = zfs_range_lock(&zv->zv_znode, offset, size, RL_READER); zgd->zgd_rl = zfs_range_lock(&zv->zv_znode, offset, size, RL_READER);
@ -1234,7 +1234,7 @@ zvol_alloc(dev_t dev, const char *name)
zvol_state_t *zv; zvol_state_t *zv;
int error = 0; int error = 0;
zv = kmem_zalloc(sizeof (zvol_state_t), KM_PUSHPAGE); zv = kmem_zalloc(sizeof (zvol_state_t), KM_SLEEP);
spin_lock_init(&zv->zv_lock); spin_lock_init(&zv->zv_lock);
list_link_init(&zv->zv_next); list_link_init(&zv->zv_next);
@ -1314,7 +1314,7 @@ __zvol_snapdev_hidden(const char *name)
char *atp; char *atp;
int error = 0; int error = 0;
parent = kmem_alloc(MAXPATHLEN, KM_PUSHPAGE); parent = kmem_alloc(MAXPATHLEN, KM_SLEEP);
(void) strlcpy(parent, name, MAXPATHLEN); (void) strlcpy(parent, name, MAXPATHLEN);
if ((atp = strrchr(parent, '@')) != NULL) { if ((atp = strrchr(parent, '@')) != NULL) {
@ -1353,7 +1353,7 @@ __zvol_create_minor(const char *name, boolean_t ignore_snapdev)
goto out; goto out;
} }
doi = kmem_alloc(sizeof (dmu_object_info_t), KM_PUSHPAGE); doi = kmem_alloc(sizeof (dmu_object_info_t), KM_SLEEP);
error = dmu_objset_own(name, DMU_OST_ZVOL, B_TRUE, zvol_tag, &os); error = dmu_objset_own(name, DMU_OST_ZVOL, B_TRUE, zvol_tag, &os);
if (error) if (error)
@ -1565,7 +1565,7 @@ zvol_rename_minors(const char *oldname, const char *newname)
oldnamelen = strlen(oldname); oldnamelen = strlen(oldname);
newnamelen = strlen(newname); newnamelen = strlen(newname);
name = kmem_alloc(MAXNAMELEN, KM_PUSHPAGE); name = kmem_alloc(MAXNAMELEN, KM_SLEEP);
mutex_enter(&zvol_state_lock); mutex_enter(&zvol_state_lock);