2025-01-04 04:14:31 +03:00
|
|
|
// SPDX-License-Identifier: BSD-2-Clause
|
2020-04-14 21:36:28 +03:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2020 iXsystems, Inc.
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' AND
|
|
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHORS OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
|
|
* SUCH DAMAGE.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
2020-07-26 06:09:50 +03:00
|
|
|
#include <sys/types.h>
|
|
|
|
#include <sys/param.h>
|
2020-04-14 21:36:28 +03:00
|
|
|
#include <sys/dmu.h>
|
|
|
|
#include <sys/dmu_impl.h>
|
|
|
|
#include <sys/dmu_tx.h>
|
|
|
|
#include <sys/dbuf.h>
|
|
|
|
#include <sys/dnode.h>
|
|
|
|
#include <sys/zfs_context.h>
|
|
|
|
#include <sys/dmu_objset.h>
|
|
|
|
#include <sys/dmu_traverse.h>
|
|
|
|
#include <sys/dsl_dataset.h>
|
|
|
|
#include <sys/dsl_dir.h>
|
|
|
|
#include <sys/dsl_pool.h>
|
|
|
|
#include <sys/dsl_synctask.h>
|
|
|
|
#include <sys/dsl_prop.h>
|
|
|
|
#include <sys/zfs_ioctl.h>
|
|
|
|
#include <sys/zap.h>
|
|
|
|
#include <sys/zio_checksum.h>
|
|
|
|
#include <sys/zio_compress.h>
|
|
|
|
#include <sys/sa.h>
|
|
|
|
#include <sys/zfeature.h>
|
|
|
|
#include <sys/abd.h>
|
|
|
|
#include <sys/zfs_rlock.h>
|
|
|
|
#include <sys/racct.h>
|
|
|
|
#include <sys/vm.h>
|
|
|
|
#include <sys/zfs_znode.h>
|
|
|
|
#include <sys/zfs_vnops.h>
|
|
|
|
|
2020-07-26 06:09:50 +03:00
|
|
|
#include <sys/ccompat.h>
|
2020-04-14 21:36:28 +03:00
|
|
|
|
|
|
|
#ifndef IDX_TO_OFF
|
|
|
|
#define IDX_TO_OFF(idx) (((vm_ooffset_t)(idx)) << PAGE_SHIFT)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#define VM_ALLOC_BUSY_FLAGS VM_ALLOC_SBUSY | VM_ALLOC_IGN_SBUSY
|
|
|
|
|
|
|
|
int
|
|
|
|
dmu_write_pages(objset_t *os, uint64_t object, uint64_t offset, uint64_t size,
|
|
|
|
vm_page_t *ma, dmu_tx_t *tx)
|
|
|
|
{
|
|
|
|
dmu_buf_t **dbp;
|
|
|
|
struct sf_buf *sf;
|
|
|
|
int numbufs, i;
|
|
|
|
int err;
|
Wire O_DIRECT also to Uncached I/O (#17218)
Before Direct I/O was implemented, I've implemented lighter version
I called Uncached I/O. It uses normal DMU/ARC data path with some
optimizations, but evicts data from caches as soon as possible and
reasonable. Originally I wired it only to a primarycache property,
but now completing the integration all the way up to the VFS.
While Direct I/O has the lowest possible memory bandwidth usage,
it also has a significant number of limitations. It require I/Os
to be page aligned, does not allow speculative prefetch, etc. The
Uncached I/O does not have those limitations, but instead require
additional memory copy, though still one less than regular cached
I/O. As such it should fill the gap in between. Considering this
I've disabled annoying EINVAL errors on misaligned requests, adding
a tunable for those who wants to test their applications.
To pass the information between the layers I had to change a number
of APIs. But as side effect upper layers can now control not only
the caching, but also speculative prefetch. I haven't wired it to
VFS yet, since it require looking on some OS specifics. But while
there I've implemented speculative prefetch of indirect blocks for
Direct I/O, controllable via all the same mechanisms.
Signed-off-by: Alexander Motin <mav@FreeBSD.org>
Sponsored by: iXsystems, Inc.
Fixes #17027
Reviewed-by: Rob Norris <robn@despairlabs.com>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
2025-05-14 00:26:55 +03:00
|
|
|
dmu_flags_t flags = 0;
|
2020-04-14 21:36:28 +03:00
|
|
|
|
|
|
|
if (size == 0)
|
|
|
|
return (0);
|
|
|
|
|
|
|
|
err = dmu_buf_hold_array(os, object, offset, size,
|
|
|
|
FALSE, FTAG, &numbufs, &dbp);
|
|
|
|
if (err)
|
|
|
|
return (err);
|
|
|
|
|
|
|
|
for (i = 0; i < numbufs; i++) {
|
|
|
|
int tocpy, copied, thiscpy;
|
|
|
|
int bufoff;
|
|
|
|
dmu_buf_t *db = dbp[i];
|
|
|
|
caddr_t va;
|
|
|
|
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3U(size, >, 0);
|
2020-04-14 21:36:28 +03:00
|
|
|
ASSERT3U(db->db_size, >=, PAGESIZE);
|
|
|
|
|
|
|
|
bufoff = offset - db->db_offset;
|
|
|
|
tocpy = (int)MIN(db->db_size - bufoff, size);
|
|
|
|
|
|
|
|
ASSERT(i == 0 || i == numbufs-1 || tocpy == db->db_size);
|
|
|
|
|
Wire O_DIRECT also to Uncached I/O (#17218)
Before Direct I/O was implemented, I've implemented lighter version
I called Uncached I/O. It uses normal DMU/ARC data path with some
optimizations, but evicts data from caches as soon as possible and
reasonable. Originally I wired it only to a primarycache property,
but now completing the integration all the way up to the VFS.
While Direct I/O has the lowest possible memory bandwidth usage,
it also has a significant number of limitations. It require I/Os
to be page aligned, does not allow speculative prefetch, etc. The
Uncached I/O does not have those limitations, but instead require
additional memory copy, though still one less than regular cached
I/O. As such it should fill the gap in between. Considering this
I've disabled annoying EINVAL errors on misaligned requests, adding
a tunable for those who wants to test their applications.
To pass the information between the layers I had to change a number
of APIs. But as side effect upper layers can now control not only
the caching, but also speculative prefetch. I haven't wired it to
VFS yet, since it require looking on some OS specifics. But while
there I've implemented speculative prefetch of indirect blocks for
Direct I/O, controllable via all the same mechanisms.
Signed-off-by: Alexander Motin <mav@FreeBSD.org>
Sponsored by: iXsystems, Inc.
Fixes #17027
Reviewed-by: Rob Norris <robn@despairlabs.com>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
2025-05-14 00:26:55 +03:00
|
|
|
if (tocpy == db->db_size) {
|
dmu: Allow buffer fills to fail
When ZFS overwrites a whole block, it does not bother to read the
old content from disk. It is a good optimization, but if the buffer
fill fails due to page fault or something else, the buffer ends up
corrupted, neither keeping old content, nor getting the new one.
On FreeBSD this is additionally complicated by page faults being
blocked by VFS layer, always returning EFAULT on attempt to write
from mmap()'ed but not yet cached address range. Normally it is
not a big problem, since after original failure VFS will retry the
write after reading the required data. The problem becomes worse
in specific case when somebody tries to write into a file its own
mmap()'ed content from the same location. In that situation the
only copy of the data is getting corrupted on the page fault and
the following retries only fixate the status quo. Block cloning
makes this issue easier to reproduce, since it does not read the
old data, unlike traditional file copy, that may work by chance.
This patch provides the fill status to dmu_buf_fill_done(), that
in case of error can destroy the corrupted buffer as if no write
happened. One more complication in case of block cloning is that
if error is possible during fill, dmu_buf_will_fill() must read
the data via fall-back to dmu_buf_will_dirty(). It is required
to allow in case of error restoring the buffer to a state after
the cloning, not not before it, that would happen if we just call
dbuf_undirty().
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Rob Norris <robn@despairlabs.com>
Signed-off-by: Alexander Motin <mav@FreeBSD.org>
Sponsored by: iXsystems, Inc.
Closes #15665
2023-12-15 20:51:41 +03:00
|
|
|
dmu_buf_will_fill(db, tx, B_FALSE);
|
Wire O_DIRECT also to Uncached I/O (#17218)
Before Direct I/O was implemented, I've implemented lighter version
I called Uncached I/O. It uses normal DMU/ARC data path with some
optimizations, but evicts data from caches as soon as possible and
reasonable. Originally I wired it only to a primarycache property,
but now completing the integration all the way up to the VFS.
While Direct I/O has the lowest possible memory bandwidth usage,
it also has a significant number of limitations. It require I/Os
to be page aligned, does not allow speculative prefetch, etc. The
Uncached I/O does not have those limitations, but instead require
additional memory copy, though still one less than regular cached
I/O. As such it should fill the gap in between. Considering this
I've disabled annoying EINVAL errors on misaligned requests, adding
a tunable for those who wants to test their applications.
To pass the information between the layers I had to change a number
of APIs. But as side effect upper layers can now control not only
the caching, but also speculative prefetch. I haven't wired it to
VFS yet, since it require looking on some OS specifics. But while
there I've implemented speculative prefetch of indirect blocks for
Direct I/O, controllable via all the same mechanisms.
Signed-off-by: Alexander Motin <mav@FreeBSD.org>
Sponsored by: iXsystems, Inc.
Fixes #17027
Reviewed-by: Rob Norris <robn@despairlabs.com>
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
2025-05-14 00:26:55 +03:00
|
|
|
} else {
|
|
|
|
if (i == numbufs - 1 && bufoff + tocpy < db->db_size) {
|
|
|
|
if (bufoff == 0)
|
|
|
|
flags |= DMU_PARTIAL_FIRST;
|
|
|
|
else
|
|
|
|
flags |= DMU_PARTIAL_MORE;
|
|
|
|
}
|
|
|
|
dmu_buf_will_dirty_flags(db, tx, flags);
|
|
|
|
}
|
2020-04-14 21:36:28 +03:00
|
|
|
|
|
|
|
for (copied = 0; copied < tocpy; copied += PAGESIZE) {
|
|
|
|
ASSERT3U(ptoa((*ma)->pindex), ==,
|
|
|
|
db->db_offset + bufoff);
|
|
|
|
thiscpy = MIN(PAGESIZE, tocpy - copied);
|
|
|
|
va = zfs_map_page(*ma, &sf);
|
2024-12-03 04:08:40 +03:00
|
|
|
ASSERT(db->db_data != NULL);
|
2022-02-25 16:26:54 +03:00
|
|
|
memcpy((char *)db->db_data + bufoff, va, thiscpy);
|
2020-04-14 21:36:28 +03:00
|
|
|
zfs_unmap_page(sf);
|
|
|
|
ma += 1;
|
|
|
|
bufoff += PAGESIZE;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (tocpy == db->db_size)
|
dmu: Allow buffer fills to fail
When ZFS overwrites a whole block, it does not bother to read the
old content from disk. It is a good optimization, but if the buffer
fill fails due to page fault or something else, the buffer ends up
corrupted, neither keeping old content, nor getting the new one.
On FreeBSD this is additionally complicated by page faults being
blocked by VFS layer, always returning EFAULT on attempt to write
from mmap()'ed but not yet cached address range. Normally it is
not a big problem, since after original failure VFS will retry the
write after reading the required data. The problem becomes worse
in specific case when somebody tries to write into a file its own
mmap()'ed content from the same location. In that situation the
only copy of the data is getting corrupted on the page fault and
the following retries only fixate the status quo. Block cloning
makes this issue easier to reproduce, since it does not read the
old data, unlike traditional file copy, that may work by chance.
This patch provides the fill status to dmu_buf_fill_done(), that
in case of error can destroy the corrupted buffer as if no write
happened. One more complication in case of block cloning is that
if error is possible during fill, dmu_buf_will_fill() must read
the data via fall-back to dmu_buf_will_dirty(). It is required
to allow in case of error restoring the buffer to a state after
the cloning, not not before it, that would happen if we just call
dbuf_undirty().
Reviewed-by: Brian Behlendorf <behlendorf1@llnl.gov>
Reviewed-by: Rob Norris <robn@despairlabs.com>
Signed-off-by: Alexander Motin <mav@FreeBSD.org>
Sponsored by: iXsystems, Inc.
Closes #15665
2023-12-15 20:51:41 +03:00
|
|
|
dmu_buf_fill_done(db, tx, B_FALSE);
|
2020-04-14 21:36:28 +03:00
|
|
|
|
|
|
|
offset += tocpy;
|
|
|
|
size -= tocpy;
|
|
|
|
}
|
|
|
|
dmu_buf_rele_array(dbp, numbufs, FTAG);
|
|
|
|
return (err);
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
dmu_read_pages(objset_t *os, uint64_t object, vm_page_t *ma, int count,
|
|
|
|
int *rbehind, int *rahead, int last_size)
|
|
|
|
{
|
|
|
|
struct sf_buf *sf;
|
|
|
|
vm_object_t vmobj;
|
|
|
|
vm_page_t m;
|
|
|
|
dmu_buf_t **dbp;
|
|
|
|
dmu_buf_t *db;
|
|
|
|
caddr_t va;
|
|
|
|
int numbufs, i;
|
|
|
|
int bufoff, pgoff, tocpy;
|
|
|
|
int mi, di;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
ASSERT3U(ma[0]->pindex + count - 1, ==, ma[count - 1]->pindex);
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(last_size, <=, PAGE_SIZE);
|
2020-04-14 21:36:28 +03:00
|
|
|
|
|
|
|
err = dmu_buf_hold_array(os, object, IDX_TO_OFF(ma[0]->pindex),
|
|
|
|
IDX_TO_OFF(count - 1) + last_size, TRUE, FTAG, &numbufs, &dbp);
|
|
|
|
if (err != 0)
|
|
|
|
return (err);
|
|
|
|
|
2020-07-26 06:07:44 +03:00
|
|
|
#ifdef ZFS_DEBUG
|
2020-04-14 21:36:28 +03:00
|
|
|
IMPLY(last_size < PAGE_SIZE, *rahead == 0);
|
|
|
|
if (dbp[0]->db_offset != 0 || numbufs > 1) {
|
|
|
|
for (i = 0; i < numbufs; i++) {
|
|
|
|
ASSERT(ISP2(dbp[i]->db_size));
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3U((dbp[i]->db_offset % dbp[i]->db_size), ==, 0);
|
2020-04-14 21:36:28 +03:00
|
|
|
ASSERT3U(dbp[i]->db_size, ==, dbp[0]->db_size);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
vmobj = ma[0]->object;
|
|
|
|
|
|
|
|
db = dbp[0];
|
|
|
|
for (i = 0; i < *rbehind; i++) {
|
2020-04-17 19:30:26 +03:00
|
|
|
m = vm_page_grab_unlocked(vmobj, ma[0]->pindex - 1 - i,
|
2020-04-14 21:36:28 +03:00
|
|
|
VM_ALLOC_NORMAL | VM_ALLOC_NOWAIT | VM_ALLOC_BUSY_FLAGS);
|
|
|
|
if (m == NULL)
|
|
|
|
break;
|
|
|
|
if (!vm_page_none_valid(m)) {
|
|
|
|
ASSERT3U(m->valid, ==, VM_PAGE_BITS_ALL);
|
2024-08-06 02:56:45 +03:00
|
|
|
vm_page_sunbusy(m);
|
2020-04-14 21:36:28 +03:00
|
|
|
break;
|
|
|
|
}
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3U(m->dirty, ==, 0);
|
2020-04-17 19:30:26 +03:00
|
|
|
ASSERT(!pmap_page_is_write_mapped(m));
|
2020-04-14 21:36:28 +03:00
|
|
|
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3U(db->db_size, >, PAGE_SIZE);
|
2020-04-14 21:36:28 +03:00
|
|
|
bufoff = IDX_TO_OFF(m->pindex) % db->db_size;
|
|
|
|
va = zfs_map_page(m, &sf);
|
2024-12-03 04:08:40 +03:00
|
|
|
ASSERT(db->db_data != NULL);
|
2022-02-25 16:26:54 +03:00
|
|
|
memcpy(va, (char *)db->db_data + bufoff, PAGESIZE);
|
2020-04-14 21:36:28 +03:00
|
|
|
zfs_unmap_page(sf);
|
|
|
|
vm_page_valid(m);
|
|
|
|
if ((m->busy_lock & VPB_BIT_WAITERS) != 0)
|
|
|
|
vm_page_activate(m);
|
|
|
|
else
|
|
|
|
vm_page_deactivate(m);
|
2024-08-06 02:56:45 +03:00
|
|
|
vm_page_sunbusy(m);
|
2020-04-14 21:36:28 +03:00
|
|
|
}
|
|
|
|
*rbehind = i;
|
|
|
|
|
|
|
|
bufoff = IDX_TO_OFF(ma[0]->pindex) % db->db_size;
|
|
|
|
pgoff = 0;
|
|
|
|
for (mi = 0, di = 0; mi < count && di < numbufs; ) {
|
|
|
|
if (pgoff == 0) {
|
|
|
|
m = ma[mi];
|
|
|
|
if (m != bogus_page) {
|
|
|
|
vm_page_assert_xbusied(m);
|
|
|
|
ASSERT(vm_page_none_valid(m));
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3U(m->dirty, ==, 0);
|
2020-04-17 19:30:26 +03:00
|
|
|
ASSERT(!pmap_page_is_write_mapped(m));
|
2020-04-14 21:36:28 +03:00
|
|
|
va = zfs_map_page(m, &sf);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (bufoff == 0)
|
|
|
|
db = dbp[di];
|
|
|
|
|
|
|
|
if (m != bogus_page) {
|
|
|
|
ASSERT3U(IDX_TO_OFF(m->pindex) + pgoff, ==,
|
|
|
|
db->db_offset + bufoff);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We do not need to clamp the copy size by the file
|
|
|
|
* size as the last block is zero-filled beyond the
|
|
|
|
* end of file anyway.
|
|
|
|
*/
|
|
|
|
tocpy = MIN(db->db_size - bufoff, PAGESIZE - pgoff);
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(tocpy, >=, 0);
|
2024-12-03 04:08:40 +03:00
|
|
|
if (m != bogus_page) {
|
|
|
|
ASSERT(db->db_data != NULL);
|
2022-02-25 16:26:54 +03:00
|
|
|
memcpy(va + pgoff, (char *)db->db_data + bufoff, tocpy);
|
2024-12-03 04:08:40 +03:00
|
|
|
}
|
2020-04-14 21:36:28 +03:00
|
|
|
|
|
|
|
pgoff += tocpy;
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(pgoff, >=, 0);
|
|
|
|
ASSERT3S(pgoff, <=, PAGESIZE);
|
2020-04-14 21:36:28 +03:00
|
|
|
if (pgoff == PAGESIZE) {
|
|
|
|
if (m != bogus_page) {
|
|
|
|
zfs_unmap_page(sf);
|
|
|
|
vm_page_valid(m);
|
|
|
|
}
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(mi, <, count);
|
2020-04-14 21:36:28 +03:00
|
|
|
mi++;
|
|
|
|
pgoff = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
bufoff += tocpy;
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(bufoff, >=, 0);
|
|
|
|
ASSERT3S(bufoff, <=, db->db_size);
|
2020-04-14 21:36:28 +03:00
|
|
|
if (bufoff == db->db_size) {
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(di, <, numbufs);
|
2020-04-14 21:36:28 +03:00
|
|
|
di++;
|
|
|
|
bufoff = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-07-26 06:07:44 +03:00
|
|
|
#ifdef ZFS_DEBUG
|
2020-04-14 21:36:28 +03:00
|
|
|
/*
|
|
|
|
* Three possibilities:
|
|
|
|
* - last requested page ends at a buffer boundary and , thus,
|
|
|
|
* all pages and buffers have been iterated;
|
|
|
|
* - all requested pages are filled, but the last buffer
|
|
|
|
* has not been exhausted;
|
|
|
|
* the read-ahead is possible only in this case;
|
|
|
|
* - all buffers have been read, but the last page has not been
|
|
|
|
* fully filled;
|
|
|
|
* this is only possible if the file has only a single buffer
|
|
|
|
* with a size that is not a multiple of the page size.
|
|
|
|
*/
|
|
|
|
if (mi == count) {
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(di, >=, numbufs - 1);
|
2020-04-14 21:36:28 +03:00
|
|
|
IMPLY(*rahead != 0, di == numbufs - 1);
|
|
|
|
IMPLY(*rahead != 0, bufoff != 0);
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT0(pgoff);
|
2020-04-14 21:36:28 +03:00
|
|
|
}
|
|
|
|
if (di == numbufs) {
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(mi, >=, count - 1);
|
|
|
|
ASSERT0(*rahead);
|
2020-04-14 21:36:28 +03:00
|
|
|
IMPLY(pgoff == 0, mi == count);
|
|
|
|
if (pgoff != 0) {
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(mi, ==, count - 1);
|
|
|
|
ASSERT3U((dbp[0]->db_size & PAGE_MASK), !=, 0);
|
2020-04-14 21:36:28 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
if (pgoff != 0) {
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3P(m, !=, bogus_page);
|
2022-02-25 16:26:54 +03:00
|
|
|
memset(va + pgoff, 0, PAGESIZE - pgoff);
|
2020-04-14 21:36:28 +03:00
|
|
|
zfs_unmap_page(sf);
|
|
|
|
vm_page_valid(m);
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < *rahead; i++) {
|
2020-04-17 19:30:26 +03:00
|
|
|
m = vm_page_grab_unlocked(vmobj, ma[count - 1]->pindex + 1 + i,
|
2020-04-14 21:36:28 +03:00
|
|
|
VM_ALLOC_NORMAL | VM_ALLOC_NOWAIT | VM_ALLOC_BUSY_FLAGS);
|
|
|
|
if (m == NULL)
|
|
|
|
break;
|
|
|
|
if (!vm_page_none_valid(m)) {
|
|
|
|
ASSERT3U(m->valid, ==, VM_PAGE_BITS_ALL);
|
2024-08-06 02:56:45 +03:00
|
|
|
vm_page_sunbusy(m);
|
2020-04-14 21:36:28 +03:00
|
|
|
break;
|
|
|
|
}
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3U(m->dirty, ==, 0);
|
2021-02-28 04:23:09 +03:00
|
|
|
ASSERT(!pmap_page_is_write_mapped(m));
|
2020-04-14 21:36:28 +03:00
|
|
|
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3U(db->db_size, >, PAGE_SIZE);
|
2020-04-14 21:36:28 +03:00
|
|
|
bufoff = IDX_TO_OFF(m->pindex) % db->db_size;
|
|
|
|
tocpy = MIN(db->db_size - bufoff, PAGESIZE);
|
|
|
|
va = zfs_map_page(m, &sf);
|
2024-12-03 04:08:40 +03:00
|
|
|
ASSERT(db->db_data != NULL);
|
2022-02-25 16:26:54 +03:00
|
|
|
memcpy(va, (char *)db->db_data + bufoff, tocpy);
|
2020-04-14 21:36:28 +03:00
|
|
|
if (tocpy < PAGESIZE) {
|
2021-05-01 02:36:10 +03:00
|
|
|
ASSERT3S(i, ==, *rahead - 1);
|
|
|
|
ASSERT3U((db->db_size & PAGE_MASK), !=, 0);
|
2022-02-25 16:26:54 +03:00
|
|
|
memset(va + tocpy, 0, PAGESIZE - tocpy);
|
2020-04-14 21:36:28 +03:00
|
|
|
}
|
|
|
|
zfs_unmap_page(sf);
|
|
|
|
vm_page_valid(m);
|
|
|
|
if ((m->busy_lock & VPB_BIT_WAITERS) != 0)
|
|
|
|
vm_page_activate(m);
|
|
|
|
else
|
|
|
|
vm_page_deactivate(m);
|
2024-08-06 02:56:45 +03:00
|
|
|
vm_page_sunbusy(m);
|
2020-04-14 21:36:28 +03:00
|
|
|
}
|
|
|
|
*rahead = i;
|
|
|
|
|
|
|
|
dmu_buf_rele_array(dbp, numbufs, FTAG);
|
|
|
|
return (0);
|
|
|
|
}
|