diff --git a/module/os/linux/zfs/vdev_disk.c b/module/os/linux/zfs/vdev_disk.c index 29e54b39a..929ff9a25 100644 --- a/module/os/linux/zfs/vdev_disk.c +++ b/module/os/linux/zfs/vdev_disk.c @@ -614,7 +614,7 @@ static inline uint_t vdev_bio_max_segs(struct block_device *bdev) { /* - * Smallest of the device max segs and the tuneable max segs. Minimum + * Smallest of the device max segs and the tunable max segs. Minimum * 4, so there's room to finish split pages if they come up. */ const uint_t dev_max_segs = queue_max_segments(bdev_get_queue(bdev)); diff --git a/module/zfs/ddt.c b/module/zfs/ddt.c index de35f748f..2fce4f393 100644 --- a/module/zfs/ddt.c +++ b/module/zfs/ddt.c @@ -153,7 +153,7 @@ * storage object (ie ZAP) as normal. OpenZFS will try hard to flush enough to * keep up with the rate of change on dedup entries, but not so much that it * would impact overall throughput, and not using too much memory. See the - * zfs_dedup_log_* tuneables in zfs(4) for more details. + * zfs_dedup_log_* tunables in zfs(4) for more details. * * ## Repair IO * diff --git a/module/zfs/dmu.c b/module/zfs/dmu.c index bddb90f29..8b6b10d8d 100644 --- a/module/zfs/dmu.c +++ b/module/zfs/dmu.c @@ -2370,7 +2370,7 @@ dmu_write_policy(objset_t *os, dnode_t *dn, int level, int wp, zio_prop_t *zp) if (dmu_ddt_copies > 0) { /* - * If this tuneable is set, and this is a write for a + * If this tunable is set, and this is a write for a * dedup entry store (zap or log), then we treat it * something like ZFS_REDUNDANT_METADATA_MOST on a * regular dataset: this many copies, and one more for