1420 lines
36 KiB
C
1420 lines
36 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Copyright (c) 2021-2022, NVIDIA CORPORATION & AFFILIATES.
|
|
*
|
|
* The io_pagetable is the top of datastructure that maps IOVA's to PFNs. The
|
|
* PFNs can be placed into an iommu_domain, or returned to the caller as a page
|
|
* list for access by an in-kernel user.
|
|
*
|
|
* The datastructure uses the iopt_pages to optimize the storage of the PFNs
|
|
* between the domains and xarray.
|
|
*/
|
|
#include <linux/iommufd.h>
|
|
#include <linux/lockdep.h>
|
|
#include <linux/iommu.h>
|
|
#include <linux/sched/mm.h>
|
|
#include <linux/err.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/errno.h>
|
|
#include <uapi/linux/iommufd.h>
|
|
|
|
#include "io_pagetable.h"
|
|
#include "double_span.h"
|
|
|
|
struct iopt_pages_list {
|
|
struct iopt_pages *pages;
|
|
struct iopt_area *area;
|
|
struct list_head next;
|
|
unsigned long start_byte;
|
|
unsigned long length;
|
|
};
|
|
|
|
struct iopt_area *iopt_area_contig_init(struct iopt_area_contig_iter *iter,
|
|
struct io_pagetable *iopt,
|
|
unsigned long iova,
|
|
unsigned long last_iova)
|
|
{
|
|
lockdep_assert_held(&iopt->iova_rwsem);
|
|
|
|
iter->cur_iova = iova;
|
|
iter->last_iova = last_iova;
|
|
iter->area = iopt_area_iter_first(iopt, iova, iova);
|
|
if (!iter->area)
|
|
return NULL;
|
|
if (!iter->area->pages) {
|
|
iter->area = NULL;
|
|
return NULL;
|
|
}
|
|
return iter->area;
|
|
}
|
|
|
|
struct iopt_area *iopt_area_contig_next(struct iopt_area_contig_iter *iter)
|
|
{
|
|
unsigned long last_iova;
|
|
|
|
if (!iter->area)
|
|
return NULL;
|
|
last_iova = iopt_area_last_iova(iter->area);
|
|
if (iter->last_iova <= last_iova)
|
|
return NULL;
|
|
|
|
iter->cur_iova = last_iova + 1;
|
|
iter->area = iopt_area_iter_next(iter->area, iter->cur_iova,
|
|
iter->last_iova);
|
|
if (!iter->area)
|
|
return NULL;
|
|
if (iter->cur_iova != iopt_area_iova(iter->area) ||
|
|
!iter->area->pages) {
|
|
iter->area = NULL;
|
|
return NULL;
|
|
}
|
|
return iter->area;
|
|
}
|
|
|
|
static bool __alloc_iova_check_hole(struct interval_tree_double_span_iter *span,
|
|
unsigned long length,
|
|
unsigned long iova_alignment,
|
|
unsigned long page_offset)
|
|
{
|
|
if (span->is_used || span->last_hole - span->start_hole < length - 1)
|
|
return false;
|
|
|
|
span->start_hole = ALIGN(span->start_hole, iova_alignment) |
|
|
page_offset;
|
|
if (span->start_hole > span->last_hole ||
|
|
span->last_hole - span->start_hole < length - 1)
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
static bool __alloc_iova_check_used(struct interval_tree_span_iter *span,
|
|
unsigned long length,
|
|
unsigned long iova_alignment,
|
|
unsigned long page_offset)
|
|
{
|
|
if (span->is_hole || span->last_used - span->start_used < length - 1)
|
|
return false;
|
|
|
|
span->start_used = ALIGN(span->start_used, iova_alignment) |
|
|
page_offset;
|
|
if (span->start_used > span->last_used ||
|
|
span->last_used - span->start_used < length - 1)
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
/*
|
|
* Automatically find a block of IOVA that is not being used and not reserved.
|
|
* Does not return a 0 IOVA even if it is valid.
|
|
*/
|
|
static int iopt_alloc_iova(struct io_pagetable *iopt, unsigned long *iova,
|
|
unsigned long uptr, unsigned long length)
|
|
{
|
|
unsigned long page_offset = uptr % PAGE_SIZE;
|
|
struct interval_tree_double_span_iter used_span;
|
|
struct interval_tree_span_iter allowed_span;
|
|
unsigned long iova_alignment;
|
|
|
|
lockdep_assert_held(&iopt->iova_rwsem);
|
|
|
|
/* Protect roundup_pow-of_two() from overflow */
|
|
if (length == 0 || length >= ULONG_MAX / 2)
|
|
return -EOVERFLOW;
|
|
|
|
/*
|
|
* Keep alignment present in the uptr when building the IOVA, this
|
|
* increases the chance we can map a THP.
|
|
*/
|
|
if (!uptr)
|
|
iova_alignment = roundup_pow_of_two(length);
|
|
else
|
|
iova_alignment = min_t(unsigned long,
|
|
roundup_pow_of_two(length),
|
|
1UL << __ffs64(uptr));
|
|
|
|
if (iova_alignment < iopt->iova_alignment)
|
|
return -EINVAL;
|
|
|
|
interval_tree_for_each_span(&allowed_span, &iopt->allowed_itree,
|
|
PAGE_SIZE, ULONG_MAX - PAGE_SIZE) {
|
|
if (RB_EMPTY_ROOT(&iopt->allowed_itree.rb_root)) {
|
|
allowed_span.start_used = PAGE_SIZE;
|
|
allowed_span.last_used = ULONG_MAX - PAGE_SIZE;
|
|
allowed_span.is_hole = false;
|
|
}
|
|
|
|
if (!__alloc_iova_check_used(&allowed_span, length,
|
|
iova_alignment, page_offset))
|
|
continue;
|
|
|
|
interval_tree_for_each_double_span(
|
|
&used_span, &iopt->reserved_itree, &iopt->area_itree,
|
|
allowed_span.start_used, allowed_span.last_used) {
|
|
if (!__alloc_iova_check_hole(&used_span, length,
|
|
iova_alignment,
|
|
page_offset))
|
|
continue;
|
|
|
|
*iova = used_span.start_hole;
|
|
return 0;
|
|
}
|
|
}
|
|
return -ENOSPC;
|
|
}
|
|
|
|
static int iopt_check_iova(struct io_pagetable *iopt, unsigned long iova,
|
|
unsigned long length)
|
|
{
|
|
unsigned long last;
|
|
|
|
lockdep_assert_held(&iopt->iova_rwsem);
|
|
|
|
if ((iova & (iopt->iova_alignment - 1)))
|
|
return -EINVAL;
|
|
|
|
if (check_add_overflow(iova, length - 1, &last))
|
|
return -EOVERFLOW;
|
|
|
|
/* No reserved IOVA intersects the range */
|
|
if (iopt_reserved_iter_first(iopt, iova, last))
|
|
return -EINVAL;
|
|
|
|
/* Check that there is not already a mapping in the range */
|
|
if (iopt_area_iter_first(iopt, iova, last))
|
|
return -EEXIST;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* The area takes a slice of the pages from start_bytes to start_byte + length
|
|
*/
|
|
static int iopt_insert_area(struct io_pagetable *iopt, struct iopt_area *area,
|
|
struct iopt_pages *pages, unsigned long iova,
|
|
unsigned long start_byte, unsigned long length,
|
|
int iommu_prot)
|
|
{
|
|
lockdep_assert_held_write(&iopt->iova_rwsem);
|
|
|
|
if ((iommu_prot & IOMMU_WRITE) && !pages->writable)
|
|
return -EPERM;
|
|
|
|
area->iommu_prot = iommu_prot;
|
|
area->page_offset = start_byte % PAGE_SIZE;
|
|
if (area->page_offset & (iopt->iova_alignment - 1))
|
|
return -EINVAL;
|
|
|
|
area->node.start = iova;
|
|
if (check_add_overflow(iova, length - 1, &area->node.last))
|
|
return -EOVERFLOW;
|
|
|
|
area->pages_node.start = start_byte / PAGE_SIZE;
|
|
if (check_add_overflow(start_byte, length - 1, &area->pages_node.last))
|
|
return -EOVERFLOW;
|
|
area->pages_node.last = area->pages_node.last / PAGE_SIZE;
|
|
if (WARN_ON(area->pages_node.last >= pages->npages))
|
|
return -EOVERFLOW;
|
|
|
|
/*
|
|
* The area is inserted with a NULL pages indicating it is not fully
|
|
* initialized yet.
|
|
*/
|
|
area->iopt = iopt;
|
|
interval_tree_insert(&area->node, &iopt->area_itree);
|
|
return 0;
|
|
}
|
|
|
|
static struct iopt_area *iopt_area_alloc(void)
|
|
{
|
|
struct iopt_area *area;
|
|
|
|
area = kzalloc(sizeof(*area), GFP_KERNEL_ACCOUNT);
|
|
if (!area)
|
|
return NULL;
|
|
RB_CLEAR_NODE(&area->node.rb);
|
|
RB_CLEAR_NODE(&area->pages_node.rb);
|
|
return area;
|
|
}
|
|
|
|
static int iopt_alloc_area_pages(struct io_pagetable *iopt,
|
|
struct list_head *pages_list,
|
|
unsigned long length, unsigned long *dst_iova,
|
|
int iommu_prot, unsigned int flags)
|
|
{
|
|
struct iopt_pages_list *elm;
|
|
unsigned long iova;
|
|
int rc = 0;
|
|
|
|
list_for_each_entry(elm, pages_list, next) {
|
|
elm->area = iopt_area_alloc();
|
|
if (!elm->area)
|
|
return -ENOMEM;
|
|
}
|
|
|
|
down_write(&iopt->iova_rwsem);
|
|
if ((length & (iopt->iova_alignment - 1)) || !length) {
|
|
rc = -EINVAL;
|
|
goto out_unlock;
|
|
}
|
|
|
|
if (flags & IOPT_ALLOC_IOVA) {
|
|
/* Use the first entry to guess the ideal IOVA alignment */
|
|
elm = list_first_entry(pages_list, struct iopt_pages_list,
|
|
next);
|
|
rc = iopt_alloc_iova(
|
|
iopt, dst_iova,
|
|
(uintptr_t)elm->pages->uptr + elm->start_byte, length);
|
|
if (rc)
|
|
goto out_unlock;
|
|
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) &&
|
|
WARN_ON(iopt_check_iova(iopt, *dst_iova, length))) {
|
|
rc = -EINVAL;
|
|
goto out_unlock;
|
|
}
|
|
} else {
|
|
rc = iopt_check_iova(iopt, *dst_iova, length);
|
|
if (rc)
|
|
goto out_unlock;
|
|
}
|
|
|
|
/*
|
|
* Areas are created with a NULL pages so that the IOVA space is
|
|
* reserved and we can unlock the iova_rwsem.
|
|
*/
|
|
iova = *dst_iova;
|
|
list_for_each_entry(elm, pages_list, next) {
|
|
rc = iopt_insert_area(iopt, elm->area, elm->pages, iova,
|
|
elm->start_byte, elm->length, iommu_prot);
|
|
if (rc)
|
|
goto out_unlock;
|
|
iova += elm->length;
|
|
}
|
|
|
|
out_unlock:
|
|
up_write(&iopt->iova_rwsem);
|
|
return rc;
|
|
}
|
|
|
|
static void iopt_abort_area(struct iopt_area *area)
|
|
{
|
|
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
|
|
WARN_ON(area->pages);
|
|
if (area->iopt) {
|
|
down_write(&area->iopt->iova_rwsem);
|
|
interval_tree_remove(&area->node, &area->iopt->area_itree);
|
|
up_write(&area->iopt->iova_rwsem);
|
|
}
|
|
kfree(area);
|
|
}
|
|
|
|
void iopt_free_pages_list(struct list_head *pages_list)
|
|
{
|
|
struct iopt_pages_list *elm;
|
|
|
|
while ((elm = list_first_entry_or_null(pages_list,
|
|
struct iopt_pages_list, next))) {
|
|
if (elm->area)
|
|
iopt_abort_area(elm->area);
|
|
if (elm->pages)
|
|
iopt_put_pages(elm->pages);
|
|
list_del(&elm->next);
|
|
kfree(elm);
|
|
}
|
|
}
|
|
|
|
static int iopt_fill_domains_pages(struct list_head *pages_list)
|
|
{
|
|
struct iopt_pages_list *undo_elm;
|
|
struct iopt_pages_list *elm;
|
|
int rc;
|
|
|
|
list_for_each_entry(elm, pages_list, next) {
|
|
rc = iopt_area_fill_domains(elm->area, elm->pages);
|
|
if (rc)
|
|
goto err_undo;
|
|
}
|
|
return 0;
|
|
|
|
err_undo:
|
|
list_for_each_entry(undo_elm, pages_list, next) {
|
|
if (undo_elm == elm)
|
|
break;
|
|
iopt_area_unfill_domains(undo_elm->area, undo_elm->pages);
|
|
}
|
|
return rc;
|
|
}
|
|
|
|
int iopt_map_pages(struct io_pagetable *iopt, struct list_head *pages_list,
|
|
unsigned long length, unsigned long *dst_iova,
|
|
int iommu_prot, unsigned int flags)
|
|
{
|
|
struct iopt_pages_list *elm;
|
|
int rc;
|
|
|
|
rc = iopt_alloc_area_pages(iopt, pages_list, length, dst_iova,
|
|
iommu_prot, flags);
|
|
if (rc)
|
|
return rc;
|
|
|
|
down_read(&iopt->domains_rwsem);
|
|
rc = iopt_fill_domains_pages(pages_list);
|
|
if (rc)
|
|
goto out_unlock_domains;
|
|
|
|
down_write(&iopt->iova_rwsem);
|
|
list_for_each_entry(elm, pages_list, next) {
|
|
/*
|
|
* area->pages must be set inside the domains_rwsem to ensure
|
|
* any newly added domains will get filled. Moves the reference
|
|
* in from the list.
|
|
*/
|
|
elm->area->pages = elm->pages;
|
|
elm->pages = NULL;
|
|
elm->area = NULL;
|
|
}
|
|
up_write(&iopt->iova_rwsem);
|
|
out_unlock_domains:
|
|
up_read(&iopt->domains_rwsem);
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* iopt_map_user_pages() - Map a user VA to an iova in the io page table
|
|
* @ictx: iommufd_ctx the iopt is part of
|
|
* @iopt: io_pagetable to act on
|
|
* @iova: If IOPT_ALLOC_IOVA is set this is unused on input and contains
|
|
* the chosen iova on output. Otherwise is the iova to map to on input
|
|
* @uptr: User VA to map
|
|
* @length: Number of bytes to map
|
|
* @iommu_prot: Combination of IOMMU_READ/WRITE/etc bits for the mapping
|
|
* @flags: IOPT_ALLOC_IOVA or zero
|
|
*
|
|
* iova, uptr, and length must be aligned to iova_alignment. For domain backed
|
|
* page tables this will pin the pages and load them into the domain at iova.
|
|
* For non-domain page tables this will only setup a lazy reference and the
|
|
* caller must use iopt_access_pages() to touch them.
|
|
*
|
|
* iopt_unmap_iova() must be called to undo this before the io_pagetable can be
|
|
* destroyed.
|
|
*/
|
|
int iopt_map_user_pages(struct iommufd_ctx *ictx, struct io_pagetable *iopt,
|
|
unsigned long *iova, void __user *uptr,
|
|
unsigned long length, int iommu_prot,
|
|
unsigned int flags)
|
|
{
|
|
struct iopt_pages_list elm = {};
|
|
LIST_HEAD(pages_list);
|
|
int rc;
|
|
|
|
elm.pages = iopt_alloc_pages(uptr, length, iommu_prot & IOMMU_WRITE);
|
|
if (IS_ERR(elm.pages))
|
|
return PTR_ERR(elm.pages);
|
|
if (ictx->account_mode == IOPT_PAGES_ACCOUNT_MM &&
|
|
elm.pages->account_mode == IOPT_PAGES_ACCOUNT_USER)
|
|
elm.pages->account_mode = IOPT_PAGES_ACCOUNT_MM;
|
|
elm.start_byte = uptr - elm.pages->uptr;
|
|
elm.length = length;
|
|
list_add(&elm.next, &pages_list);
|
|
|
|
rc = iopt_map_pages(iopt, &pages_list, length, iova, iommu_prot, flags);
|
|
if (rc) {
|
|
if (elm.area)
|
|
iopt_abort_area(elm.area);
|
|
if (elm.pages)
|
|
iopt_put_pages(elm.pages);
|
|
return rc;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
struct iova_bitmap_fn_arg {
|
|
unsigned long flags;
|
|
struct io_pagetable *iopt;
|
|
struct iommu_domain *domain;
|
|
struct iommu_dirty_bitmap *dirty;
|
|
};
|
|
|
|
static int __iommu_read_and_clear_dirty(struct iova_bitmap *bitmap,
|
|
unsigned long iova, size_t length,
|
|
void *opaque)
|
|
{
|
|
struct iopt_area *area;
|
|
struct iopt_area_contig_iter iter;
|
|
struct iova_bitmap_fn_arg *arg = opaque;
|
|
struct iommu_domain *domain = arg->domain;
|
|
struct iommu_dirty_bitmap *dirty = arg->dirty;
|
|
const struct iommu_dirty_ops *ops = domain->dirty_ops;
|
|
unsigned long last_iova = iova + length - 1;
|
|
unsigned long flags = arg->flags;
|
|
int ret;
|
|
|
|
iopt_for_each_contig_area(&iter, area, arg->iopt, iova, last_iova) {
|
|
unsigned long last = min(last_iova, iopt_area_last_iova(area));
|
|
|
|
ret = ops->read_and_clear_dirty(domain, iter.cur_iova,
|
|
last - iter.cur_iova + 1, flags,
|
|
dirty);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
if (!iopt_area_contig_done(&iter))
|
|
return -EINVAL;
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
iommu_read_and_clear_dirty(struct iommu_domain *domain,
|
|
struct io_pagetable *iopt, unsigned long flags,
|
|
struct iommu_hwpt_get_dirty_bitmap *bitmap)
|
|
{
|
|
const struct iommu_dirty_ops *ops = domain->dirty_ops;
|
|
struct iommu_iotlb_gather gather;
|
|
struct iommu_dirty_bitmap dirty;
|
|
struct iova_bitmap_fn_arg arg;
|
|
struct iova_bitmap *iter;
|
|
int ret = 0;
|
|
|
|
if (!ops || !ops->read_and_clear_dirty)
|
|
return -EOPNOTSUPP;
|
|
|
|
iter = iova_bitmap_alloc(bitmap->iova, bitmap->length,
|
|
bitmap->page_size,
|
|
u64_to_user_ptr(bitmap->data));
|
|
if (IS_ERR(iter))
|
|
return -ENOMEM;
|
|
|
|
iommu_dirty_bitmap_init(&dirty, iter, &gather);
|
|
|
|
arg.flags = flags;
|
|
arg.iopt = iopt;
|
|
arg.domain = domain;
|
|
arg.dirty = &dirty;
|
|
iova_bitmap_for_each(iter, &arg, __iommu_read_and_clear_dirty);
|
|
|
|
if (!(flags & IOMMU_DIRTY_NO_CLEAR))
|
|
iommu_iotlb_sync(domain, &gather);
|
|
|
|
iova_bitmap_free(iter);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int iommufd_check_iova_range(struct io_pagetable *iopt,
|
|
struct iommu_hwpt_get_dirty_bitmap *bitmap)
|
|
{
|
|
size_t iommu_pgsize = iopt->iova_alignment;
|
|
u64 last_iova;
|
|
|
|
if (check_add_overflow(bitmap->iova, bitmap->length - 1, &last_iova))
|
|
return -EOVERFLOW;
|
|
|
|
if (bitmap->iova > ULONG_MAX || last_iova > ULONG_MAX)
|
|
return -EOVERFLOW;
|
|
|
|
if ((bitmap->iova & (iommu_pgsize - 1)) ||
|
|
((last_iova + 1) & (iommu_pgsize - 1)))
|
|
return -EINVAL;
|
|
|
|
if (!bitmap->page_size)
|
|
return -EINVAL;
|
|
|
|
if ((bitmap->iova & (bitmap->page_size - 1)) ||
|
|
((last_iova + 1) & (bitmap->page_size - 1)))
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int iopt_read_and_clear_dirty_data(struct io_pagetable *iopt,
|
|
struct iommu_domain *domain,
|
|
unsigned long flags,
|
|
struct iommu_hwpt_get_dirty_bitmap *bitmap)
|
|
{
|
|
int ret;
|
|
|
|
ret = iommufd_check_iova_range(iopt, bitmap);
|
|
if (ret)
|
|
return ret;
|
|
|
|
down_read(&iopt->iova_rwsem);
|
|
ret = iommu_read_and_clear_dirty(domain, iopt, flags, bitmap);
|
|
up_read(&iopt->iova_rwsem);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int iopt_clear_dirty_data(struct io_pagetable *iopt,
|
|
struct iommu_domain *domain)
|
|
{
|
|
const struct iommu_dirty_ops *ops = domain->dirty_ops;
|
|
struct iommu_iotlb_gather gather;
|
|
struct iommu_dirty_bitmap dirty;
|
|
struct iopt_area *area;
|
|
int ret = 0;
|
|
|
|
lockdep_assert_held_read(&iopt->iova_rwsem);
|
|
|
|
iommu_dirty_bitmap_init(&dirty, NULL, &gather);
|
|
|
|
for (area = iopt_area_iter_first(iopt, 0, ULONG_MAX); area;
|
|
area = iopt_area_iter_next(area, 0, ULONG_MAX)) {
|
|
if (!area->pages)
|
|
continue;
|
|
|
|
ret = ops->read_and_clear_dirty(domain, iopt_area_iova(area),
|
|
iopt_area_length(area), 0,
|
|
&dirty);
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
iommu_iotlb_sync(domain, &gather);
|
|
return ret;
|
|
}
|
|
|
|
int iopt_set_dirty_tracking(struct io_pagetable *iopt,
|
|
struct iommu_domain *domain, bool enable)
|
|
{
|
|
const struct iommu_dirty_ops *ops = domain->dirty_ops;
|
|
int ret = 0;
|
|
|
|
if (!ops)
|
|
return -EOPNOTSUPP;
|
|
|
|
down_read(&iopt->iova_rwsem);
|
|
|
|
/* Clear dirty bits from PTEs to ensure a clean snapshot */
|
|
if (enable) {
|
|
ret = iopt_clear_dirty_data(iopt, domain);
|
|
if (ret)
|
|
goto out_unlock;
|
|
}
|
|
|
|
ret = ops->set_dirty_tracking(domain, enable);
|
|
|
|
out_unlock:
|
|
up_read(&iopt->iova_rwsem);
|
|
return ret;
|
|
}
|
|
|
|
int iopt_get_pages(struct io_pagetable *iopt, unsigned long iova,
|
|
unsigned long length, struct list_head *pages_list)
|
|
{
|
|
struct iopt_area_contig_iter iter;
|
|
unsigned long last_iova;
|
|
struct iopt_area *area;
|
|
int rc;
|
|
|
|
if (!length)
|
|
return -EINVAL;
|
|
if (check_add_overflow(iova, length - 1, &last_iova))
|
|
return -EOVERFLOW;
|
|
|
|
down_read(&iopt->iova_rwsem);
|
|
iopt_for_each_contig_area(&iter, area, iopt, iova, last_iova) {
|
|
struct iopt_pages_list *elm;
|
|
unsigned long last = min(last_iova, iopt_area_last_iova(area));
|
|
|
|
elm = kzalloc(sizeof(*elm), GFP_KERNEL_ACCOUNT);
|
|
if (!elm) {
|
|
rc = -ENOMEM;
|
|
goto err_free;
|
|
}
|
|
elm->start_byte = iopt_area_start_byte(area, iter.cur_iova);
|
|
elm->pages = area->pages;
|
|
elm->length = (last - iter.cur_iova) + 1;
|
|
kref_get(&elm->pages->kref);
|
|
list_add_tail(&elm->next, pages_list);
|
|
}
|
|
if (!iopt_area_contig_done(&iter)) {
|
|
rc = -ENOENT;
|
|
goto err_free;
|
|
}
|
|
up_read(&iopt->iova_rwsem);
|
|
return 0;
|
|
err_free:
|
|
up_read(&iopt->iova_rwsem);
|
|
iopt_free_pages_list(pages_list);
|
|
return rc;
|
|
}
|
|
|
|
static int iopt_unmap_iova_range(struct io_pagetable *iopt, unsigned long start,
|
|
unsigned long last, unsigned long *unmapped)
|
|
{
|
|
struct iopt_area *area;
|
|
unsigned long unmapped_bytes = 0;
|
|
unsigned int tries = 0;
|
|
int rc = -ENOENT;
|
|
|
|
/*
|
|
* The domains_rwsem must be held in read mode any time any area->pages
|
|
* is NULL. This prevents domain attach/detatch from running
|
|
* concurrently with cleaning up the area.
|
|
*/
|
|
again:
|
|
down_read(&iopt->domains_rwsem);
|
|
down_write(&iopt->iova_rwsem);
|
|
while ((area = iopt_area_iter_first(iopt, start, last))) {
|
|
unsigned long area_last = iopt_area_last_iova(area);
|
|
unsigned long area_first = iopt_area_iova(area);
|
|
struct iopt_pages *pages;
|
|
|
|
/* Userspace should not race map/unmap's of the same area */
|
|
if (!area->pages) {
|
|
rc = -EBUSY;
|
|
goto out_unlock_iova;
|
|
}
|
|
|
|
if (area_first < start || area_last > last) {
|
|
rc = -ENOENT;
|
|
goto out_unlock_iova;
|
|
}
|
|
|
|
if (area_first != start)
|
|
tries = 0;
|
|
|
|
/*
|
|
* num_accesses writers must hold the iova_rwsem too, so we can
|
|
* safely read it under the write side of the iovam_rwsem
|
|
* without the pages->mutex.
|
|
*/
|
|
if (area->num_accesses) {
|
|
size_t length = iopt_area_length(area);
|
|
|
|
start = area_first;
|
|
area->prevent_access = true;
|
|
up_write(&iopt->iova_rwsem);
|
|
up_read(&iopt->domains_rwsem);
|
|
|
|
iommufd_access_notify_unmap(iopt, area_first, length);
|
|
/* Something is not responding to unmap requests. */
|
|
tries++;
|
|
if (WARN_ON(tries > 100))
|
|
return -EDEADLOCK;
|
|
goto again;
|
|
}
|
|
|
|
pages = area->pages;
|
|
area->pages = NULL;
|
|
up_write(&iopt->iova_rwsem);
|
|
|
|
iopt_area_unfill_domains(area, pages);
|
|
iopt_abort_area(area);
|
|
iopt_put_pages(pages);
|
|
|
|
unmapped_bytes += area_last - area_first + 1;
|
|
|
|
down_write(&iopt->iova_rwsem);
|
|
}
|
|
if (unmapped_bytes)
|
|
rc = 0;
|
|
|
|
out_unlock_iova:
|
|
up_write(&iopt->iova_rwsem);
|
|
up_read(&iopt->domains_rwsem);
|
|
if (unmapped)
|
|
*unmapped = unmapped_bytes;
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* iopt_unmap_iova() - Remove a range of iova
|
|
* @iopt: io_pagetable to act on
|
|
* @iova: Starting iova to unmap
|
|
* @length: Number of bytes to unmap
|
|
* @unmapped: Return number of bytes unmapped
|
|
*
|
|
* The requested range must be a superset of existing ranges.
|
|
* Splitting/truncating IOVA mappings is not allowed.
|
|
*/
|
|
int iopt_unmap_iova(struct io_pagetable *iopt, unsigned long iova,
|
|
unsigned long length, unsigned long *unmapped)
|
|
{
|
|
unsigned long iova_last;
|
|
|
|
if (!length)
|
|
return -EINVAL;
|
|
|
|
if (check_add_overflow(iova, length - 1, &iova_last))
|
|
return -EOVERFLOW;
|
|
|
|
return iopt_unmap_iova_range(iopt, iova, iova_last, unmapped);
|
|
}
|
|
|
|
int iopt_unmap_all(struct io_pagetable *iopt, unsigned long *unmapped)
|
|
{
|
|
int rc;
|
|
|
|
rc = iopt_unmap_iova_range(iopt, 0, ULONG_MAX, unmapped);
|
|
/* If the IOVAs are empty then unmap all succeeds */
|
|
if (rc == -ENOENT)
|
|
return 0;
|
|
return rc;
|
|
}
|
|
|
|
/* The caller must always free all the nodes in the allowed_iova rb_root. */
|
|
int iopt_set_allow_iova(struct io_pagetable *iopt,
|
|
struct rb_root_cached *allowed_iova)
|
|
{
|
|
struct iopt_allowed *allowed;
|
|
|
|
down_write(&iopt->iova_rwsem);
|
|
swap(*allowed_iova, iopt->allowed_itree);
|
|
|
|
for (allowed = iopt_allowed_iter_first(iopt, 0, ULONG_MAX); allowed;
|
|
allowed = iopt_allowed_iter_next(allowed, 0, ULONG_MAX)) {
|
|
if (iopt_reserved_iter_first(iopt, allowed->node.start,
|
|
allowed->node.last)) {
|
|
swap(*allowed_iova, iopt->allowed_itree);
|
|
up_write(&iopt->iova_rwsem);
|
|
return -EADDRINUSE;
|
|
}
|
|
}
|
|
up_write(&iopt->iova_rwsem);
|
|
return 0;
|
|
}
|
|
|
|
int iopt_reserve_iova(struct io_pagetable *iopt, unsigned long start,
|
|
unsigned long last, void *owner)
|
|
{
|
|
struct iopt_reserved *reserved;
|
|
|
|
lockdep_assert_held_write(&iopt->iova_rwsem);
|
|
|
|
if (iopt_area_iter_first(iopt, start, last) ||
|
|
iopt_allowed_iter_first(iopt, start, last))
|
|
return -EADDRINUSE;
|
|
|
|
reserved = kzalloc(sizeof(*reserved), GFP_KERNEL_ACCOUNT);
|
|
if (!reserved)
|
|
return -ENOMEM;
|
|
reserved->node.start = start;
|
|
reserved->node.last = last;
|
|
reserved->owner = owner;
|
|
interval_tree_insert(&reserved->node, &iopt->reserved_itree);
|
|
return 0;
|
|
}
|
|
|
|
static void __iopt_remove_reserved_iova(struct io_pagetable *iopt, void *owner)
|
|
{
|
|
struct iopt_reserved *reserved, *next;
|
|
|
|
lockdep_assert_held_write(&iopt->iova_rwsem);
|
|
|
|
for (reserved = iopt_reserved_iter_first(iopt, 0, ULONG_MAX); reserved;
|
|
reserved = next) {
|
|
next = iopt_reserved_iter_next(reserved, 0, ULONG_MAX);
|
|
|
|
if (reserved->owner == owner) {
|
|
interval_tree_remove(&reserved->node,
|
|
&iopt->reserved_itree);
|
|
kfree(reserved);
|
|
}
|
|
}
|
|
}
|
|
|
|
void iopt_remove_reserved_iova(struct io_pagetable *iopt, void *owner)
|
|
{
|
|
down_write(&iopt->iova_rwsem);
|
|
__iopt_remove_reserved_iova(iopt, owner);
|
|
up_write(&iopt->iova_rwsem);
|
|
}
|
|
|
|
void iopt_init_table(struct io_pagetable *iopt)
|
|
{
|
|
init_rwsem(&iopt->iova_rwsem);
|
|
init_rwsem(&iopt->domains_rwsem);
|
|
iopt->area_itree = RB_ROOT_CACHED;
|
|
iopt->allowed_itree = RB_ROOT_CACHED;
|
|
iopt->reserved_itree = RB_ROOT_CACHED;
|
|
xa_init_flags(&iopt->domains, XA_FLAGS_ACCOUNT);
|
|
xa_init_flags(&iopt->access_list, XA_FLAGS_ALLOC);
|
|
|
|
/*
|
|
* iopt's start as SW tables that can use the entire size_t IOVA space
|
|
* due to the use of size_t in the APIs. They have no alignment
|
|
* restriction.
|
|
*/
|
|
iopt->iova_alignment = 1;
|
|
}
|
|
|
|
void iopt_destroy_table(struct io_pagetable *iopt)
|
|
{
|
|
struct interval_tree_node *node;
|
|
|
|
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
|
|
iopt_remove_reserved_iova(iopt, NULL);
|
|
|
|
while ((node = interval_tree_iter_first(&iopt->allowed_itree, 0,
|
|
ULONG_MAX))) {
|
|
interval_tree_remove(node, &iopt->allowed_itree);
|
|
kfree(container_of(node, struct iopt_allowed, node));
|
|
}
|
|
|
|
WARN_ON(!RB_EMPTY_ROOT(&iopt->reserved_itree.rb_root));
|
|
WARN_ON(!xa_empty(&iopt->domains));
|
|
WARN_ON(!xa_empty(&iopt->access_list));
|
|
WARN_ON(!RB_EMPTY_ROOT(&iopt->area_itree.rb_root));
|
|
}
|
|
|
|
/**
|
|
* iopt_unfill_domain() - Unfill a domain with PFNs
|
|
* @iopt: io_pagetable to act on
|
|
* @domain: domain to unfill
|
|
*
|
|
* This is used when removing a domain from the iopt. Every area in the iopt
|
|
* will be unmapped from the domain. The domain must already be removed from the
|
|
* domains xarray.
|
|
*/
|
|
static void iopt_unfill_domain(struct io_pagetable *iopt,
|
|
struct iommu_domain *domain)
|
|
{
|
|
struct iopt_area *area;
|
|
|
|
lockdep_assert_held(&iopt->iova_rwsem);
|
|
lockdep_assert_held_write(&iopt->domains_rwsem);
|
|
|
|
/*
|
|
* Some other domain is holding all the pfns still, rapidly unmap this
|
|
* domain.
|
|
*/
|
|
if (iopt->next_domain_id != 0) {
|
|
/* Pick an arbitrary remaining domain to act as storage */
|
|
struct iommu_domain *storage_domain =
|
|
xa_load(&iopt->domains, 0);
|
|
|
|
for (area = iopt_area_iter_first(iopt, 0, ULONG_MAX); area;
|
|
area = iopt_area_iter_next(area, 0, ULONG_MAX)) {
|
|
struct iopt_pages *pages = area->pages;
|
|
|
|
if (!pages)
|
|
continue;
|
|
|
|
mutex_lock(&pages->mutex);
|
|
if (IS_ENABLED(CONFIG_IOMMUFD_TEST))
|
|
WARN_ON(!area->storage_domain);
|
|
if (area->storage_domain == domain)
|
|
area->storage_domain = storage_domain;
|
|
mutex_unlock(&pages->mutex);
|
|
|
|
iopt_area_unmap_domain(area, domain);
|
|
}
|
|
return;
|
|
}
|
|
|
|
for (area = iopt_area_iter_first(iopt, 0, ULONG_MAX); area;
|
|
area = iopt_area_iter_next(area, 0, ULONG_MAX)) {
|
|
struct iopt_pages *pages = area->pages;
|
|
|
|
if (!pages)
|
|
continue;
|
|
|
|
mutex_lock(&pages->mutex);
|
|
interval_tree_remove(&area->pages_node, &pages->domains_itree);
|
|
WARN_ON(area->storage_domain != domain);
|
|
area->storage_domain = NULL;
|
|
iopt_area_unfill_domain(area, pages, domain);
|
|
mutex_unlock(&pages->mutex);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* iopt_fill_domain() - Fill a domain with PFNs
|
|
* @iopt: io_pagetable to act on
|
|
* @domain: domain to fill
|
|
*
|
|
* Fill the domain with PFNs from every area in the iopt. On failure the domain
|
|
* is left unchanged.
|
|
*/
|
|
static int iopt_fill_domain(struct io_pagetable *iopt,
|
|
struct iommu_domain *domain)
|
|
{
|
|
struct iopt_area *end_area;
|
|
struct iopt_area *area;
|
|
int rc;
|
|
|
|
lockdep_assert_held(&iopt->iova_rwsem);
|
|
lockdep_assert_held_write(&iopt->domains_rwsem);
|
|
|
|
for (area = iopt_area_iter_first(iopt, 0, ULONG_MAX); area;
|
|
area = iopt_area_iter_next(area, 0, ULONG_MAX)) {
|
|
struct iopt_pages *pages = area->pages;
|
|
|
|
if (!pages)
|
|
continue;
|
|
|
|
mutex_lock(&pages->mutex);
|
|
rc = iopt_area_fill_domain(area, domain);
|
|
if (rc) {
|
|
mutex_unlock(&pages->mutex);
|
|
goto out_unfill;
|
|
}
|
|
if (!area->storage_domain) {
|
|
WARN_ON(iopt->next_domain_id != 0);
|
|
area->storage_domain = domain;
|
|
interval_tree_insert(&area->pages_node,
|
|
&pages->domains_itree);
|
|
}
|
|
mutex_unlock(&pages->mutex);
|
|
}
|
|
return 0;
|
|
|
|
out_unfill:
|
|
end_area = area;
|
|
for (area = iopt_area_iter_first(iopt, 0, ULONG_MAX); area;
|
|
area = iopt_area_iter_next(area, 0, ULONG_MAX)) {
|
|
struct iopt_pages *pages = area->pages;
|
|
|
|
if (area == end_area)
|
|
break;
|
|
if (!pages)
|
|
continue;
|
|
mutex_lock(&pages->mutex);
|
|
if (iopt->next_domain_id == 0) {
|
|
interval_tree_remove(&area->pages_node,
|
|
&pages->domains_itree);
|
|
area->storage_domain = NULL;
|
|
}
|
|
iopt_area_unfill_domain(area, pages, domain);
|
|
mutex_unlock(&pages->mutex);
|
|
}
|
|
return rc;
|
|
}
|
|
|
|
/* All existing area's conform to an increased page size */
|
|
static int iopt_check_iova_alignment(struct io_pagetable *iopt,
|
|
unsigned long new_iova_alignment)
|
|
{
|
|
unsigned long align_mask = new_iova_alignment - 1;
|
|
struct iopt_area *area;
|
|
|
|
lockdep_assert_held(&iopt->iova_rwsem);
|
|
lockdep_assert_held(&iopt->domains_rwsem);
|
|
|
|
for (area = iopt_area_iter_first(iopt, 0, ULONG_MAX); area;
|
|
area = iopt_area_iter_next(area, 0, ULONG_MAX))
|
|
if ((iopt_area_iova(area) & align_mask) ||
|
|
(iopt_area_length(area) & align_mask) ||
|
|
(area->page_offset & align_mask))
|
|
return -EADDRINUSE;
|
|
|
|
if (IS_ENABLED(CONFIG_IOMMUFD_TEST)) {
|
|
struct iommufd_access *access;
|
|
unsigned long index;
|
|
|
|
xa_for_each(&iopt->access_list, index, access)
|
|
if (WARN_ON(access->iova_alignment >
|
|
new_iova_alignment))
|
|
return -EADDRINUSE;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int iopt_table_add_domain(struct io_pagetable *iopt,
|
|
struct iommu_domain *domain)
|
|
{
|
|
const struct iommu_domain_geometry *geometry = &domain->geometry;
|
|
struct iommu_domain *iter_domain;
|
|
unsigned int new_iova_alignment;
|
|
unsigned long index;
|
|
int rc;
|
|
|
|
down_write(&iopt->domains_rwsem);
|
|
down_write(&iopt->iova_rwsem);
|
|
|
|
xa_for_each(&iopt->domains, index, iter_domain) {
|
|
if (WARN_ON(iter_domain == domain)) {
|
|
rc = -EEXIST;
|
|
goto out_unlock;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* The io page size drives the iova_alignment. Internally the iopt_pages
|
|
* works in PAGE_SIZE units and we adjust when mapping sub-PAGE_SIZE
|
|
* objects into the iommu_domain.
|
|
*
|
|
* A iommu_domain must always be able to accept PAGE_SIZE to be
|
|
* compatible as we can't guarantee higher contiguity.
|
|
*/
|
|
new_iova_alignment = max_t(unsigned long,
|
|
1UL << __ffs(domain->pgsize_bitmap),
|
|
iopt->iova_alignment);
|
|
if (new_iova_alignment > PAGE_SIZE) {
|
|
rc = -EINVAL;
|
|
goto out_unlock;
|
|
}
|
|
if (new_iova_alignment != iopt->iova_alignment) {
|
|
rc = iopt_check_iova_alignment(iopt, new_iova_alignment);
|
|
if (rc)
|
|
goto out_unlock;
|
|
}
|
|
|
|
/* No area exists that is outside the allowed domain aperture */
|
|
if (geometry->aperture_start != 0) {
|
|
rc = iopt_reserve_iova(iopt, 0, geometry->aperture_start - 1,
|
|
domain);
|
|
if (rc)
|
|
goto out_reserved;
|
|
}
|
|
if (geometry->aperture_end != ULONG_MAX) {
|
|
rc = iopt_reserve_iova(iopt, geometry->aperture_end + 1,
|
|
ULONG_MAX, domain);
|
|
if (rc)
|
|
goto out_reserved;
|
|
}
|
|
|
|
rc = xa_reserve(&iopt->domains, iopt->next_domain_id, GFP_KERNEL);
|
|
if (rc)
|
|
goto out_reserved;
|
|
|
|
rc = iopt_fill_domain(iopt, domain);
|
|
if (rc)
|
|
goto out_release;
|
|
|
|
iopt->iova_alignment = new_iova_alignment;
|
|
xa_store(&iopt->domains, iopt->next_domain_id, domain, GFP_KERNEL);
|
|
iopt->next_domain_id++;
|
|
up_write(&iopt->iova_rwsem);
|
|
up_write(&iopt->domains_rwsem);
|
|
return 0;
|
|
out_release:
|
|
xa_release(&iopt->domains, iopt->next_domain_id);
|
|
out_reserved:
|
|
__iopt_remove_reserved_iova(iopt, domain);
|
|
out_unlock:
|
|
up_write(&iopt->iova_rwsem);
|
|
up_write(&iopt->domains_rwsem);
|
|
return rc;
|
|
}
|
|
|
|
static int iopt_calculate_iova_alignment(struct io_pagetable *iopt)
|
|
{
|
|
unsigned long new_iova_alignment;
|
|
struct iommufd_access *access;
|
|
struct iommu_domain *domain;
|
|
unsigned long index;
|
|
|
|
lockdep_assert_held_write(&iopt->iova_rwsem);
|
|
lockdep_assert_held(&iopt->domains_rwsem);
|
|
|
|
/* See batch_iommu_map_small() */
|
|
if (iopt->disable_large_pages)
|
|
new_iova_alignment = PAGE_SIZE;
|
|
else
|
|
new_iova_alignment = 1;
|
|
|
|
xa_for_each(&iopt->domains, index, domain)
|
|
new_iova_alignment = max_t(unsigned long,
|
|
1UL << __ffs(domain->pgsize_bitmap),
|
|
new_iova_alignment);
|
|
xa_for_each(&iopt->access_list, index, access)
|
|
new_iova_alignment = max_t(unsigned long,
|
|
access->iova_alignment,
|
|
new_iova_alignment);
|
|
|
|
if (new_iova_alignment > iopt->iova_alignment) {
|
|
int rc;
|
|
|
|
rc = iopt_check_iova_alignment(iopt, new_iova_alignment);
|
|
if (rc)
|
|
return rc;
|
|
}
|
|
iopt->iova_alignment = new_iova_alignment;
|
|
return 0;
|
|
}
|
|
|
|
void iopt_table_remove_domain(struct io_pagetable *iopt,
|
|
struct iommu_domain *domain)
|
|
{
|
|
struct iommu_domain *iter_domain = NULL;
|
|
unsigned long index;
|
|
|
|
down_write(&iopt->domains_rwsem);
|
|
down_write(&iopt->iova_rwsem);
|
|
|
|
xa_for_each(&iopt->domains, index, iter_domain)
|
|
if (iter_domain == domain)
|
|
break;
|
|
if (WARN_ON(iter_domain != domain) || index >= iopt->next_domain_id)
|
|
goto out_unlock;
|
|
|
|
/*
|
|
* Compress the xarray to keep it linear by swapping the entry to erase
|
|
* with the tail entry and shrinking the tail.
|
|
*/
|
|
iopt->next_domain_id--;
|
|
iter_domain = xa_erase(&iopt->domains, iopt->next_domain_id);
|
|
if (index != iopt->next_domain_id)
|
|
xa_store(&iopt->domains, index, iter_domain, GFP_KERNEL);
|
|
|
|
iopt_unfill_domain(iopt, domain);
|
|
__iopt_remove_reserved_iova(iopt, domain);
|
|
|
|
WARN_ON(iopt_calculate_iova_alignment(iopt));
|
|
out_unlock:
|
|
up_write(&iopt->iova_rwsem);
|
|
up_write(&iopt->domains_rwsem);
|
|
}
|
|
|
|
/**
|
|
* iopt_area_split - Split an area into two parts at iova
|
|
* @area: The area to split
|
|
* @iova: Becomes the last of a new area
|
|
*
|
|
* This splits an area into two. It is part of the VFIO compatibility to allow
|
|
* poking a hole in the mapping. The two areas continue to point at the same
|
|
* iopt_pages, just with different starting bytes.
|
|
*/
|
|
static int iopt_area_split(struct iopt_area *area, unsigned long iova)
|
|
{
|
|
unsigned long alignment = area->iopt->iova_alignment;
|
|
unsigned long last_iova = iopt_area_last_iova(area);
|
|
unsigned long start_iova = iopt_area_iova(area);
|
|
unsigned long new_start = iova + 1;
|
|
struct io_pagetable *iopt = area->iopt;
|
|
struct iopt_pages *pages = area->pages;
|
|
struct iopt_area *lhs;
|
|
struct iopt_area *rhs;
|
|
int rc;
|
|
|
|
lockdep_assert_held_write(&iopt->iova_rwsem);
|
|
|
|
if (iova == start_iova || iova == last_iova)
|
|
return 0;
|
|
|
|
if (!pages || area->prevent_access)
|
|
return -EBUSY;
|
|
|
|
if (new_start & (alignment - 1) ||
|
|
iopt_area_start_byte(area, new_start) & (alignment - 1))
|
|
return -EINVAL;
|
|
|
|
lhs = iopt_area_alloc();
|
|
if (!lhs)
|
|
return -ENOMEM;
|
|
|
|
rhs = iopt_area_alloc();
|
|
if (!rhs) {
|
|
rc = -ENOMEM;
|
|
goto err_free_lhs;
|
|
}
|
|
|
|
mutex_lock(&pages->mutex);
|
|
/*
|
|
* Splitting is not permitted if an access exists, we don't track enough
|
|
* information to split existing accesses.
|
|
*/
|
|
if (area->num_accesses) {
|
|
rc = -EINVAL;
|
|
goto err_unlock;
|
|
}
|
|
|
|
/*
|
|
* Splitting is not permitted if a domain could have been mapped with
|
|
* huge pages.
|
|
*/
|
|
if (area->storage_domain && !iopt->disable_large_pages) {
|
|
rc = -EINVAL;
|
|
goto err_unlock;
|
|
}
|
|
|
|
interval_tree_remove(&area->node, &iopt->area_itree);
|
|
rc = iopt_insert_area(iopt, lhs, area->pages, start_iova,
|
|
iopt_area_start_byte(area, start_iova),
|
|
(new_start - 1) - start_iova + 1,
|
|
area->iommu_prot);
|
|
if (WARN_ON(rc))
|
|
goto err_insert;
|
|
|
|
rc = iopt_insert_area(iopt, rhs, area->pages, new_start,
|
|
iopt_area_start_byte(area, new_start),
|
|
last_iova - new_start + 1, area->iommu_prot);
|
|
if (WARN_ON(rc))
|
|
goto err_remove_lhs;
|
|
|
|
/*
|
|
* If the original area has filled a domain, domains_itree has to be
|
|
* updated.
|
|
*/
|
|
if (area->storage_domain) {
|
|
interval_tree_remove(&area->pages_node, &pages->domains_itree);
|
|
interval_tree_insert(&lhs->pages_node, &pages->domains_itree);
|
|
interval_tree_insert(&rhs->pages_node, &pages->domains_itree);
|
|
}
|
|
|
|
lhs->storage_domain = area->storage_domain;
|
|
lhs->pages = area->pages;
|
|
rhs->storage_domain = area->storage_domain;
|
|
rhs->pages = area->pages;
|
|
kref_get(&rhs->pages->kref);
|
|
kfree(area);
|
|
mutex_unlock(&pages->mutex);
|
|
|
|
/*
|
|
* No change to domains or accesses because the pages hasn't been
|
|
* changed
|
|
*/
|
|
return 0;
|
|
|
|
err_remove_lhs:
|
|
interval_tree_remove(&lhs->node, &iopt->area_itree);
|
|
err_insert:
|
|
interval_tree_insert(&area->node, &iopt->area_itree);
|
|
err_unlock:
|
|
mutex_unlock(&pages->mutex);
|
|
kfree(rhs);
|
|
err_free_lhs:
|
|
kfree(lhs);
|
|
return rc;
|
|
}
|
|
|
|
int iopt_cut_iova(struct io_pagetable *iopt, unsigned long *iovas,
|
|
size_t num_iovas)
|
|
{
|
|
int rc = 0;
|
|
int i;
|
|
|
|
down_write(&iopt->iova_rwsem);
|
|
for (i = 0; i < num_iovas; i++) {
|
|
struct iopt_area *area;
|
|
|
|
area = iopt_area_iter_first(iopt, iovas[i], iovas[i]);
|
|
if (!area)
|
|
continue;
|
|
rc = iopt_area_split(area, iovas[i]);
|
|
if (rc)
|
|
break;
|
|
}
|
|
up_write(&iopt->iova_rwsem);
|
|
return rc;
|
|
}
|
|
|
|
void iopt_enable_large_pages(struct io_pagetable *iopt)
|
|
{
|
|
int rc;
|
|
|
|
down_write(&iopt->domains_rwsem);
|
|
down_write(&iopt->iova_rwsem);
|
|
WRITE_ONCE(iopt->disable_large_pages, false);
|
|
rc = iopt_calculate_iova_alignment(iopt);
|
|
WARN_ON(rc);
|
|
up_write(&iopt->iova_rwsem);
|
|
up_write(&iopt->domains_rwsem);
|
|
}
|
|
|
|
int iopt_disable_large_pages(struct io_pagetable *iopt)
|
|
{
|
|
int rc = 0;
|
|
|
|
down_write(&iopt->domains_rwsem);
|
|
down_write(&iopt->iova_rwsem);
|
|
if (iopt->disable_large_pages)
|
|
goto out_unlock;
|
|
|
|
/* Won't do it if domains already have pages mapped in them */
|
|
if (!xa_empty(&iopt->domains) &&
|
|
!RB_EMPTY_ROOT(&iopt->area_itree.rb_root)) {
|
|
rc = -EINVAL;
|
|
goto out_unlock;
|
|
}
|
|
|
|
WRITE_ONCE(iopt->disable_large_pages, true);
|
|
rc = iopt_calculate_iova_alignment(iopt);
|
|
if (rc)
|
|
WRITE_ONCE(iopt->disable_large_pages, false);
|
|
out_unlock:
|
|
up_write(&iopt->iova_rwsem);
|
|
up_write(&iopt->domains_rwsem);
|
|
return rc;
|
|
}
|
|
|
|
int iopt_add_access(struct io_pagetable *iopt, struct iommufd_access *access)
|
|
{
|
|
u32 new_id;
|
|
int rc;
|
|
|
|
down_write(&iopt->domains_rwsem);
|
|
down_write(&iopt->iova_rwsem);
|
|
rc = xa_alloc(&iopt->access_list, &new_id, access, xa_limit_16b,
|
|
GFP_KERNEL_ACCOUNT);
|
|
|
|
if (rc)
|
|
goto out_unlock;
|
|
|
|
rc = iopt_calculate_iova_alignment(iopt);
|
|
if (rc) {
|
|
xa_erase(&iopt->access_list, new_id);
|
|
goto out_unlock;
|
|
}
|
|
access->iopt_access_list_id = new_id;
|
|
|
|
out_unlock:
|
|
up_write(&iopt->iova_rwsem);
|
|
up_write(&iopt->domains_rwsem);
|
|
return rc;
|
|
}
|
|
|
|
void iopt_remove_access(struct io_pagetable *iopt,
|
|
struct iommufd_access *access,
|
|
u32 iopt_access_list_id)
|
|
{
|
|
down_write(&iopt->domains_rwsem);
|
|
down_write(&iopt->iova_rwsem);
|
|
WARN_ON(xa_erase(&iopt->access_list, iopt_access_list_id) != access);
|
|
WARN_ON(iopt_calculate_iova_alignment(iopt));
|
|
up_write(&iopt->iova_rwsem);
|
|
up_write(&iopt->domains_rwsem);
|
|
}
|
|
|
|
/* Narrow the valid_iova_itree to include reserved ranges from a device. */
|
|
int iopt_table_enforce_dev_resv_regions(struct io_pagetable *iopt,
|
|
struct device *dev,
|
|
phys_addr_t *sw_msi_start)
|
|
{
|
|
struct iommu_resv_region *resv;
|
|
LIST_HEAD(resv_regions);
|
|
unsigned int num_hw_msi = 0;
|
|
unsigned int num_sw_msi = 0;
|
|
int rc;
|
|
|
|
if (iommufd_should_fail())
|
|
return -EINVAL;
|
|
|
|
down_write(&iopt->iova_rwsem);
|
|
/* FIXME: drivers allocate memory but there is no failure propogated */
|
|
iommu_get_resv_regions(dev, &resv_regions);
|
|
|
|
list_for_each_entry(resv, &resv_regions, list) {
|
|
if (resv->type == IOMMU_RESV_DIRECT_RELAXABLE)
|
|
continue;
|
|
|
|
if (sw_msi_start && resv->type == IOMMU_RESV_MSI)
|
|
num_hw_msi++;
|
|
if (sw_msi_start && resv->type == IOMMU_RESV_SW_MSI) {
|
|
*sw_msi_start = resv->start;
|
|
num_sw_msi++;
|
|
}
|
|
|
|
rc = iopt_reserve_iova(iopt, resv->start,
|
|
resv->length - 1 + resv->start, dev);
|
|
if (rc)
|
|
goto out_reserved;
|
|
}
|
|
|
|
/* Drivers must offer sane combinations of regions */
|
|
if (WARN_ON(num_sw_msi && num_hw_msi) || WARN_ON(num_sw_msi > 1)) {
|
|
rc = -EINVAL;
|
|
goto out_reserved;
|
|
}
|
|
|
|
rc = 0;
|
|
goto out_free_resv;
|
|
|
|
out_reserved:
|
|
__iopt_remove_reserved_iova(iopt, dev);
|
|
out_free_resv:
|
|
iommu_put_resv_regions(dev, &resv_regions);
|
|
up_write(&iopt->iova_rwsem);
|
|
return rc;
|
|
}
|