407 lines
		
	
	
		
			9.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			407 lines
		
	
	
		
			9.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
// SPDX-License-Identifier: GPL-2.0-only
 | 
						|
/*
 | 
						|
 * Copyright (c) 2021-2022, NVIDIA CORPORATION & AFFILIATES
 | 
						|
 */
 | 
						|
#include <linux/interval_tree.h>
 | 
						|
#include <linux/iommu.h>
 | 
						|
#include <linux/iommufd.h>
 | 
						|
#include <uapi/linux/iommufd.h>
 | 
						|
 | 
						|
#include "io_pagetable.h"
 | 
						|
 | 
						|
void iommufd_ioas_destroy(struct iommufd_object *obj)
 | 
						|
{
 | 
						|
	struct iommufd_ioas *ioas = container_of(obj, struct iommufd_ioas, obj);
 | 
						|
	int rc;
 | 
						|
 | 
						|
	rc = iopt_unmap_all(&ioas->iopt, NULL);
 | 
						|
	WARN_ON(rc && rc != -ENOENT);
 | 
						|
	iopt_destroy_table(&ioas->iopt);
 | 
						|
	mutex_destroy(&ioas->mutex);
 | 
						|
}
 | 
						|
 | 
						|
struct iommufd_ioas *iommufd_ioas_alloc(struct iommufd_ctx *ictx)
 | 
						|
{
 | 
						|
	struct iommufd_ioas *ioas;
 | 
						|
 | 
						|
	ioas = iommufd_object_alloc(ictx, ioas, IOMMUFD_OBJ_IOAS);
 | 
						|
	if (IS_ERR(ioas))
 | 
						|
		return ioas;
 | 
						|
 | 
						|
	iopt_init_table(&ioas->iopt);
 | 
						|
	INIT_LIST_HEAD(&ioas->hwpt_list);
 | 
						|
	mutex_init(&ioas->mutex);
 | 
						|
	return ioas;
 | 
						|
}
 | 
						|
 | 
						|
int iommufd_ioas_alloc_ioctl(struct iommufd_ucmd *ucmd)
 | 
						|
{
 | 
						|
	struct iommu_ioas_alloc *cmd = ucmd->cmd;
 | 
						|
	struct iommufd_ioas *ioas;
 | 
						|
	int rc;
 | 
						|
 | 
						|
	if (cmd->flags)
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
 | 
						|
	ioas = iommufd_ioas_alloc(ucmd->ictx);
 | 
						|
	if (IS_ERR(ioas))
 | 
						|
		return PTR_ERR(ioas);
 | 
						|
 | 
						|
	cmd->out_ioas_id = ioas->obj.id;
 | 
						|
	rc = iommufd_ucmd_respond(ucmd, sizeof(*cmd));
 | 
						|
	if (rc)
 | 
						|
		goto out_table;
 | 
						|
	iommufd_object_finalize(ucmd->ictx, &ioas->obj);
 | 
						|
	return 0;
 | 
						|
 | 
						|
out_table:
 | 
						|
	iommufd_object_abort_and_destroy(ucmd->ictx, &ioas->obj);
 | 
						|
	return rc;
 | 
						|
}
 | 
						|
 | 
						|
int iommufd_ioas_iova_ranges(struct iommufd_ucmd *ucmd)
 | 
						|
{
 | 
						|
	struct iommu_iova_range __user *ranges;
 | 
						|
	struct iommu_ioas_iova_ranges *cmd = ucmd->cmd;
 | 
						|
	struct iommufd_ioas *ioas;
 | 
						|
	struct interval_tree_span_iter span;
 | 
						|
	u32 max_iovas;
 | 
						|
	int rc;
 | 
						|
 | 
						|
	if (cmd->__reserved)
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
 | 
						|
	ioas = iommufd_get_ioas(ucmd->ictx, cmd->ioas_id);
 | 
						|
	if (IS_ERR(ioas))
 | 
						|
		return PTR_ERR(ioas);
 | 
						|
 | 
						|
	down_read(&ioas->iopt.iova_rwsem);
 | 
						|
	max_iovas = cmd->num_iovas;
 | 
						|
	ranges = u64_to_user_ptr(cmd->allowed_iovas);
 | 
						|
	cmd->num_iovas = 0;
 | 
						|
	cmd->out_iova_alignment = ioas->iopt.iova_alignment;
 | 
						|
	interval_tree_for_each_span(&span, &ioas->iopt.reserved_itree, 0,
 | 
						|
				    ULONG_MAX) {
 | 
						|
		if (!span.is_hole)
 | 
						|
			continue;
 | 
						|
		if (cmd->num_iovas < max_iovas) {
 | 
						|
			struct iommu_iova_range elm = {
 | 
						|
				.start = span.start_hole,
 | 
						|
				.last = span.last_hole,
 | 
						|
			};
 | 
						|
 | 
						|
			if (copy_to_user(&ranges[cmd->num_iovas], &elm,
 | 
						|
					 sizeof(elm))) {
 | 
						|
				rc = -EFAULT;
 | 
						|
				goto out_put;
 | 
						|
			}
 | 
						|
		}
 | 
						|
		cmd->num_iovas++;
 | 
						|
	}
 | 
						|
	rc = iommufd_ucmd_respond(ucmd, sizeof(*cmd));
 | 
						|
	if (rc)
 | 
						|
		goto out_put;
 | 
						|
	if (cmd->num_iovas > max_iovas)
 | 
						|
		rc = -EMSGSIZE;
 | 
						|
out_put:
 | 
						|
	up_read(&ioas->iopt.iova_rwsem);
 | 
						|
	iommufd_put_object(ucmd->ictx, &ioas->obj);
 | 
						|
	return rc;
 | 
						|
}
 | 
						|
 | 
						|
static int iommufd_ioas_load_iovas(struct rb_root_cached *itree,
 | 
						|
				   struct iommu_iova_range __user *ranges,
 | 
						|
				   u32 num)
 | 
						|
{
 | 
						|
	u32 i;
 | 
						|
 | 
						|
	for (i = 0; i != num; i++) {
 | 
						|
		struct iommu_iova_range range;
 | 
						|
		struct iopt_allowed *allowed;
 | 
						|
 | 
						|
		if (copy_from_user(&range, ranges + i, sizeof(range)))
 | 
						|
			return -EFAULT;
 | 
						|
 | 
						|
		if (range.start >= range.last)
 | 
						|
			return -EINVAL;
 | 
						|
 | 
						|
		if (interval_tree_iter_first(itree, range.start, range.last))
 | 
						|
			return -EINVAL;
 | 
						|
 | 
						|
		allowed = kzalloc(sizeof(*allowed), GFP_KERNEL_ACCOUNT);
 | 
						|
		if (!allowed)
 | 
						|
			return -ENOMEM;
 | 
						|
		allowed->node.start = range.start;
 | 
						|
		allowed->node.last = range.last;
 | 
						|
 | 
						|
		interval_tree_insert(&allowed->node, itree);
 | 
						|
	}
 | 
						|
	return 0;
 | 
						|
}
 | 
						|
 | 
						|
int iommufd_ioas_allow_iovas(struct iommufd_ucmd *ucmd)
 | 
						|
{
 | 
						|
	struct iommu_ioas_allow_iovas *cmd = ucmd->cmd;
 | 
						|
	struct rb_root_cached allowed_iova = RB_ROOT_CACHED;
 | 
						|
	struct interval_tree_node *node;
 | 
						|
	struct iommufd_ioas *ioas;
 | 
						|
	struct io_pagetable *iopt;
 | 
						|
	int rc = 0;
 | 
						|
 | 
						|
	if (cmd->__reserved)
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
 | 
						|
	ioas = iommufd_get_ioas(ucmd->ictx, cmd->ioas_id);
 | 
						|
	if (IS_ERR(ioas))
 | 
						|
		return PTR_ERR(ioas);
 | 
						|
	iopt = &ioas->iopt;
 | 
						|
 | 
						|
	rc = iommufd_ioas_load_iovas(&allowed_iova,
 | 
						|
				     u64_to_user_ptr(cmd->allowed_iovas),
 | 
						|
				     cmd->num_iovas);
 | 
						|
	if (rc)
 | 
						|
		goto out_free;
 | 
						|
 | 
						|
	/*
 | 
						|
	 * We want the allowed tree update to be atomic, so we have to keep the
 | 
						|
	 * original nodes around, and keep track of the new nodes as we allocate
 | 
						|
	 * memory for them. The simplest solution is to have a new/old tree and
 | 
						|
	 * then swap new for old. On success we free the old tree, on failure we
 | 
						|
	 * free the new tree.
 | 
						|
	 */
 | 
						|
	rc = iopt_set_allow_iova(iopt, &allowed_iova);
 | 
						|
out_free:
 | 
						|
	while ((node = interval_tree_iter_first(&allowed_iova, 0, ULONG_MAX))) {
 | 
						|
		interval_tree_remove(node, &allowed_iova);
 | 
						|
		kfree(container_of(node, struct iopt_allowed, node));
 | 
						|
	}
 | 
						|
	iommufd_put_object(ucmd->ictx, &ioas->obj);
 | 
						|
	return rc;
 | 
						|
}
 | 
						|
 | 
						|
static int conv_iommu_prot(u32 map_flags)
 | 
						|
{
 | 
						|
	/*
 | 
						|
	 * We provide no manual cache coherency ioctls to userspace and most
 | 
						|
	 * architectures make the CPU ops for cache flushing privileged.
 | 
						|
	 * Therefore we require the underlying IOMMU to support CPU coherent
 | 
						|
	 * operation. Support for IOMMU_CACHE is enforced by the
 | 
						|
	 * IOMMU_CAP_CACHE_COHERENCY test during bind.
 | 
						|
	 */
 | 
						|
	int iommu_prot = IOMMU_CACHE;
 | 
						|
 | 
						|
	if (map_flags & IOMMU_IOAS_MAP_WRITEABLE)
 | 
						|
		iommu_prot |= IOMMU_WRITE;
 | 
						|
	if (map_flags & IOMMU_IOAS_MAP_READABLE)
 | 
						|
		iommu_prot |= IOMMU_READ;
 | 
						|
	return iommu_prot;
 | 
						|
}
 | 
						|
 | 
						|
int iommufd_ioas_map(struct iommufd_ucmd *ucmd)
 | 
						|
{
 | 
						|
	struct iommu_ioas_map *cmd = ucmd->cmd;
 | 
						|
	unsigned long iova = cmd->iova;
 | 
						|
	struct iommufd_ioas *ioas;
 | 
						|
	unsigned int flags = 0;
 | 
						|
	int rc;
 | 
						|
 | 
						|
	if ((cmd->flags &
 | 
						|
	     ~(IOMMU_IOAS_MAP_FIXED_IOVA | IOMMU_IOAS_MAP_WRITEABLE |
 | 
						|
	       IOMMU_IOAS_MAP_READABLE)) ||
 | 
						|
	    cmd->__reserved)
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
	if (cmd->iova >= ULONG_MAX || cmd->length >= ULONG_MAX)
 | 
						|
		return -EOVERFLOW;
 | 
						|
 | 
						|
	if (!(cmd->flags &
 | 
						|
	      (IOMMU_IOAS_MAP_WRITEABLE | IOMMU_IOAS_MAP_READABLE)))
 | 
						|
		return -EINVAL;
 | 
						|
 | 
						|
	ioas = iommufd_get_ioas(ucmd->ictx, cmd->ioas_id);
 | 
						|
	if (IS_ERR(ioas))
 | 
						|
		return PTR_ERR(ioas);
 | 
						|
 | 
						|
	if (!(cmd->flags & IOMMU_IOAS_MAP_FIXED_IOVA))
 | 
						|
		flags = IOPT_ALLOC_IOVA;
 | 
						|
	rc = iopt_map_user_pages(ucmd->ictx, &ioas->iopt, &iova,
 | 
						|
				 u64_to_user_ptr(cmd->user_va), cmd->length,
 | 
						|
				 conv_iommu_prot(cmd->flags), flags);
 | 
						|
	if (rc)
 | 
						|
		goto out_put;
 | 
						|
 | 
						|
	cmd->iova = iova;
 | 
						|
	rc = iommufd_ucmd_respond(ucmd, sizeof(*cmd));
 | 
						|
out_put:
 | 
						|
	iommufd_put_object(ucmd->ictx, &ioas->obj);
 | 
						|
	return rc;
 | 
						|
}
 | 
						|
 | 
						|
int iommufd_ioas_copy(struct iommufd_ucmd *ucmd)
 | 
						|
{
 | 
						|
	struct iommu_ioas_copy *cmd = ucmd->cmd;
 | 
						|
	struct iommufd_ioas *src_ioas;
 | 
						|
	struct iommufd_ioas *dst_ioas;
 | 
						|
	unsigned int flags = 0;
 | 
						|
	LIST_HEAD(pages_list);
 | 
						|
	unsigned long iova;
 | 
						|
	int rc;
 | 
						|
 | 
						|
	iommufd_test_syz_conv_iova_id(ucmd, cmd->src_ioas_id, &cmd->src_iova,
 | 
						|
				      &cmd->flags);
 | 
						|
 | 
						|
	if ((cmd->flags &
 | 
						|
	     ~(IOMMU_IOAS_MAP_FIXED_IOVA | IOMMU_IOAS_MAP_WRITEABLE |
 | 
						|
	       IOMMU_IOAS_MAP_READABLE)))
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
	if (cmd->length >= ULONG_MAX || cmd->src_iova >= ULONG_MAX ||
 | 
						|
	    cmd->dst_iova >= ULONG_MAX)
 | 
						|
		return -EOVERFLOW;
 | 
						|
 | 
						|
	if (!(cmd->flags &
 | 
						|
	      (IOMMU_IOAS_MAP_WRITEABLE | IOMMU_IOAS_MAP_READABLE)))
 | 
						|
		return -EINVAL;
 | 
						|
 | 
						|
	src_ioas = iommufd_get_ioas(ucmd->ictx, cmd->src_ioas_id);
 | 
						|
	if (IS_ERR(src_ioas))
 | 
						|
		return PTR_ERR(src_ioas);
 | 
						|
	rc = iopt_get_pages(&src_ioas->iopt, cmd->src_iova, cmd->length,
 | 
						|
			    &pages_list);
 | 
						|
	iommufd_put_object(ucmd->ictx, &src_ioas->obj);
 | 
						|
	if (rc)
 | 
						|
		return rc;
 | 
						|
 | 
						|
	dst_ioas = iommufd_get_ioas(ucmd->ictx, cmd->dst_ioas_id);
 | 
						|
	if (IS_ERR(dst_ioas)) {
 | 
						|
		rc = PTR_ERR(dst_ioas);
 | 
						|
		goto out_pages;
 | 
						|
	}
 | 
						|
 | 
						|
	if (!(cmd->flags & IOMMU_IOAS_MAP_FIXED_IOVA))
 | 
						|
		flags = IOPT_ALLOC_IOVA;
 | 
						|
	iova = cmd->dst_iova;
 | 
						|
	rc = iopt_map_pages(&dst_ioas->iopt, &pages_list, cmd->length, &iova,
 | 
						|
			    conv_iommu_prot(cmd->flags), flags);
 | 
						|
	if (rc)
 | 
						|
		goto out_put_dst;
 | 
						|
 | 
						|
	cmd->dst_iova = iova;
 | 
						|
	rc = iommufd_ucmd_respond(ucmd, sizeof(*cmd));
 | 
						|
out_put_dst:
 | 
						|
	iommufd_put_object(ucmd->ictx, &dst_ioas->obj);
 | 
						|
out_pages:
 | 
						|
	iopt_free_pages_list(&pages_list);
 | 
						|
	return rc;
 | 
						|
}
 | 
						|
 | 
						|
int iommufd_ioas_unmap(struct iommufd_ucmd *ucmd)
 | 
						|
{
 | 
						|
	struct iommu_ioas_unmap *cmd = ucmd->cmd;
 | 
						|
	struct iommufd_ioas *ioas;
 | 
						|
	unsigned long unmapped = 0;
 | 
						|
	int rc;
 | 
						|
 | 
						|
	ioas = iommufd_get_ioas(ucmd->ictx, cmd->ioas_id);
 | 
						|
	if (IS_ERR(ioas))
 | 
						|
		return PTR_ERR(ioas);
 | 
						|
 | 
						|
	if (cmd->iova == 0 && cmd->length == U64_MAX) {
 | 
						|
		rc = iopt_unmap_all(&ioas->iopt, &unmapped);
 | 
						|
		if (rc)
 | 
						|
			goto out_put;
 | 
						|
	} else {
 | 
						|
		if (cmd->iova >= ULONG_MAX || cmd->length >= ULONG_MAX) {
 | 
						|
			rc = -EOVERFLOW;
 | 
						|
			goto out_put;
 | 
						|
		}
 | 
						|
		rc = iopt_unmap_iova(&ioas->iopt, cmd->iova, cmd->length,
 | 
						|
				     &unmapped);
 | 
						|
		if (rc)
 | 
						|
			goto out_put;
 | 
						|
	}
 | 
						|
 | 
						|
	cmd->length = unmapped;
 | 
						|
	rc = iommufd_ucmd_respond(ucmd, sizeof(*cmd));
 | 
						|
 | 
						|
out_put:
 | 
						|
	iommufd_put_object(ucmd->ictx, &ioas->obj);
 | 
						|
	return rc;
 | 
						|
}
 | 
						|
 | 
						|
int iommufd_option_rlimit_mode(struct iommu_option *cmd,
 | 
						|
			       struct iommufd_ctx *ictx)
 | 
						|
{
 | 
						|
	if (cmd->object_id)
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
 | 
						|
	if (cmd->op == IOMMU_OPTION_OP_GET) {
 | 
						|
		cmd->val64 = ictx->account_mode == IOPT_PAGES_ACCOUNT_MM;
 | 
						|
		return 0;
 | 
						|
	}
 | 
						|
	if (cmd->op == IOMMU_OPTION_OP_SET) {
 | 
						|
		int rc = 0;
 | 
						|
 | 
						|
		if (!capable(CAP_SYS_RESOURCE))
 | 
						|
			return -EPERM;
 | 
						|
 | 
						|
		xa_lock(&ictx->objects);
 | 
						|
		if (!xa_empty(&ictx->objects)) {
 | 
						|
			rc = -EBUSY;
 | 
						|
		} else {
 | 
						|
			if (cmd->val64 == 0)
 | 
						|
				ictx->account_mode = IOPT_PAGES_ACCOUNT_USER;
 | 
						|
			else if (cmd->val64 == 1)
 | 
						|
				ictx->account_mode = IOPT_PAGES_ACCOUNT_MM;
 | 
						|
			else
 | 
						|
				rc = -EINVAL;
 | 
						|
		}
 | 
						|
		xa_unlock(&ictx->objects);
 | 
						|
 | 
						|
		return rc;
 | 
						|
	}
 | 
						|
	return -EOPNOTSUPP;
 | 
						|
}
 | 
						|
 | 
						|
static int iommufd_ioas_option_huge_pages(struct iommu_option *cmd,
 | 
						|
					  struct iommufd_ioas *ioas)
 | 
						|
{
 | 
						|
	if (cmd->op == IOMMU_OPTION_OP_GET) {
 | 
						|
		cmd->val64 = !ioas->iopt.disable_large_pages;
 | 
						|
		return 0;
 | 
						|
	}
 | 
						|
	if (cmd->op == IOMMU_OPTION_OP_SET) {
 | 
						|
		if (cmd->val64 == 0)
 | 
						|
			return iopt_disable_large_pages(&ioas->iopt);
 | 
						|
		if (cmd->val64 == 1) {
 | 
						|
			iopt_enable_large_pages(&ioas->iopt);
 | 
						|
			return 0;
 | 
						|
		}
 | 
						|
		return -EINVAL;
 | 
						|
	}
 | 
						|
	return -EOPNOTSUPP;
 | 
						|
}
 | 
						|
 | 
						|
int iommufd_ioas_option(struct iommufd_ucmd *ucmd)
 | 
						|
{
 | 
						|
	struct iommu_option *cmd = ucmd->cmd;
 | 
						|
	struct iommufd_ioas *ioas;
 | 
						|
	int rc = 0;
 | 
						|
 | 
						|
	if (cmd->__reserved)
 | 
						|
		return -EOPNOTSUPP;
 | 
						|
 | 
						|
	ioas = iommufd_get_ioas(ucmd->ictx, cmd->object_id);
 | 
						|
	if (IS_ERR(ioas))
 | 
						|
		return PTR_ERR(ioas);
 | 
						|
 | 
						|
	switch (cmd->option_id) {
 | 
						|
	case IOMMU_OPTION_HUGE_PAGES:
 | 
						|
		rc = iommufd_ioas_option_huge_pages(cmd, ioas);
 | 
						|
		break;
 | 
						|
	default:
 | 
						|
		rc = -EOPNOTSUPP;
 | 
						|
	}
 | 
						|
 | 
						|
	iommufd_put_object(ucmd->ictx, &ioas->obj);
 | 
						|
	return rc;
 | 
						|
}
 |