diff options
Diffstat (limited to 'drivers/iommu/iommufd/main.c')
| -rw-r--r-- | drivers/iommu/iommufd/main.c | 460 | 
1 files changed, 460 insertions, 0 deletions
diff --git a/drivers/iommu/iommufd/main.c b/drivers/iommu/iommufd/main.c new file mode 100644 index 000000000000..083e6fcbe10a --- /dev/null +++ b/drivers/iommu/iommufd/main.c @@ -0,0 +1,460 @@ +// SPDX-License-Identifier: GPL-2.0-only +/* Copyright (C) 2021 Intel Corporation + * Copyright (c) 2021-2022, NVIDIA CORPORATION & AFFILIATES + * + * iommufd provides control over the IOMMU HW objects created by IOMMU kernel + * drivers. IOMMU HW objects revolve around IO page tables that map incoming DMA + * addresses (IOVA) to CPU addresses. + */ +#define pr_fmt(fmt) "iommufd: " fmt + +#include <linux/file.h> +#include <linux/fs.h> +#include <linux/module.h> +#include <linux/slab.h> +#include <linux/miscdevice.h> +#include <linux/mutex.h> +#include <linux/bug.h> +#include <uapi/linux/iommufd.h> +#include <linux/iommufd.h> + +#include "io_pagetable.h" +#include "iommufd_private.h" +#include "iommufd_test.h" + +struct iommufd_object_ops { +	void (*destroy)(struct iommufd_object *obj); +}; +static const struct iommufd_object_ops iommufd_object_ops[]; +static struct miscdevice vfio_misc_dev; + +struct iommufd_object *_iommufd_object_alloc(struct iommufd_ctx *ictx, +					     size_t size, +					     enum iommufd_object_type type) +{ +	struct iommufd_object *obj; +	int rc; + +	obj = kzalloc(size, GFP_KERNEL_ACCOUNT); +	if (!obj) +		return ERR_PTR(-ENOMEM); +	obj->type = type; +	init_rwsem(&obj->destroy_rwsem); +	refcount_set(&obj->users, 1); + +	/* +	 * Reserve an ID in the xarray but do not publish the pointer yet since +	 * the caller hasn't initialized it yet. Once the pointer is published +	 * in the xarray and visible to other threads we can't reliably destroy +	 * it anymore, so the caller must complete all errorable operations +	 * before calling iommufd_object_finalize(). +	 */ +	rc = xa_alloc(&ictx->objects, &obj->id, XA_ZERO_ENTRY, +		      xa_limit_32b, GFP_KERNEL_ACCOUNT); +	if (rc) +		goto out_free; +	return obj; +out_free: +	kfree(obj); +	return ERR_PTR(rc); +} + +/* + * Allow concurrent access to the object. + * + * Once another thread can see the object pointer it can prevent object + * destruction. Expect for special kernel-only objects there is no in-kernel way + * to reliably destroy a single object. Thus all APIs that are creating objects + * must use iommufd_object_abort() to handle their errors and only call + * iommufd_object_finalize() once object creation cannot fail. + */ +void iommufd_object_finalize(struct iommufd_ctx *ictx, +			     struct iommufd_object *obj) +{ +	void *old; + +	old = xa_store(&ictx->objects, obj->id, obj, GFP_KERNEL); +	/* obj->id was returned from xa_alloc() so the xa_store() cannot fail */ +	WARN_ON(old); +} + +/* Undo _iommufd_object_alloc() if iommufd_object_finalize() was not called */ +void iommufd_object_abort(struct iommufd_ctx *ictx, struct iommufd_object *obj) +{ +	void *old; + +	old = xa_erase(&ictx->objects, obj->id); +	WARN_ON(old); +	kfree(obj); +} + +/* + * Abort an object that has been fully initialized and needs destroy, but has + * not been finalized. + */ +void iommufd_object_abort_and_destroy(struct iommufd_ctx *ictx, +				      struct iommufd_object *obj) +{ +	iommufd_object_ops[obj->type].destroy(obj); +	iommufd_object_abort(ictx, obj); +} + +struct iommufd_object *iommufd_get_object(struct iommufd_ctx *ictx, u32 id, +					  enum iommufd_object_type type) +{ +	struct iommufd_object *obj; + +	if (iommufd_should_fail()) +		return ERR_PTR(-ENOENT); + +	xa_lock(&ictx->objects); +	obj = xa_load(&ictx->objects, id); +	if (!obj || (type != IOMMUFD_OBJ_ANY && obj->type != type) || +	    !iommufd_lock_obj(obj)) +		obj = ERR_PTR(-ENOENT); +	xa_unlock(&ictx->objects); +	return obj; +} + +/* + * The caller holds a users refcount and wants to destroy the object. Returns + * true if the object was destroyed. In all cases the caller no longer has a + * reference on obj. + */ +bool iommufd_object_destroy_user(struct iommufd_ctx *ictx, +				 struct iommufd_object *obj) +{ +	/* +	 * The purpose of the destroy_rwsem is to ensure deterministic +	 * destruction of objects used by external drivers and destroyed by this +	 * function. Any temporary increment of the refcount must hold the read +	 * side of this, such as during ioctl execution. +	 */ +	down_write(&obj->destroy_rwsem); +	xa_lock(&ictx->objects); +	refcount_dec(&obj->users); +	if (!refcount_dec_if_one(&obj->users)) { +		xa_unlock(&ictx->objects); +		up_write(&obj->destroy_rwsem); +		return false; +	} +	__xa_erase(&ictx->objects, obj->id); +	if (ictx->vfio_ioas && &ictx->vfio_ioas->obj == obj) +		ictx->vfio_ioas = NULL; +	xa_unlock(&ictx->objects); +	up_write(&obj->destroy_rwsem); + +	iommufd_object_ops[obj->type].destroy(obj); +	kfree(obj); +	return true; +} + +static int iommufd_destroy(struct iommufd_ucmd *ucmd) +{ +	struct iommu_destroy *cmd = ucmd->cmd; +	struct iommufd_object *obj; + +	obj = iommufd_get_object(ucmd->ictx, cmd->id, IOMMUFD_OBJ_ANY); +	if (IS_ERR(obj)) +		return PTR_ERR(obj); +	iommufd_ref_to_users(obj); +	/* See iommufd_ref_to_users() */ +	if (!iommufd_object_destroy_user(ucmd->ictx, obj)) +		return -EBUSY; +	return 0; +} + +static int iommufd_fops_open(struct inode *inode, struct file *filp) +{ +	struct iommufd_ctx *ictx; + +	ictx = kzalloc(sizeof(*ictx), GFP_KERNEL_ACCOUNT); +	if (!ictx) +		return -ENOMEM; + +	/* +	 * For compatibility with VFIO when /dev/vfio/vfio is opened we default +	 * to the same rlimit accounting as vfio uses. +	 */ +	if (IS_ENABLED(CONFIG_IOMMUFD_VFIO_CONTAINER) && +	    filp->private_data == &vfio_misc_dev) { +		ictx->account_mode = IOPT_PAGES_ACCOUNT_MM; +		pr_info_once("IOMMUFD is providing /dev/vfio/vfio, not VFIO.\n"); +	} + +	xa_init_flags(&ictx->objects, XA_FLAGS_ALLOC1 | XA_FLAGS_ACCOUNT); +	ictx->file = filp; +	filp->private_data = ictx; +	return 0; +} + +static int iommufd_fops_release(struct inode *inode, struct file *filp) +{ +	struct iommufd_ctx *ictx = filp->private_data; +	struct iommufd_object *obj; + +	/* +	 * The objects in the xarray form a graph of "users" counts, and we have +	 * to destroy them in a depth first manner. Leaf objects will reduce the +	 * users count of interior objects when they are destroyed. +	 * +	 * Repeatedly destroying all the "1 users" leaf objects will progress +	 * until the entire list is destroyed. If this can't progress then there +	 * is some bug related to object refcounting. +	 */ +	while (!xa_empty(&ictx->objects)) { +		unsigned int destroyed = 0; +		unsigned long index; + +		xa_for_each(&ictx->objects, index, obj) { +			if (!refcount_dec_if_one(&obj->users)) +				continue; +			destroyed++; +			xa_erase(&ictx->objects, index); +			iommufd_object_ops[obj->type].destroy(obj); +			kfree(obj); +		} +		/* Bug related to users refcount */ +		if (WARN_ON(!destroyed)) +			break; +	} +	kfree(ictx); +	return 0; +} + +static int iommufd_option(struct iommufd_ucmd *ucmd) +{ +	struct iommu_option *cmd = ucmd->cmd; +	int rc; + +	if (cmd->__reserved) +		return -EOPNOTSUPP; + +	switch (cmd->option_id) { +	case IOMMU_OPTION_RLIMIT_MODE: +		rc = iommufd_option_rlimit_mode(cmd, ucmd->ictx); +		break; +	case IOMMU_OPTION_HUGE_PAGES: +		rc = iommufd_ioas_option(ucmd); +		break; +	default: +		return -EOPNOTSUPP; +	} +	if (rc) +		return rc; +	if (copy_to_user(&((struct iommu_option __user *)ucmd->ubuffer)->val64, +			 &cmd->val64, sizeof(cmd->val64))) +		return -EFAULT; +	return 0; +} + +union ucmd_buffer { +	struct iommu_destroy destroy; +	struct iommu_ioas_alloc alloc; +	struct iommu_ioas_allow_iovas allow_iovas; +	struct iommu_ioas_iova_ranges iova_ranges; +	struct iommu_ioas_map map; +	struct iommu_ioas_unmap unmap; +#ifdef CONFIG_IOMMUFD_TEST +	struct iommu_test_cmd test; +#endif +}; + +struct iommufd_ioctl_op { +	unsigned int size; +	unsigned int min_size; +	unsigned int ioctl_num; +	int (*execute)(struct iommufd_ucmd *ucmd); +}; + +#define IOCTL_OP(_ioctl, _fn, _struct, _last)                                  \ +	[_IOC_NR(_ioctl) - IOMMUFD_CMD_BASE] = {                               \ +		.size = sizeof(_struct) +                                      \ +			BUILD_BUG_ON_ZERO(sizeof(union ucmd_buffer) <          \ +					  sizeof(_struct)),                    \ +		.min_size = offsetofend(_struct, _last),                       \ +		.ioctl_num = _ioctl,                                           \ +		.execute = _fn,                                                \ +	} +static const struct iommufd_ioctl_op iommufd_ioctl_ops[] = { +	IOCTL_OP(IOMMU_DESTROY, iommufd_destroy, struct iommu_destroy, id), +	IOCTL_OP(IOMMU_IOAS_ALLOC, iommufd_ioas_alloc_ioctl, +		 struct iommu_ioas_alloc, out_ioas_id), +	IOCTL_OP(IOMMU_IOAS_ALLOW_IOVAS, iommufd_ioas_allow_iovas, +		 struct iommu_ioas_allow_iovas, allowed_iovas), +	IOCTL_OP(IOMMU_IOAS_COPY, iommufd_ioas_copy, struct iommu_ioas_copy, +		 src_iova), +	IOCTL_OP(IOMMU_IOAS_IOVA_RANGES, iommufd_ioas_iova_ranges, +		 struct iommu_ioas_iova_ranges, out_iova_alignment), +	IOCTL_OP(IOMMU_IOAS_MAP, iommufd_ioas_map, struct iommu_ioas_map, +		 iova), +	IOCTL_OP(IOMMU_IOAS_UNMAP, iommufd_ioas_unmap, struct iommu_ioas_unmap, +		 length), +	IOCTL_OP(IOMMU_OPTION, iommufd_option, struct iommu_option, +		 val64), +	IOCTL_OP(IOMMU_VFIO_IOAS, iommufd_vfio_ioas, struct iommu_vfio_ioas, +		 __reserved), +#ifdef CONFIG_IOMMUFD_TEST +	IOCTL_OP(IOMMU_TEST_CMD, iommufd_test, struct iommu_test_cmd, last), +#endif +}; + +static long iommufd_fops_ioctl(struct file *filp, unsigned int cmd, +			       unsigned long arg) +{ +	struct iommufd_ctx *ictx = filp->private_data; +	const struct iommufd_ioctl_op *op; +	struct iommufd_ucmd ucmd = {}; +	union ucmd_buffer buf; +	unsigned int nr; +	int ret; + +	nr = _IOC_NR(cmd); +	if (nr < IOMMUFD_CMD_BASE || +	    (nr - IOMMUFD_CMD_BASE) >= ARRAY_SIZE(iommufd_ioctl_ops)) +		return iommufd_vfio_ioctl(ictx, cmd, arg); + +	ucmd.ictx = ictx; +	ucmd.ubuffer = (void __user *)arg; +	ret = get_user(ucmd.user_size, (u32 __user *)ucmd.ubuffer); +	if (ret) +		return ret; + +	op = &iommufd_ioctl_ops[nr - IOMMUFD_CMD_BASE]; +	if (op->ioctl_num != cmd) +		return -ENOIOCTLCMD; +	if (ucmd.user_size < op->min_size) +		return -EINVAL; + +	ucmd.cmd = &buf; +	ret = copy_struct_from_user(ucmd.cmd, op->size, ucmd.ubuffer, +				    ucmd.user_size); +	if (ret) +		return ret; +	ret = op->execute(&ucmd); +	return ret; +} + +static const struct file_operations iommufd_fops = { +	.owner = THIS_MODULE, +	.open = iommufd_fops_open, +	.release = iommufd_fops_release, +	.unlocked_ioctl = iommufd_fops_ioctl, +}; + +/** + * iommufd_ctx_get - Get a context reference + * @ictx: Context to get + * + * The caller must already hold a valid reference to ictx. + */ +void iommufd_ctx_get(struct iommufd_ctx *ictx) +{ +	get_file(ictx->file); +} +EXPORT_SYMBOL_NS_GPL(iommufd_ctx_get, IOMMUFD); + +/** + * iommufd_ctx_from_file - Acquires a reference to the iommufd context + * @file: File to obtain the reference from + * + * Returns a pointer to the iommufd_ctx, otherwise ERR_PTR. The struct file + * remains owned by the caller and the caller must still do fput. On success + * the caller is responsible to call iommufd_ctx_put(). + */ +struct iommufd_ctx *iommufd_ctx_from_file(struct file *file) +{ +	struct iommufd_ctx *ictx; + +	if (file->f_op != &iommufd_fops) +		return ERR_PTR(-EBADFD); +	ictx = file->private_data; +	iommufd_ctx_get(ictx); +	return ictx; +} +EXPORT_SYMBOL_NS_GPL(iommufd_ctx_from_file, IOMMUFD); + +/** + * iommufd_ctx_put - Put back a reference + * @ictx: Context to put back + */ +void iommufd_ctx_put(struct iommufd_ctx *ictx) +{ +	fput(ictx->file); +} +EXPORT_SYMBOL_NS_GPL(iommufd_ctx_put, IOMMUFD); + +static const struct iommufd_object_ops iommufd_object_ops[] = { +	[IOMMUFD_OBJ_ACCESS] = { +		.destroy = iommufd_access_destroy_object, +	}, +	[IOMMUFD_OBJ_DEVICE] = { +		.destroy = iommufd_device_destroy, +	}, +	[IOMMUFD_OBJ_IOAS] = { +		.destroy = iommufd_ioas_destroy, +	}, +	[IOMMUFD_OBJ_HW_PAGETABLE] = { +		.destroy = iommufd_hw_pagetable_destroy, +	}, +#ifdef CONFIG_IOMMUFD_TEST +	[IOMMUFD_OBJ_SELFTEST] = { +		.destroy = iommufd_selftest_destroy, +	}, +#endif +}; + +static struct miscdevice iommu_misc_dev = { +	.minor = MISC_DYNAMIC_MINOR, +	.name = "iommu", +	.fops = &iommufd_fops, +	.nodename = "iommu", +	.mode = 0660, +}; + + +static struct miscdevice vfio_misc_dev = { +	.minor = VFIO_MINOR, +	.name = "vfio", +	.fops = &iommufd_fops, +	.nodename = "vfio/vfio", +	.mode = 0666, +}; + +static int __init iommufd_init(void) +{ +	int ret; + +	ret = misc_register(&iommu_misc_dev); +	if (ret) +		return ret; + +	if (IS_ENABLED(CONFIG_IOMMUFD_VFIO_CONTAINER)) { +		ret = misc_register(&vfio_misc_dev); +		if (ret) +			goto err_misc; +	} +	iommufd_test_init(); +	return 0; +err_misc: +	misc_deregister(&iommu_misc_dev); +	return ret; +} + +static void __exit iommufd_exit(void) +{ +	iommufd_test_exit(); +	if (IS_ENABLED(CONFIG_IOMMUFD_VFIO_CONTAINER)) +		misc_deregister(&vfio_misc_dev); +	misc_deregister(&iommu_misc_dev); +} + +module_init(iommufd_init); +module_exit(iommufd_exit); + +#if IS_ENABLED(CONFIG_IOMMUFD_VFIO_CONTAINER) +MODULE_ALIAS_MISCDEV(VFIO_MINOR); +MODULE_ALIAS("devname:vfio/vfio"); +#endif +MODULE_DESCRIPTION("I/O Address Space Management for passthrough devices"); +MODULE_LICENSE("GPL");  | 
