236 lines
		
	
	
		
			5.3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			236 lines
		
	
	
		
			5.3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * ACRN HSM irqfd: use eventfd objects to inject virtual interrupts
 | |
|  *
 | |
|  * Copyright (C) 2020 Intel Corporation. All rights reserved.
 | |
|  *
 | |
|  * Authors:
 | |
|  *	Shuo Liu <shuo.a.liu@intel.com>
 | |
|  *	Yakui Zhao <yakui.zhao@intel.com>
 | |
|  */
 | |
| 
 | |
| #include <linux/eventfd.h>
 | |
| #include <linux/file.h>
 | |
| #include <linux/poll.h>
 | |
| #include <linux/slab.h>
 | |
| 
 | |
| #include "acrn_drv.h"
 | |
| 
 | |
| static LIST_HEAD(acrn_irqfd_clients);
 | |
| static DEFINE_MUTEX(acrn_irqfds_mutex);
 | |
| 
 | |
| /**
 | |
|  * struct hsm_irqfd - Properties of HSM irqfd
 | |
|  * @vm:		Associated VM pointer
 | |
|  * @wait:	Entry of wait-queue
 | |
|  * @shutdown:	Async shutdown work
 | |
|  * @eventfd:	Associated eventfd
 | |
|  * @list:	Entry within &acrn_vm.irqfds of irqfds of a VM
 | |
|  * @pt:		Structure for select/poll on the associated eventfd
 | |
|  * @msi:	MSI data
 | |
|  */
 | |
| struct hsm_irqfd {
 | |
| 	struct acrn_vm		*vm;
 | |
| 	wait_queue_entry_t	wait;
 | |
| 	struct work_struct	shutdown;
 | |
| 	struct eventfd_ctx	*eventfd;
 | |
| 	struct list_head	list;
 | |
| 	poll_table		pt;
 | |
| 	struct acrn_msi_entry	msi;
 | |
| };
 | |
| 
 | |
| static void acrn_irqfd_inject(struct hsm_irqfd *irqfd)
 | |
| {
 | |
| 	struct acrn_vm *vm = irqfd->vm;
 | |
| 
 | |
| 	acrn_msi_inject(vm, irqfd->msi.msi_addr,
 | |
| 			irqfd->msi.msi_data);
 | |
| }
 | |
| 
 | |
| static void hsm_irqfd_shutdown(struct hsm_irqfd *irqfd)
 | |
| {
 | |
| 	u64 cnt;
 | |
| 
 | |
| 	lockdep_assert_held(&irqfd->vm->irqfds_lock);
 | |
| 
 | |
| 	/* remove from wait queue */
 | |
| 	list_del_init(&irqfd->list);
 | |
| 	eventfd_ctx_remove_wait_queue(irqfd->eventfd, &irqfd->wait, &cnt);
 | |
| 	eventfd_ctx_put(irqfd->eventfd);
 | |
| 	kfree(irqfd);
 | |
| }
 | |
| 
 | |
| static void hsm_irqfd_shutdown_work(struct work_struct *work)
 | |
| {
 | |
| 	struct hsm_irqfd *irqfd;
 | |
| 	struct acrn_vm *vm;
 | |
| 
 | |
| 	irqfd = container_of(work, struct hsm_irqfd, shutdown);
 | |
| 	vm = irqfd->vm;
 | |
| 	mutex_lock(&vm->irqfds_lock);
 | |
| 	if (!list_empty(&irqfd->list))
 | |
| 		hsm_irqfd_shutdown(irqfd);
 | |
| 	mutex_unlock(&vm->irqfds_lock);
 | |
| }
 | |
| 
 | |
| /* Called with wqh->lock held and interrupts disabled */
 | |
| static int hsm_irqfd_wakeup(wait_queue_entry_t *wait, unsigned int mode,
 | |
| 			    int sync, void *key)
 | |
| {
 | |
| 	unsigned long poll_bits = (unsigned long)key;
 | |
| 	struct hsm_irqfd *irqfd;
 | |
| 	struct acrn_vm *vm;
 | |
| 
 | |
| 	irqfd = container_of(wait, struct hsm_irqfd, wait);
 | |
| 	vm = irqfd->vm;
 | |
| 	if (poll_bits & POLLIN)
 | |
| 		/* An event has been signaled, inject an interrupt */
 | |
| 		acrn_irqfd_inject(irqfd);
 | |
| 
 | |
| 	if (poll_bits & POLLHUP)
 | |
| 		/* Do shutdown work in thread to hold wqh->lock */
 | |
| 		queue_work(vm->irqfd_wq, &irqfd->shutdown);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void hsm_irqfd_poll_func(struct file *file, wait_queue_head_t *wqh,
 | |
| 				poll_table *pt)
 | |
| {
 | |
| 	struct hsm_irqfd *irqfd;
 | |
| 
 | |
| 	irqfd = container_of(pt, struct hsm_irqfd, pt);
 | |
| 	add_wait_queue(wqh, &irqfd->wait);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Assign an eventfd to a VM and create a HSM irqfd associated with the
 | |
|  * eventfd. The properties of the HSM irqfd are built from a &struct
 | |
|  * acrn_irqfd.
 | |
|  */
 | |
| static int acrn_irqfd_assign(struct acrn_vm *vm, struct acrn_irqfd *args)
 | |
| {
 | |
| 	struct eventfd_ctx *eventfd = NULL;
 | |
| 	struct hsm_irqfd *irqfd, *tmp;
 | |
| 	__poll_t events;
 | |
| 	struct fd f;
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	irqfd = kzalloc(sizeof(*irqfd), GFP_KERNEL);
 | |
| 	if (!irqfd)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	irqfd->vm = vm;
 | |
| 	memcpy(&irqfd->msi, &args->msi, sizeof(args->msi));
 | |
| 	INIT_LIST_HEAD(&irqfd->list);
 | |
| 	INIT_WORK(&irqfd->shutdown, hsm_irqfd_shutdown_work);
 | |
| 
 | |
| 	f = fdget(args->fd);
 | |
| 	if (!f.file) {
 | |
| 		ret = -EBADF;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	eventfd = eventfd_ctx_fileget(f.file);
 | |
| 	if (IS_ERR(eventfd)) {
 | |
| 		ret = PTR_ERR(eventfd);
 | |
| 		goto fail;
 | |
| 	}
 | |
| 
 | |
| 	irqfd->eventfd = eventfd;
 | |
| 
 | |
| 	/*
 | |
| 	 * Install custom wake-up handling to be notified whenever underlying
 | |
| 	 * eventfd is signaled.
 | |
| 	 */
 | |
| 	init_waitqueue_func_entry(&irqfd->wait, hsm_irqfd_wakeup);
 | |
| 	init_poll_funcptr(&irqfd->pt, hsm_irqfd_poll_func);
 | |
| 
 | |
| 	mutex_lock(&vm->irqfds_lock);
 | |
| 	list_for_each_entry(tmp, &vm->irqfds, list) {
 | |
| 		if (irqfd->eventfd != tmp->eventfd)
 | |
| 			continue;
 | |
| 		ret = -EBUSY;
 | |
| 		mutex_unlock(&vm->irqfds_lock);
 | |
| 		goto fail;
 | |
| 	}
 | |
| 	list_add_tail(&irqfd->list, &vm->irqfds);
 | |
| 	mutex_unlock(&vm->irqfds_lock);
 | |
| 
 | |
| 	/* Check the pending event in this stage */
 | |
| 	events = vfs_poll(f.file, &irqfd->pt);
 | |
| 
 | |
| 	if (events & EPOLLIN)
 | |
| 		acrn_irqfd_inject(irqfd);
 | |
| 
 | |
| 	fdput(f);
 | |
| 	return 0;
 | |
| fail:
 | |
| 	if (eventfd && !IS_ERR(eventfd))
 | |
| 		eventfd_ctx_put(eventfd);
 | |
| 
 | |
| 	fdput(f);
 | |
| out:
 | |
| 	kfree(irqfd);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int acrn_irqfd_deassign(struct acrn_vm *vm,
 | |
| 			       struct acrn_irqfd *args)
 | |
| {
 | |
| 	struct hsm_irqfd *irqfd, *tmp;
 | |
| 	struct eventfd_ctx *eventfd;
 | |
| 
 | |
| 	eventfd = eventfd_ctx_fdget(args->fd);
 | |
| 	if (IS_ERR(eventfd))
 | |
| 		return PTR_ERR(eventfd);
 | |
| 
 | |
| 	mutex_lock(&vm->irqfds_lock);
 | |
| 	list_for_each_entry_safe(irqfd, tmp, &vm->irqfds, list) {
 | |
| 		if (irqfd->eventfd == eventfd) {
 | |
| 			hsm_irqfd_shutdown(irqfd);
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 	mutex_unlock(&vm->irqfds_lock);
 | |
| 	eventfd_ctx_put(eventfd);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int acrn_irqfd_config(struct acrn_vm *vm, struct acrn_irqfd *args)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	if (args->flags & ACRN_IRQFD_FLAG_DEASSIGN)
 | |
| 		ret = acrn_irqfd_deassign(vm, args);
 | |
| 	else
 | |
| 		ret = acrn_irqfd_assign(vm, args);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int acrn_irqfd_init(struct acrn_vm *vm)
 | |
| {
 | |
| 	INIT_LIST_HEAD(&vm->irqfds);
 | |
| 	mutex_init(&vm->irqfds_lock);
 | |
| 	vm->irqfd_wq = alloc_workqueue("acrn_irqfd-%u", 0, 0, vm->vmid);
 | |
| 	if (!vm->irqfd_wq)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	dev_dbg(acrn_dev.this_device, "VM %u irqfd init.\n", vm->vmid);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void acrn_irqfd_deinit(struct acrn_vm *vm)
 | |
| {
 | |
| 	struct hsm_irqfd *irqfd, *next;
 | |
| 
 | |
| 	dev_dbg(acrn_dev.this_device, "VM %u irqfd deinit.\n", vm->vmid);
 | |
| 	destroy_workqueue(vm->irqfd_wq);
 | |
| 	mutex_lock(&vm->irqfds_lock);
 | |
| 	list_for_each_entry_safe(irqfd, next, &vm->irqfds, list)
 | |
| 		hsm_irqfd_shutdown(irqfd);
 | |
| 	mutex_unlock(&vm->irqfds_lock);
 | |
| }
 |