123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382 |
- // SPDX-License-Identifier: GPL-2.0
- /*
- * Copyright (c) 2023 MediaTek Inc.
- */
- #include <linux/eventfd.h>
- #include <linux/syscalls.h>
- #include <linux/gzvm_drv.h>
- #include "gzvm_common.h"
- struct gzvm_irq_ack_notifier {
- struct hlist_node link;
- unsigned int gsi;
- void (*irq_acked)(struct gzvm_irq_ack_notifier *ian);
- };
- /**
- * struct gzvm_kernel_irqfd: gzvm kernel irqfd descriptor.
- * @gzvm: Pointer to struct gzvm.
- * @wait: Wait queue entry.
- * @gsi: Used for level IRQ fast-path.
- * @eventfd: Used for setup/shutdown.
- * @list: struct list_head.
- * @pt: struct poll_table_struct.
- * @shutdown: struct work_struct.
- */
- struct gzvm_kernel_irqfd {
- struct gzvm *gzvm;
- wait_queue_entry_t wait;
- int gsi;
- struct eventfd_ctx *eventfd;
- struct list_head list;
- poll_table pt;
- struct work_struct shutdown;
- };
- static struct workqueue_struct *irqfd_cleanup_wq;
- /**
- * irqfd_set_irq(): irqfd to inject virtual interrupt.
- * @gzvm: Pointer to gzvm.
- * @irq: This is spi interrupt number (starts from 0 instead of 32).
- * @level: irq triggered level.
- */
- static void irqfd_set_irq(struct gzvm *gzvm, u32 irq, int level)
- {
- if (level)
- gzvm_irqchip_inject_irq(gzvm, 0, irq, level);
- }
- /**
- * irqfd_shutdown() - Race-free decouple logic (ordering is critical).
- * @work: Pointer to work_struct.
- */
- static void irqfd_shutdown(struct work_struct *work)
- {
- struct gzvm_kernel_irqfd *irqfd =
- container_of(work, struct gzvm_kernel_irqfd, shutdown);
- struct gzvm *gzvm = irqfd->gzvm;
- u64 cnt;
- /* Make sure irqfd has been initialized in assign path. */
- synchronize_srcu(&gzvm->irq_srcu);
- /*
- * Synchronize with the wait-queue and unhook ourselves to prevent
- * further events.
- */
- eventfd_ctx_remove_wait_queue(irqfd->eventfd, &irqfd->wait, &cnt);
- /*
- * It is now safe to release the object's resources
- */
- eventfd_ctx_put(irqfd->eventfd);
- kfree(irqfd);
- }
- /**
- * irqfd_is_active() - Assumes gzvm->irqfds.lock is held.
- * @irqfd: Pointer to gzvm_kernel_irqfd.
- *
- * Return:
- * * true - irqfd is active.
- */
- static bool irqfd_is_active(struct gzvm_kernel_irqfd *irqfd)
- {
- return list_empty(&irqfd->list) ? false : true;
- }
- /**
- * irqfd_deactivate() - Mark the irqfd as inactive and schedule it for removal.
- * assumes gzvm->irqfds.lock is held.
- * @irqfd: Pointer to gzvm_kernel_irqfd.
- */
- static void irqfd_deactivate(struct gzvm_kernel_irqfd *irqfd)
- {
- if (!irqfd_is_active(irqfd))
- return;
- list_del_init(&irqfd->list);
- queue_work(irqfd_cleanup_wq, &irqfd->shutdown);
- }
- /**
- * irqfd_wakeup() - Callback of irqfd wait queue, would be woken by writing to
- * irqfd to do virtual interrupt injection.
- * @wait: Pointer to wait_queue_entry_t.
- * @mode: Unused.
- * @sync: Unused.
- * @key: Get flags about Epoll events.
- *
- * Return:
- * * 0 - Success
- */
- static int irqfd_wakeup(wait_queue_entry_t *wait, unsigned int mode, int sync,
- void *key)
- {
- struct gzvm_kernel_irqfd *irqfd =
- container_of(wait, struct gzvm_kernel_irqfd, wait);
- __poll_t flags = key_to_poll(key);
- struct gzvm *gzvm = irqfd->gzvm;
- if (flags & EPOLLIN) {
- u64 cnt;
- eventfd_ctx_do_read(irqfd->eventfd, &cnt);
- /* gzvm's irq injection is not blocked, don't need workq */
- irqfd_set_irq(gzvm, irqfd->gsi, 1);
- }
- if (flags & EPOLLHUP) {
- /* The eventfd is closing, detach from GZVM */
- unsigned long iflags;
- spin_lock_irqsave(&gzvm->irqfds.lock, iflags);
- /*
- * Do more check if someone deactivated the irqfd before
- * we could acquire the irqfds.lock.
- */
- if (irqfd_is_active(irqfd))
- irqfd_deactivate(irqfd);
- spin_unlock_irqrestore(&gzvm->irqfds.lock, iflags);
- }
- return 0;
- }
- static void irqfd_ptable_queue_proc(struct file *file, wait_queue_head_t *wqh,
- poll_table *pt)
- {
- struct gzvm_kernel_irqfd *irqfd =
- container_of(pt, struct gzvm_kernel_irqfd, pt);
- add_wait_queue_priority(wqh, &irqfd->wait);
- }
- static int gzvm_irqfd_assign(struct gzvm *gzvm, struct gzvm_irqfd *args)
- {
- struct gzvm_kernel_irqfd *irqfd, *tmp;
- struct fd f;
- struct eventfd_ctx *eventfd = NULL;
- int ret;
- int idx;
- irqfd = kzalloc(sizeof(*irqfd), GFP_KERNEL_ACCOUNT);
- if (!irqfd)
- return -ENOMEM;
- irqfd->gzvm = gzvm;
- irqfd->gsi = args->gsi;
- INIT_LIST_HEAD(&irqfd->list);
- INIT_WORK(&irqfd->shutdown, irqfd_shutdown);
- f = fdget(args->fd);
- if (!f.file) {
- ret = -EBADF;
- goto out;
- }
- eventfd = eventfd_ctx_fileget(f.file);
- if (IS_ERR(eventfd)) {
- ret = PTR_ERR(eventfd);
- goto fail;
- }
- irqfd->eventfd = eventfd;
- /*
- * Install our own custom wake-up handling so we are notified via
- * a callback whenever someone signals the underlying eventfd
- */
- init_waitqueue_func_entry(&irqfd->wait, irqfd_wakeup);
- init_poll_funcptr(&irqfd->pt, irqfd_ptable_queue_proc);
- spin_lock_irq(&gzvm->irqfds.lock);
- ret = 0;
- list_for_each_entry(tmp, &gzvm->irqfds.items, list) {
- if (irqfd->eventfd != tmp->eventfd)
- continue;
- /* This fd is used for another irq already. */
- pr_err("already used: gsi=%d fd=%d\n", args->gsi, args->fd);
- ret = -EBUSY;
- spin_unlock_irq(&gzvm->irqfds.lock);
- goto fail;
- }
- idx = srcu_read_lock(&gzvm->irq_srcu);
- list_add_tail(&irqfd->list, &gzvm->irqfds.items);
- spin_unlock_irq(&gzvm->irqfds.lock);
- vfs_poll(f.file, &irqfd->pt);
- srcu_read_unlock(&gzvm->irq_srcu, idx);
- /*
- * do not drop the file until the irqfd is fully initialized, otherwise
- * we might race against the EPOLLHUP
- */
- fdput(f);
- return 0;
- fail:
- if (eventfd && !IS_ERR(eventfd))
- eventfd_ctx_put(eventfd);
- fdput(f);
- out:
- kfree(irqfd);
- return ret;
- }
- static void gzvm_notify_acked_gsi(struct gzvm *gzvm, int gsi)
- {
- struct gzvm_irq_ack_notifier *gian;
- hlist_for_each_entry_srcu(gian, &gzvm->irq_ack_notifier_list,
- link, srcu_read_lock_held(&gzvm->irq_srcu))
- if (gian->gsi == gsi)
- gian->irq_acked(gian);
- }
- void gzvm_notify_acked_irq(struct gzvm *gzvm, unsigned int gsi)
- {
- int idx;
- idx = srcu_read_lock(&gzvm->irq_srcu);
- gzvm_notify_acked_gsi(gzvm, gsi);
- srcu_read_unlock(&gzvm->irq_srcu, idx);
- }
- /**
- * gzvm_irqfd_deassign() - Shutdown any irqfd's that match fd+gsi.
- * @gzvm: Pointer to gzvm.
- * @args: Pointer to gzvm_irqfd.
- *
- * Return:
- * * 0 - Success.
- * * Negative value - Failure.
- */
- static int gzvm_irqfd_deassign(struct gzvm *gzvm, struct gzvm_irqfd *args)
- {
- struct gzvm_kernel_irqfd *irqfd, *tmp;
- struct eventfd_ctx *eventfd;
- eventfd = eventfd_ctx_fdget(args->fd);
- if (IS_ERR(eventfd))
- return PTR_ERR(eventfd);
- spin_lock_irq(&gzvm->irqfds.lock);
- list_for_each_entry_safe(irqfd, tmp, &gzvm->irqfds.items, list) {
- if (irqfd->eventfd == eventfd && irqfd->gsi == args->gsi)
- irqfd_deactivate(irqfd);
- }
- spin_unlock_irq(&gzvm->irqfds.lock);
- eventfd_ctx_put(eventfd);
- /*
- * Block until we know all outstanding shutdown jobs have completed
- * so that we guarantee there will not be any more interrupts on this
- * gsi once this deassign function returns.
- */
- flush_workqueue(irqfd_cleanup_wq);
- return 0;
- }
- int gzvm_irqfd(struct gzvm *gzvm, struct gzvm_irqfd *args)
- {
- for (int i = 0; i < ARRAY_SIZE(args->pad); i++) {
- if (args->pad[i])
- return -EINVAL;
- }
- if (args->flags &
- ~(GZVM_IRQFD_FLAG_DEASSIGN | GZVM_IRQFD_FLAG_RESAMPLE))
- return -EINVAL;
- if (args->flags & GZVM_IRQFD_FLAG_DEASSIGN)
- return gzvm_irqfd_deassign(gzvm, args);
- return gzvm_irqfd_assign(gzvm, args);
- }
- /**
- * gzvm_vm_irqfd_init() - Initialize irqfd data structure per VM
- *
- * @gzvm: Pointer to struct gzvm.
- *
- * Return:
- * * 0 - Success.
- * * Negative - Failure.
- */
- int gzvm_vm_irqfd_init(struct gzvm *gzvm)
- {
- mutex_init(&gzvm->irq_lock);
- spin_lock_init(&gzvm->irqfds.lock);
- INIT_LIST_HEAD(&gzvm->irqfds.items);
- if (init_srcu_struct(&gzvm->irq_srcu))
- return -EINVAL;
- INIT_HLIST_HEAD(&gzvm->irq_ack_notifier_list);
- return 0;
- }
- /**
- * gzvm_vm_irqfd_release() - This function is called as the gzvm VM fd is being
- * released. Shutdown all irqfds that still remain open.
- * @gzvm: Pointer to gzvm.
- */
- void gzvm_vm_irqfd_release(struct gzvm *gzvm)
- {
- struct gzvm_kernel_irqfd *irqfd, *tmp;
- spin_lock_irq(&gzvm->irqfds.lock);
- list_for_each_entry_safe(irqfd, tmp, &gzvm->irqfds.items, list)
- irqfd_deactivate(irqfd);
- spin_unlock_irq(&gzvm->irqfds.lock);
- /*
- * Block until we know all outstanding shutdown jobs have completed.
- */
- flush_workqueue(irqfd_cleanup_wq);
- }
- /**
- * gzvm_drv_irqfd_init() - Erase flushing work items when a VM exits.
- *
- * Return:
- * * 0 - Success.
- * * Negative - Failure.
- *
- * Create a host-wide workqueue for issuing deferred shutdown requests
- * aggregated from all vm* instances. We need our own isolated
- * queue to ease flushing work items when a VM exits.
- */
- int gzvm_drv_irqfd_init(void)
- {
- irqfd_cleanup_wq = alloc_workqueue("gzvm-irqfd-cleanup", 0, 0);
- if (!irqfd_cleanup_wq)
- return -ENOMEM;
- return 0;
- }
- void gzvm_drv_irqfd_exit(void)
- {
- destroy_workqueue(irqfd_cleanup_wq);
- }
|