/* * Copyright (C) 2018 Intel Corporation. All rights reserved. * * SPDX-License-Identifier: BSD-3-Clause */ #include #include #include /* passthrough device link */ struct list_head ptdev_list; spinlock_t ptdev_lock; /* * entry could both be in ptdev_list and vm->softirq_dev_entry_list. * When release entry, we need make sure entry deleted from both * lists. We have to require two locks and the lock sequence is: * ptdev_lock * vm->softirq_dev_lock */ static void ptdev_enqueue_softirq(struct ptdev_remapping_info *entry) { uint64_t rflags; /* enqueue request in order, SOFTIRQ_PTDEV will pickup */ spinlock_irqsave_obtain(&entry->vm->softirq_dev_lock, &rflags); /* avoid adding recursively */ list_del(&entry->softirq_node); /* TODO: assert if entry already in list */ list_add_tail(&entry->softirq_node, &entry->vm->softirq_dev_entry_list); spinlock_irqrestore_release(&entry->vm->softirq_dev_lock, rflags); fire_softirq(SOFTIRQ_PTDEV); } struct ptdev_remapping_info* ptdev_dequeue_softirq(struct vm *vm) { uint64_t rflags; struct ptdev_remapping_info *entry = NULL; spinlock_irqsave_obtain(&vm->softirq_dev_lock, &rflags); if (!list_empty(&vm->softirq_dev_entry_list)) { entry = get_first_item(&vm->softirq_dev_entry_list, struct ptdev_remapping_info, softirq_node); list_del_init(&entry->softirq_node); } spinlock_irqrestore_release(&vm->softirq_dev_lock, rflags); return entry; } /* require ptdev_lock protect */ struct ptdev_remapping_info * alloc_entry(struct vm *vm, uint32_t intr_type) { struct ptdev_remapping_info *entry; /* allocate */ entry = calloc(1U, sizeof(*entry)); ASSERT(entry != NULL, "alloc memory failed"); entry->intr_type = intr_type; entry->vm = vm; INIT_LIST_HEAD(&entry->softirq_node); INIT_LIST_HEAD(&entry->entry_node); atomic_clear32(&entry->active, ACTIVE_FLAG); list_add(&entry->entry_node, &ptdev_list); return entry; } /* require ptdev_lock protect */ void release_entry(struct ptdev_remapping_info *entry) { uint64_t rflags; /* remove entry from ptdev_list */ list_del_init(&entry->entry_node); /* * remove entry from softirq list.the ptdev_lock * is required before calling release_entry. */ spinlock_irqsave_obtain(&entry->vm->softirq_dev_lock, &rflags); list_del_init(&entry->softirq_node); spinlock_irqrestore_release(&entry->vm->softirq_dev_lock, rflags); free(entry); } /* require ptdev_lock protect */ static void release_all_entries(struct vm *vm) { struct ptdev_remapping_info *entry; struct list_head *pos, *tmp; list_for_each_safe(pos, tmp, &ptdev_list) { entry = list_entry(pos, struct ptdev_remapping_info, entry_node); if (entry->vm == vm) release_entry(entry); } } /* interrupt context */ static void ptdev_interrupt_handler(__unused uint32_t irq, void *data) { struct ptdev_remapping_info *entry = (struct ptdev_remapping_info *) data; ptdev_enqueue_softirq(entry); } /* active intr with irq registering */ void ptdev_activate_entry(struct ptdev_remapping_info *entry, uint32_t phys_irq) { int32_t retval; /* register and allocate host vector/irq */ retval = request_irq(phys_irq, ptdev_interrupt_handler, (void *)entry, IRQF_PT); ASSERT(retval >= 0, "dev register failed"); entry->allocated_pirq = (uint32_t)retval; atomic_set32(&entry->active, ACTIVE_FLAG); } void ptdev_deactivate_entry(struct ptdev_remapping_info *entry) { uint64_t rflags; atomic_clear32(&entry->active, ACTIVE_FLAG); free_irq(entry->allocated_pirq); entry->allocated_pirq = IRQ_INVALID; /* remove from softirq list if added */ spinlock_irqsave_obtain(&entry->vm->softirq_dev_lock, &rflags); list_del_init(&entry->softirq_node); spinlock_irqrestore_release(&entry->vm->softirq_dev_lock, rflags); } void ptdev_init(void) { if (get_cpu_id() > 0) return; INIT_LIST_HEAD(&ptdev_list); spinlock_init(&ptdev_lock); register_softirq(SOFTIRQ_PTDEV, ptdev_softirq); } void ptdev_release_all_entries(struct vm *vm) { /* VM already down */ spinlock_obtain(&ptdev_lock); release_all_entries(vm); spinlock_release(&ptdev_lock); }