/* * Copyright (C) 2018 Intel Corporation. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in * the documentation and/or other materials provided with the * distribution. * * Neither the name of Intel Corporation nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include #include #include #include #include #define ACTIVE_FLAG 0x1 /* any non zero should be okay */ #define INTX_ID_IOAPIC (0 << 23) #define INTX_ID_PIC (1 << 23) /* SOFTIRQ_DEV_ASSIGN list for all CPUs */ static struct list_head softirq_dev_entry_list; /* * entry could both be in ptdev_list and softirq_dev_entry_list. * When release entry, we need make sure entry deleted from both * lists. We have to require two locks and the lock sequence is: * vm->ptdev_lock * softirq_dev_lock */ static spinlock_t softirq_dev_lock; static inline uint32_t entry_id_from_msix(uint16_t vbdf, int8_t index) { uint32_t id = index; id = vbdf | (id << 16) | (PTDEV_INTR_MSI << 24); return id; } static inline uint32_t entry_id_from_intx(uint8_t vpin, enum ptdev_vpin_source vpin_src) { uint32_t id = vpin; uint32_t src_id = vpin_src == PTDEV_VPIN_IOAPIC ? INTX_ID_IOAPIC : INTX_ID_PIC; id = id | src_id | (PTDEV_INTR_INTX << 24); return id; } static inline uint32_t entry_id(struct ptdev_remapping_info *entry) { uint32_t id; if (entry->type == PTDEV_INTR_INTX) id = entry_id_from_intx(entry->intx.virt_pin, entry->intx.vpin_src); else id = entry_id_from_msix(entry->virt_bdf, entry->msi.msix_entry_index); return id; } static inline bool entry_is_active(struct ptdev_remapping_info *entry) { return atomic_load_acq_int(&entry->active) == ACTIVE_FLAG; } /* require ptdev_lock protect */ static inline struct ptdev_remapping_info * _get_remapping_entry(struct vm *vm, uint32_t id) { struct ptdev_remapping_info *entry; struct list_head *pos; list_for_each(pos, &vm->ptdev_list) { entry = list_entry(pos, struct ptdev_remapping_info, entry_node); if (entry_id(entry) == id) return entry; } return NULL; } static inline struct ptdev_remapping_info * get_remapping_entry(struct vm *vm, uint32_t id) { struct ptdev_remapping_info *entry; spinlock_obtain(&vm->ptdev_lock); entry = _get_remapping_entry(vm, id); spinlock_release(&vm->ptdev_lock); return entry; } static void ptdev_enqueue_softirq(struct ptdev_remapping_info *entry) { spinlock_rflags; /* enqueue request in order, SOFTIRQ_DEV_ASSIGN will pickup */ spinlock_irqsave_obtain(&softirq_dev_lock); /* avoid adding recursively */ list_del(&entry->softirq_node); /* TODO: assert if entry already in list */ list_add_tail(&entry->softirq_node, &softirq_dev_entry_list); spinlock_irqrestore_release(&softirq_dev_lock); raise_softirq(SOFTIRQ_DEV_ASSIGN); } static struct ptdev_remapping_info* ptdev_dequeue_softirq(void) { struct ptdev_remapping_info *entry = NULL; spinlock_rflags; spinlock_irqsave_obtain(&softirq_dev_lock); if (!list_empty(&softirq_dev_entry_list)) { entry = get_first_item(&softirq_dev_entry_list, struct ptdev_remapping_info, softirq_node); list_del_init(&entry->softirq_node); } spinlock_irqrestore_release(&softirq_dev_lock); return entry; } /* interrupt context */ static int ptdev_interrupt_handler(__unused int irq, void *data) { struct ptdev_remapping_info *entry = (struct ptdev_remapping_info *) data; ptdev_enqueue_softirq(entry); return 0; } static void ptdev_update_irq_handler(struct vm *vm, struct ptdev_remapping_info *entry) { int phys_irq = dev_to_irq(entry->node); if (entry->type == PTDEV_INTR_MSI) { /* all other MSI and normal maskable */ update_irq_handler(phys_irq, common_handler_edge); } /* update irq handler for IOAPIC */ if ((entry->type == PTDEV_INTR_INTX) && (entry->intx.vpin_src == PTDEV_VPIN_IOAPIC)) { uint64_t rte; bool trigger_lvl = false; /* VPIN_IOAPIC src means we have vioapic enabled */ vioapic_get_rte(vm, entry->intx.virt_pin, &rte); if ((rte & IOAPIC_RTE_TRGRMOD) == IOAPIC_RTE_TRGRLVL) trigger_lvl = true; if (trigger_lvl) update_irq_handler(phys_irq, common_dev_handler_level); else update_irq_handler(phys_irq, common_handler_edge); } /* update irq handler for PIC */ if ((entry->type == PTDEV_INTR_INTX) && (phys_irq < NR_LEGACY_IRQ) && (entry->intx.vpin_src == PTDEV_VPIN_PIC)) { enum vpic_trigger trigger; /* VPIN_PIC src means we have vpic enabled */ vpic_get_irq_trigger(vm, entry->intx.virt_pin, &trigger); if (trigger == LEVEL_TRIGGER) update_irq_handler(phys_irq, common_dev_handler_level); else update_irq_handler(phys_irq, common_handler_edge); } } /* require ptdev_lock protect */ static struct ptdev_remapping_info * alloc_entry(struct vm *vm, enum ptdev_intr_type type) { struct ptdev_remapping_info *entry; /* allocate */ entry = calloc(1, sizeof(*entry)); ASSERT(entry, "alloc memory failed"); entry->type = type; entry->vm = vm; atomic_clear_int(&entry->active, ACTIVE_FLAG); list_add(&entry->entry_node, &vm->ptdev_list); return entry; } /* require ptdev_lock protect */ static void release_entry(struct ptdev_remapping_info *entry) { spinlock_rflags; /* remove entry from ptdev_list */ list_del_init(&entry->entry_node); /* * remove entry from softirq list.the ptdev_lock * is required before calling release_entry. */ spinlock_irqsave_obtain(&softirq_dev_lock); list_del_init(&entry->softirq_node); spinlock_irqrestore_release(&softirq_dev_lock); free(entry); } /* require ptdev_lock protect */ static void release_all_entry(struct vm *vm) { struct ptdev_remapping_info *entry; struct list_head *pos, *tmp; list_for_each_safe(pos, tmp, &vm->ptdev_list) { entry = list_entry(pos, struct ptdev_remapping_info, entry_node); release_entry(entry); } } /* active intr with irq registering */ static struct ptdev_remapping_info * ptdev_activate_entry(struct ptdev_remapping_info *entry, int phys_irq, bool lowpri) { struct dev_handler_node *node; /* register and allocate host vector/irq */ node = normal_register_handler(phys_irq, ptdev_interrupt_handler, (void *)entry, true, lowpri, "dev assign"); ASSERT(node != NULL, "dev register failed"); entry->node = node; atomic_set_int(&entry->active, ACTIVE_FLAG); return entry; } static void ptdev_deactivate_entry(struct ptdev_remapping_info *entry) { spinlock_rflags; atomic_clear_int(&entry->active, ACTIVE_FLAG); unregister_handler_common(entry->node); entry->node = NULL; /* remove from softirq list if added */ spinlock_irqsave_obtain(&softirq_dev_lock); list_del_init(&entry->softirq_node); spinlock_irqrestore_release(&softirq_dev_lock); } static void check_deactive_pic_intx(struct vm *vm, uint8_t phys_pin) { struct ptdev_remapping_info *entry; struct list_head *pos; if (phys_pin >= NR_LEGACY_IRQ) return; spinlock_obtain(&vm->ptdev_lock); list_for_each(pos, &vm->ptdev_list) { entry = list_entry(pos, struct ptdev_remapping_info, entry_node); if (entry->type == PTDEV_INTR_INTX && entry->intx.vpin_src == PTDEV_VPIN_PIC && entry->intx.phys_pin == phys_pin && entry_is_active(entry)) { GSI_MASK_IRQ(pin_to_irq(phys_pin)); ptdev_deactivate_entry(entry); dev_dbg(ACRN_DBG_IRQ, "IOAPIC pin=%d pirq=%d deassigned ", phys_pin, pin_to_irq(phys_pin)); dev_dbg(ACRN_DBG_IRQ, "from vm%d vPIC vpin=%d", entry->vm->attr.id, entry->intx.virt_pin); } } spinlock_release(&vm->ptdev_lock); } static bool ptdev_native_owned_intx(struct vm *vm, struct ptdev_intx_info *info) { /* vm0 pin 4 (uart) is owned by hypervisor */ if (is_vm0(vm) && info->virt_pin == 4) return true; else return false; } static void ptdev_build_native_msi(struct vm *vm, struct ptdev_msi_info *info, int vector) { uint64_t vdmask, pdmask; uint32_t dest, delmode; bool phys; /* native destination cpu mask */ dest = (info->vmsi_addr >> 12) & 0xff; phys = ((info->vmsi_addr & (MSI_ADDR_RH | MSI_ADDR_LOG)) != (MSI_ADDR_RH | MSI_ADDR_LOG)); calcvdest(vm, &vdmask, dest, phys); pdmask = vcpumask2pcpumask(vm, vdmask); /* native delivery mode */ delmode = info->vmsi_data & APIC_DELMODE_MASK; if (delmode != APIC_DELMODE_FIXED && delmode != APIC_DELMODE_LOWPRIO) delmode = APIC_DELMODE_LOWPRIO; /* update native delivery mode & vector */ info->pmsi_data = info->vmsi_data; info->pmsi_data &= ~0x7FF; info->pmsi_data |= delmode | vector; /* update native dest mode & dest field */ info->pmsi_addr = info->vmsi_addr; info->pmsi_addr &= ~0xFF00C; info->pmsi_addr |= pdmask << 12 | MSI_ADDR_RH | MSI_ADDR_LOG; dev_dbg(ACRN_DBG_IRQ, "MSI addr:data = 0x%x:%x(V) -> 0x%x:%x(P)", info->vmsi_addr, info->vmsi_data, info->pmsi_addr, info->pmsi_data); } static uint64_t ptdev_build_native_rte(struct vm *vm, struct ptdev_intx_info *info, int vector) { uint64_t rte; if (info->vpin_src == PTDEV_VPIN_IOAPIC) { uint64_t vdmask, pdmask; uint32_t dest, low, high, delmode; bool phys; vioapic_get_rte(vm, info->virt_pin, &rte); low = rte; high = rte >> 32; /* native destination cpu mask */ phys = ((low & IOAPIC_RTE_DESTMOD) == IOAPIC_RTE_DESTPHY); dest = high >> APIC_ID_SHIFT; calcvdest(vm, &vdmask, dest, phys); pdmask = vcpumask2pcpumask(vm, vdmask); /* native delivery mode */ delmode = low & IOAPIC_RTE_DELMOD; if ((delmode != IOAPIC_RTE_DELFIXED) && (delmode != IOAPIC_RTE_DELLOPRI)) delmode = IOAPIC_RTE_DELLOPRI; /* update native delivery mode, dest mode(logical) & vector */ low &= ~(IOAPIC_RTE_DESTMOD | IOAPIC_RTE_DELMOD | IOAPIC_RTE_INTVEC); low |= IOAPIC_RTE_DESTLOG | delmode | vector; /* update native dest field */ high &= ~IOAPIC_RTE_DEST; high |= pdmask << 24; dev_dbg(ACRN_DBG_IRQ, "IOAPIC RTE = 0x%x:%x(V) -> 0x%x:%x(P)", rte >> 32, (uint32_t)rte, high, low); rte = high; rte = rte << 32 | low; } else { enum vpic_trigger trigger; int phys_irq = pin_to_irq(info->phys_pin); uint64_t native_rte; /* just update trigger mode */ ioapic_get_rte(phys_irq, &native_rte); rte = native_rte; rte &= ~IOAPIC_RTE_TRGRMOD; vpic_get_irq_trigger(vm, info->virt_pin, &trigger); if (trigger == LEVEL_TRIGGER) rte |= IOAPIC_RTE_TRGRLVL; dev_dbg(ACRN_DBG_IRQ, "IOAPIC RTE = 0x%x:%x(P) -> 0x%x:%x(P)", native_rte >> 32, (uint32_t)native_rte, rte >> 32, (uint32_t)rte); } return rte; } static struct ptdev_remapping_info * add_msix_remapping(struct vm *vm, uint16_t virt_bdf, uint16_t phys_bdf, int msix_entry_index) { struct ptdev_remapping_info *entry; spinlock_obtain(&vm->ptdev_lock); entry = _get_remapping_entry(vm, entry_id_from_msix(virt_bdf, msix_entry_index)); if (!entry) { entry = alloc_entry(vm, PTDEV_INTR_MSI); entry->virt_bdf = virt_bdf; entry->phys_bdf = phys_bdf; entry->msi.msix_entry_index = msix_entry_index; } spinlock_release(&vm->ptdev_lock); dev_dbg(ACRN_DBG_IRQ, "VM%d MSIX add vector mapping vbdf%x:pbdf%x idx=%d", entry->vm->attr.id, virt_bdf, phys_bdf, msix_entry_index); return entry; } /* deactive & remove mapping entry of msix_entry_index for virt_bdf */ static void remove_msix_remapping(struct vm *vm, uint16_t virt_bdf, int msix_entry_index) { struct ptdev_remapping_info *entry; spinlock_obtain(&vm->ptdev_lock); entry = _get_remapping_entry(vm, entry_id_from_msix(virt_bdf, msix_entry_index)); if (!entry) goto END; if (entry_is_active(entry)) /*TODO: disable MSIX device when HV can in future */ ptdev_deactivate_entry(entry); dev_dbg(ACRN_DBG_IRQ, "VM%d MSIX remove vector mapping vbdf-pbdf:0x%x-0x%x idx=%d", entry->vm->attr.id, virt_bdf, entry->phys_bdf, msix_entry_index); release_entry(entry); END: spinlock_release(&vm->ptdev_lock); } /* install virt to phys pin mapping if not existing */ static struct ptdev_remapping_info * add_intx_remapping(struct vm *vm, uint8_t virt_pin, uint8_t phys_pin, bool pic_pin) { struct ptdev_remapping_info *entry; enum ptdev_vpin_source vpin_src = pic_pin ? PTDEV_VPIN_PIC : PTDEV_VPIN_IOAPIC; spinlock_obtain(&vm->ptdev_lock); entry = _get_remapping_entry(vm, entry_id_from_intx(virt_pin, vpin_src)); if (!entry) { entry = alloc_entry(vm, PTDEV_INTR_INTX); entry->intx.vpin_src = vpin_src; } /* update existing */ entry->intx.virt_pin = virt_pin; entry->intx.phys_pin = phys_pin; spinlock_release(&vm->ptdev_lock); dev_dbg(ACRN_DBG_IRQ, "VM%d INTX add pin mapping vpin%d:ppin%d", entry->vm->attr.id, virt_pin, phys_pin); return entry; } /* deactive & remove mapping entry of virt_pin */ void remove_intx_remapping(struct vm *vm, uint8_t virt_pin, bool pic_pin) { int phys_irq, phys_pin; struct ptdev_remapping_info *entry; enum ptdev_vpin_source vpin_src = pic_pin ? PTDEV_VPIN_PIC : PTDEV_VPIN_IOAPIC; spinlock_obtain(&vm->ptdev_lock); entry = _get_remapping_entry(vm, entry_id_from_intx(virt_pin, vpin_src)); if (!entry) goto END; if (entry_is_active(entry)) { phys_pin = entry->intx.phys_pin; phys_irq = pin_to_irq(phys_pin); if (!irq_is_gsi(phys_irq)) goto END; /* disable interrupt */ GSI_MASK_IRQ(phys_irq); ptdev_deactivate_entry(entry); dev_dbg(ACRN_DBG_IRQ, "deactive %s intx entry:ppin=%d, pirq=%d ", pic_pin ? "PIC" : "IOAPIC", phys_pin, phys_irq); dev_dbg(ACRN_DBG_IRQ, "from vm%d vpin=%d\n", entry->vm->attr.id, virt_pin); } release_entry(entry); END: spinlock_release(&vm->ptdev_lock); } static void ptdev_intr_handle_irq(struct vm *vm, struct ptdev_remapping_info *entry) { switch (entry->intx.vpin_src) { case PTDEV_VPIN_IOAPIC: { uint64_t rte; bool trigger_lvl = false; /* VPIN_IOAPIC src means we have vioapic enabled */ vioapic_get_rte(vm, entry->intx.virt_pin, &rte); if ((rte & IOAPIC_RTE_TRGRMOD) == IOAPIC_RTE_TRGRLVL) trigger_lvl = true; if (trigger_lvl) vioapic_assert_irq(vm, entry->intx.virt_pin); else vioapic_pulse_irq(vm, entry->intx.virt_pin); dev_dbg(ACRN_DBG_PTIRQ, "dev-assign: irq=0x%x assert vr: 0x%x vRTE=0x%x", dev_to_irq(entry->node), irq_to_vector(dev_to_irq(entry->node)), rte); break; } case PTDEV_VPIN_PIC: { enum vpic_trigger trigger; /* VPIN_PIC src means we have vpic enabled */ vpic_get_irq_trigger(vm, entry->intx.virt_pin, &trigger); if (trigger == LEVEL_TRIGGER) vpic_assert_irq(vm, entry->intx.virt_pin); else vpic_pulse_irq(vm, entry->intx.virt_pin); break; } default: break; } } void ptdev_intx_ack(struct vm *vm, int virt_pin, enum ptdev_vpin_source vpin_src) { int phys_irq; struct ptdev_remapping_info *entry; int phys_pin; entry = get_remapping_entry(vm, entry_id_from_intx(virt_pin, vpin_src)); if (!entry) return; ASSERT(vpin_src == entry->intx.vpin_src, "PIN source not match"); phys_pin = entry->intx.phys_pin; phys_irq = pin_to_irq(phys_pin); if (!irq_is_gsi(phys_irq)) return; /* NOTE: only Level trigger will process EOI/ACK and if we got here * means we have this vioapic or vpic or both enabled */ switch (entry->intx.vpin_src) { case PTDEV_VPIN_IOAPIC: vioapic_deassert_irq(vm, virt_pin); break; case PTDEV_VPIN_PIC: vpic_deassert_irq(vm, virt_pin); default: break; } dev_dbg(ACRN_DBG_PTIRQ, "dev-assign: irq=0x%x acked vr: 0x%x", phys_irq, irq_to_vector(phys_irq)); GSI_UNMASK_IRQ(phys_irq); } /* Main entry for PCI device assignment with MSI and MSI-X * MSI can up to 8 vectors and MSI-X can up to 1024 Vectors * We use msix_entry_index to indicate coming vectors * msix_entry_index = 0 means first vector * user must provide bdf and msix_entry_index */ int ptdev_msix_remap(struct vm *vm, uint16_t virt_bdf, struct ptdev_msi_info *info) { struct ptdev_remapping_info *entry; bool lowpri = !is_vm0(vm); entry = get_remapping_entry(vm, entry_id_from_msix(virt_bdf, info->msix_entry_index)); if (!entry) { /* VM0 we add mapping dynamically */ if (is_vm0(vm)) { entry = add_msix_remapping(vm, virt_bdf, virt_bdf, info->msix_entry_index); } else { pr_err("dev-assign: msi entry not exist"); return -1; } } /* handle destroy case */ if (entry_is_active(entry) && info->vmsi_data == 0) { info->pmsi_data = 0; ptdev_deactivate_entry(entry); goto END; } if (!entry_is_active(entry)) { /* update msi source and active entry */ ptdev_activate_entry(entry, -1, lowpri); } /* build native config MSI, update to info->pmsi_xxx */ ptdev_build_native_msi(vm, info, dev_to_vector(entry->node)); entry->msi = *info; entry->msi.virt_vector = info->vmsi_data & 0xFF; entry->msi.phys_vector = dev_to_vector(entry->node); /* update irq handler according to info in guest */ ptdev_update_irq_handler(vm, entry); dev_dbg(ACRN_DBG_IRQ, "PCI %x:%x.%x MSI VR[%d] 0x%x->0x%x assigned to vm%d", (entry->virt_bdf >> 8) & 0xFF, (entry->virt_bdf >> 3) & 0x1F, (entry->virt_bdf) & 0x7, entry->msi.msix_entry_index, entry->msi.virt_vector, entry->msi.phys_vector, entry->vm->attr.id); END: return 0; } static bool vpin_masked(struct vm *vm, uint8_t virt_pin, enum ptdev_vpin_source vpin_src) { if (vpin_src == PTDEV_VPIN_IOAPIC) { uint64_t rte; vioapic_get_rte(vm, virt_pin, &rte); if ((rte & IOAPIC_RTE_INTMASK) == IOAPIC_RTE_INTMSET) return true; else return false; } else return vpic_is_pin_mask(vm->vpic, virt_pin); } static void update_active_native_ioapic(struct vm *vm, struct ptdev_intx_info *info, struct ptdev_remapping_info *entry, int phys_irq) { uint64_t rte; /* disable interrupt */ GSI_MASK_IRQ(phys_irq); /* build native IOAPIC RTE, update to info->rte */ rte = ptdev_build_native_rte(vm, info, dev_to_vector(entry->node)); /* set rte entry */ GSI_SET_RTE(phys_irq, rte | IOAPIC_RTE_INTMSET); /* update irq handler according to info in guest */ ptdev_update_irq_handler(vm, entry); /* enable interrupt */ GSI_UNMASK_IRQ(phys_irq); } /* Main entry for PCI/Legacy device assignment with INTx */ int ptdev_intx_pin_remap(struct vm *vm, struct ptdev_intx_info *info) { struct ptdev_remapping_info *entry; uint64_t rte; int phys_irq; int phys_pin; bool lowpri = !is_vm0(vm); /* * virt pin could come from vpic master, vpic slave or vioapic * while phys pin is always means for native IOAPIC * Device Model should tell us the mapping information */ /* no fix for native owned intx */ if (ptdev_native_owned_intx(vm, info)) goto END; /* query if we have virt to phys mapping */ entry = get_remapping_entry(vm, entry_id_from_intx(info->virt_pin, info->vpin_src)); if (!entry) { /* allocate entry during first unmask */ if (vpin_masked(vm, info->virt_pin, info->vpin_src)) goto END; if (is_vm0(vm)) { bool pic_pin = (info->vpin_src == PTDEV_VPIN_PIC); info->phys_pin = info->virt_pin; /* fix vPIC pin to correct native IOAPIC pin */ if (pic_pin) info->phys_pin = legacy_irq_to_pin[info->virt_pin]; entry = add_intx_remapping(vm, info->virt_pin, info->phys_pin, pic_pin); } else goto END; } /* phys_pin from native IOAPIC */ phys_pin = entry->intx.phys_pin; phys_irq = pin_to_irq(phys_pin); if (!irq_is_gsi(phys_irq)) goto END; if (entry_is_active(entry) && (info->vpin_src == PTDEV_VPIN_IOAPIC)) { /* SOS/UOS vIOAPIC pin will be dynamically free/assign * while vPIC pin is static assigned */ vioapic_get_rte(vm, info->virt_pin, &rte); if (((uint32_t)rte) == 0x10000) { /* disable interrupt */ GSI_MASK_IRQ(phys_irq); ptdev_deactivate_entry(entry); dev_dbg(ACRN_DBG_IRQ, "IOAPIC pin=%d pirq=%d deassigned ", phys_pin, phys_irq); dev_dbg(ACRN_DBG_IRQ, "from vm%d vIOAPIC vpin=%d", entry->vm->attr.id, info->virt_pin); goto END; } else { /*update rte*/ update_active_native_ioapic(vm, info, entry, phys_irq); } } else if (entry_is_active(entry) && info->vpin_src == PTDEV_VPIN_PIC) { /* only update here * deactive vPIC entry when IOAPIC take it over */ update_active_native_ioapic(vm, info, entry, phys_irq); } else { /* * both vIOAPIC & vPIC take native IOAPIC path * vIOAPIC: build native RTE according vIOAPIC configuration * vPIC: keep native RTE configuration in setup_ioapic_irq() */ if (info->vpin_src == PTDEV_VPIN_IOAPIC) check_deactive_pic_intx(vm, phys_pin); /* active entry */ ptdev_activate_entry(entry, phys_irq, lowpri); update_active_native_ioapic(vm, info, entry, phys_irq); dev_dbg(ACRN_DBG_IRQ, "IOAPIC pin=%d pirq=%d assigned to vm%d %s vpin=%d", phys_pin, phys_irq, entry->vm->attr.id, info->vpin_src == PTDEV_VPIN_PIC ? "vPIC" : "vIOAPIC", info->virt_pin); } END: return 0; } void ptdev_softirq(__unused int cpu) { while (1) { struct ptdev_remapping_info *entry = ptdev_dequeue_softirq(); struct vm *vm; if (!entry) break; /* skip any inactive entry */ if (!entry_is_active(entry)) { /* service next item */ continue; } /* TBD: need valid vm */ vm = entry->vm; /* handle real request */ if (entry->type == PTDEV_INTR_INTX) ptdev_intr_handle_irq(vm, entry); else { /* TODO: msi destmode check required */ vlapic_intr_msi(vm, entry->msi.vmsi_addr, entry->msi.vmsi_data); dev_dbg(ACRN_DBG_PTIRQ, "dev-assign: irq=0x%x MSI VR: 0x%x-0x%x", dev_to_irq(entry->node), entry->msi.virt_vector, irq_to_vector(dev_to_irq(entry->node))); dev_dbg(ACRN_DBG_PTIRQ, " vmsi_addr: 0x%x vmsi_data: 0x%x", entry->msi.vmsi_addr, entry->msi.vmsi_data); } } } void ptdev_init(void) { if (get_cpu_id() > 0) return; INIT_LIST_HEAD(&softirq_dev_entry_list); spinlock_init(&softirq_dev_lock); } void ptdev_vm_init(struct vm *vm) { INIT_LIST_HEAD(&vm->ptdev_list); spinlock_init(&vm->ptdev_lock); } void ptdev_vm_deinit(struct vm *vm) { /* VM already down */ spinlock_obtain(&vm->ptdev_lock); release_all_entry(vm); spinlock_release(&vm->ptdev_lock); } void ptdev_add_intx_remapping(struct vm *vm, __unused uint16_t virt_bdf, __unused uint16_t phys_bdf, uint8_t virt_pin, uint8_t phys_pin, bool pic_pin) { if (vm == NULL) { pr_err("ptdev_add_intx_remapping fails!\n"); return; } add_intx_remapping(vm, virt_pin, phys_pin, pic_pin); } void ptdev_remove_intx_remapping(struct vm *vm, uint8_t virt_pin, bool pic_pin) { if (vm == NULL) { pr_err("ptdev_remove_intr_remapping fails!\n"); return; } remove_intx_remapping(vm, virt_pin, pic_pin); } void ptdev_add_msix_remapping(struct vm *vm, uint16_t virt_bdf, uint16_t phys_bdf, int vector_count) { int i; for (i = 0; i < vector_count; i++) add_msix_remapping(vm, virt_bdf, phys_bdf, i); } void ptdev_remove_msix_remapping(struct vm *vm, uint16_t virt_bdf, int vector_count) { int i; if (vm == NULL) { pr_err("ptdev_remove_msix_remapping fails!\n"); return; } for (i = 0; i < vector_count; i++) remove_msix_remapping(vm, virt_bdf, i); } static void get_entry_info(struct ptdev_remapping_info *entry, char *type, int *irq, int *vector, uint64_t *dest, bool *lvl_tm, int *pin, int *vpin, int *bdf, int *vbdf) { if (entry_is_active(entry)) { if (entry->type == PTDEV_INTR_MSI) { strcpy_s(type, 16, "MSI"); *dest = (entry->msi.pmsi_addr & 0xFF000) >> 12; if (entry->msi.pmsi_data & APIC_TRIGMOD_LEVEL) *lvl_tm = true; else *lvl_tm = false; *pin = -1; *vpin = -1; *bdf = entry->phys_bdf; *vbdf = entry->virt_bdf; } else { int phys_irq = pin_to_irq(entry->intx.phys_pin); uint64_t rte = 0; if (entry->intx.vpin_src == PTDEV_VPIN_IOAPIC) strcpy_s(type, 16, "IOAPIC"); else strcpy_s(type, 16, "PIC"); ioapic_get_rte(phys_irq, &rte); *dest = ((rte >> 32) & IOAPIC_RTE_DEST) >> 24; if (rte & IOAPIC_RTE_TRGRLVL) *lvl_tm = true; else *lvl_tm = false; *pin = entry->intx.phys_pin; *vpin = entry->intx.virt_pin; *bdf = 0; *vbdf = 0; } *irq = dev_to_irq(entry->node); *vector = dev_to_vector(entry->node); } else { strcpy_s(type, 16, "NONE"); *irq = -1; *vector = 0; *dest = 0; *lvl_tm = 0; *pin = -1; *vpin = -1; *bdf = 0; *vbdf = 0; } } int get_ptdev_info(char *str, int str_max) { struct ptdev_remapping_info *entry; int len, size = str_max, irq, vector; char type[16]; uint64_t dest; bool lvl_tm; int pin, vpin, bdf, vbdf; struct list_head *pos = NULL, *vm_pos; struct vm *vm; len = snprintf(str, size, "\r\nVM\tTYPE\tIRQ\tVEC\tDEST\tTM\tPIN\tVPIN\tBDF\tVBDF"); size -= len; str += len; spinlock_obtain(&vm_list_lock); list_for_each(vm_pos, &vm_list) { vm = list_entry(pos, struct vm, list); spinlock_obtain(&vm->ptdev_lock); list_for_each(pos, &vm->ptdev_list) { entry = list_entry(pos, struct ptdev_remapping_info, entry_node); if (entry_is_active(entry)) { get_entry_info(entry, type, &irq, &vector, &dest, &lvl_tm, &pin, &vpin, &bdf, &vbdf); len = snprintf(str, size, "\r\n%d\t%s\t%d\t0x%X\t0x%X", entry->vm->attr.id, type, irq, vector, dest); size -= len; str += len; len = snprintf(str, size, "\t%s\t%d\t%d\t%x:%x.%x\t%x:%x.%x", entry_is_active(entry) ? (lvl_tm ? "level" : "edge") : "none", pin, vpin, (bdf & 0xff00) >> 8, (bdf & 0xf8) >> 3, bdf & 0x7, (vbdf & 0xff00) >> 8, (vbdf & 0xf8) >> 3, vbdf & 0x7); size -= len; str += len; } } spinlock_release(&vm->ptdev_lock); } spinlock_release(&vm_list_lock); snprintf(str, size, "\r\n"); return 0; }