/* * Copyright (c) 2011 NetApp, Inc. * Copyright (c) 2018 Intel Corporation * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * $FreeBSD$ */ #include #include "pci_priv.h" static inline bool msixcap_access(struct pci_vdev *vdev, uint32_t offset) { if (vdev->msix.capoff == 0U) { return 0; } return in_range(offset, vdev->msix.capoff, vdev->msix.caplen); } static inline bool msixtable_access(struct pci_vdev *vdev, uint32_t offset) { return in_range(offset, vdev->msix.table_offset, vdev->msix.table_count * MSIX_TABLE_ENTRY_SIZE); } static int vmsix_remap_entry(struct pci_vdev *vdev, uint32_t index, bool enable) { struct msix_table_entry *pentry; struct ptdev_msi_info info; uint64_t hva; int ret; info.is_msix = 1; info.vmsi_addr = vdev->msix.tables[index].addr; info.vmsi_data = (enable) ? vdev->msix.tables[index].data : 0U; ret = ptdev_msix_remap(vdev->vpci->vm, vdev->vbdf.value, index, &info); if (ret != 0) { return ret; } /* Write the table entry to the physical structure */ hva = vdev->msix.mmio_hva + vdev->msix.table_offset; pentry = (struct msix_table_entry *)hva + index; pentry->addr = info.pmsi_addr; pentry->data = info.pmsi_data; pentry->vector_control = vdev->msix.tables[index].vector_control; return ret; } static inline void enable_disable_msix(struct pci_vdev *vdev, bool enable) { uint32_t msgctrl; msgctrl = pci_vdev_read_cfg(vdev, vdev->msix.capoff + PCIR_MSIX_CTRL, 2U); if (enable) { msgctrl |= PCIM_MSIXCTRL_MSIX_ENABLE; } else { msgctrl &= ~PCIM_MSIXCTRL_MSIX_ENABLE; } pci_pdev_write_cfg(vdev->pdev.bdf, vdev->msix.capoff + PCIR_MSIX_CTRL, 2U, msgctrl); } /* Do MSI-X remap for all MSI-X table entries in the target device */ static int vmsix_remap(struct pci_vdev *vdev, bool enable) { uint32_t index; int ret; /* disable MSI-X during configuration */ enable_disable_msix(vdev, false); for (index = 0U; index < vdev->msix.table_count; index++) { ret = vmsix_remap_entry(vdev, index, enable); if (ret != 0) { return ret; } } /* If MSI Enable is being set, make sure INTxDIS bit is set */ if (enable) { enable_disable_pci_intx(vdev->pdev.bdf, false); } enable_disable_msix(vdev, enable); return 0; } /* Do MSI-X remap for one MSI-X table entry only */ static int vmsix_remap_one_entry(struct pci_vdev *vdev, uint32_t index, bool enable) { uint32_t msgctrl; int ret; /* disable MSI-X during configuration */ enable_disable_msix(vdev, false); ret = vmsix_remap_entry(vdev, index, enable); if (ret != 0) { return ret; } /* If MSI Enable is being set, make sure INTxDIS bit is set */ if (enable) { enable_disable_pci_intx(vdev->pdev.bdf, false); } /* Restore MSI-X Enable bit */ msgctrl = pci_vdev_read_cfg(vdev, vdev->msix.capoff + PCIR_MSIX_CTRL, 2U); if ((msgctrl & PCIM_MSIXCTRL_MSIX_ENABLE) == PCIM_MSIXCTRL_MSIX_ENABLE) { pci_pdev_write_cfg(vdev->pdev.bdf, vdev->msix.capoff + PCIR_MSIX_CTRL, 2U, msgctrl); } return ret; } static int vmsix_cfgread(struct pci_vdev *vdev, uint32_t offset, uint32_t bytes, uint32_t *val) { /* For PIO access, we emulate Capability Structures only */ if (msixcap_access(vdev, offset)) { *val = pci_vdev_read_cfg(vdev, offset, bytes); return 0; } return -ENODEV; } static int vmsix_cfgwrite(struct pci_vdev *vdev, uint32_t offset, uint32_t bytes, uint32_t val) { uint32_t msgctrl; /* Writing MSI-X Capability Structure */ if (msixcap_access(vdev, offset)) { msgctrl = pci_vdev_read_cfg(vdev, vdev->msix.capoff + PCIR_MSIX_CTRL, 2U); /* Write to vdev */ pci_vdev_write_cfg(vdev, offset, bytes, val); /* Writing Message Control field? */ if ((offset - vdev->msix.capoff) == PCIR_MSIX_CTRL) { if (((msgctrl ^ val) & PCIM_MSIXCTRL_MSIX_ENABLE) != 0U) { if (val & PCIM_MSIXCTRL_MSIX_ENABLE) { (void)vmsix_remap(vdev, true); } else { (void)vmsix_remap(vdev, false); } } if (((msgctrl ^ val) & PCIM_MSIXCTRL_FUNCTION_MASK) != 0U) { pci_pdev_write_cfg(vdev->pdev.bdf, offset, 2U, msgctrl); } } return 0; } return -ENODEV; } static void vmsix_table_rw(struct pci_vdev *vdev, struct mmio_request *mmio, uint32_t offset) { struct msix_table_entry *entry; uint32_t vector_control, entry_offset, index; bool message_changed = false; bool unmasked; /* Find out which entry it's accessing */ offset -= vdev->msix.table_offset; index = offset / MSIX_TABLE_ENTRY_SIZE; if (index >= vdev->msix.table_count) { pr_err("%s, invalid arguments %llx - %llx", __func__, mmio->value, mmio->address); return; } entry = &vdev->msix.tables[index]; entry_offset = offset % MSIX_TABLE_ENTRY_SIZE; if (mmio->direction == REQUEST_READ) { (void)memcpy_s(&mmio->value, (size_t)mmio->size, (void *)entry + entry_offset, (size_t)mmio->size); } else { /* Only DWORD and QWORD are permitted */ if ((mmio->size != 4U) && (mmio->size != 8U)) { pr_err("%s, Only DWORD and QWORD are permitted", __func__); return; } /* Save for comparison */ vector_control = entry->vector_control; /* Writing different value to Message Data/Addr? */ if (((offsetof(struct msix_table_entry, addr) == entry_offset) && (entry->addr != mmio->value)) || ((offsetof(struct msix_table_entry, data) == entry_offset) && (entry->data != (uint32_t)mmio->value))) { message_changed = true; } /* Write to pci_vdev */ (void)memcpy_s((void *)entry + entry_offset, (size_t)mmio->size, &mmio->value, (size_t)mmio->size); /* If MSI-X hasn't been enabled, do nothing */ if ((pci_vdev_read_cfg(vdev, vdev->msix.capoff + PCIR_MSIX_CTRL, 2U) & PCIM_MSIXCTRL_MSIX_ENABLE) == PCIM_MSIXCTRL_MSIX_ENABLE) { if ((((entry->vector_control ^ vector_control) & PCIM_MSIX_VCTRL_MASK) != 0U) || message_changed) { unmasked = ((entry->vector_control & PCIM_MSIX_VCTRL_MASK) == 0U); (void)vmsix_remap_one_entry(vdev, index, unmasked); } } } } static int vmsix_table_mmio_access_handler(struct io_request *io_req, void *handler_private_data) { struct mmio_request *mmio = &io_req->reqs.mmio; struct pci_vdev *vdev; uint32_t offset; uint64_t hva; vdev = (struct pci_vdev *)handler_private_data; offset = (uint32_t)(mmio->address - vdev->msix.mmio_gpa); if (msixtable_access(vdev, offset)) { vmsix_table_rw(vdev, mmio, offset); } else { hva = vdev->msix.mmio_hva + offset; /* MSI-X PBA and Capability Table could be in the same range */ if (mmio->direction == REQUEST_READ) { (void)memcpy_s(&mmio->value, (size_t)mmio->size, (const void *)hva, (size_t)mmio->size); } else { (void)memcpy_s((void *)hva, (size_t)mmio->size, &mmio->value, (size_t)mmio->size); } } return 0; } static void decode_msix_table_bar(struct pci_vdev *vdev) { uint32_t bir = vdev->msix.table_bar; union pci_bdf pbdf = vdev->pdev.bdf; uint64_t base, size; uint32_t bar_lo, bar_hi; bar_lo = pci_pdev_read_cfg(pbdf, pci_bar_offset(bir), 4U); if ((bar_lo & PCIM_BAR_SPACE) == PCIM_BAR_IO_SPACE) { /* I/O bar, should never happen */ pr_err("PCI device (%x) has MSI-X Table at IO BAR", vdev->vbdf.value); return; } /* Get the base address */ base = (uint64_t)(bar_lo & PCIM_BAR_MEM_BASE); if ((bar_lo & PCIM_BAR_MEM_TYPE) == PCIM_BAR_MEM_64) { bar_hi = pci_pdev_read_cfg(pbdf, pci_bar_offset(bir + 1U), 4U); base |= ((uint64_t)bar_hi << 32U); } vdev->msix.mmio_hva = (uint64_t)hpa2hva(base); vdev->msix.mmio_gpa = vm0_hpa2gpa(base); /* Sizing the BAR */ pci_pdev_write_cfg(pbdf, pci_bar_offset(bir), 4U, ~0U); size = pci_pdev_read_cfg(pbdf, pci_bar_offset(bir), 4U); vdev->msix.mmio_size = (size & ~(size - 1U)); if ((bar_lo & PCIM_BAR_MEM_TYPE) == PCIM_BAR_MEM_64) { pci_pdev_write_cfg(pbdf, pci_bar_offset(bir + 1U), 4U, ~0U); size = (uint64_t)pci_pdev_read_cfg(pbdf, pci_bar_offset(bir + 1U), 4U); vdev->msix.mmio_size |= (size << 32U); } /* Restore the BAR */ pci_pdev_write_cfg(pbdf, pci_bar_offset(bir), 4U, bar_lo); if ((bar_lo & PCIM_BAR_MEM_TYPE) == PCIM_BAR_MEM_64) { pci_pdev_write_cfg(pbdf, pci_bar_offset(bir + 1U), 4U, bar_hi); } } static int vmsix_init(struct pci_vdev *vdev) { uint32_t msgctrl; uint32_t table_info, i; struct msix *msix = &vdev->msix; msgctrl = pci_pdev_read_cfg(vdev->pdev.bdf, vdev->msix.capoff + PCIR_MSIX_CTRL, 2U); /* Read Table Offset and Table BIR */ table_info = pci_pdev_read_cfg(vdev->pdev.bdf, msix->capoff + PCIR_MSIX_TABLE, 4U); msix->table_bar = table_info & PCIM_MSIX_BIR_MASK; msix->table_offset = table_info & ~PCIM_MSIX_BIR_MASK; msix->table_count = (msgctrl & PCIM_MSIXCTRL_TABLE_SIZE) + 1U; /* Mask all table entries */ for (i = 0U; i < msix->table_count; i++) { msix->tables[i].vector_control |= PCIM_MSIX_VCTRL_MASK; } decode_msix_table_bar(vdev); /* * MSI-X table structures is in a 4 KB aligned range, * while it's possible that the MSI-X PBA co-reside within this * naturally aligned 4 KB address range */ if (msix->mmio_gpa != 0U) { (void)register_mmio_emulation_handler(vdev->vpci->vm, vmsix_table_mmio_access_handler, msix->mmio_gpa, msix->mmio_gpa + msix->mmio_size, vdev); } return 0; } static int vmsix_deinit(struct pci_vdev *vdev) { if (vdev->msix.mmio_gpa != 0UL) { unregister_mmio_emulation_handler(vdev->vpci->vm, vdev->msix.mmio_gpa, vdev->msix.mmio_gpa + vdev->msix.mmio_size); vdev->msix.mmio_gpa = 0U; } if (vdev->msix.table_count != 0U) { ptdev_remove_msix_remapping(vdev->vpci->vm, vdev->vbdf.value, vdev->msix.table_count); } return 0; } struct pci_vdev_ops pci_ops_vdev_msix = { .init = vmsix_init, .deinit = vmsix_deinit, .cfgwrite = vmsix_cfgwrite, .cfgread = vmsix_cfgread, };