/* * Copyright (c) 2011 NetApp, Inc. * Copyright (c) 2018 Intel Corporation * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * $FreeBSD$ */ #include #include "pci_priv.h" static uint32_t num_pci_vdev; static struct pci_vdev sharing_mode_vdev_array[CONFIG_MAX_PCI_DEV_NUM]; struct pci_vdev *sharing_mode_find_vdev(union pci_bdf pbdf) { struct pci_vdev *vdev = NULL; uint32_t i; /* in VM0, it uses phys BDF */ for (i = 0U; i < num_pci_vdev; i++) { if (sharing_mode_vdev_array[i].pdev.bdf.value == pbdf.value) { vdev = &sharing_mode_vdev_array[i]; } } return vdev; } static void sharing_mode_cfgread(__unused struct acrn_vpci *vpci, union pci_bdf bdf, uint32_t offset, uint32_t bytes, uint32_t *val) { struct pci_vdev *vdev; bool handled = false; uint32_t i; vdev = sharing_mode_find_vdev(bdf); /* vdev == NULL: Could be hit for PCI enumeration from guests */ if ((vdev == NULL) || ((bytes != 1U) && (bytes != 2U) && (bytes != 4U))) { *val = ~0U; } else { for (i = 0U; (i < vdev->nr_ops) && !handled; i++) { if (vdev->ops[i].cfgread != NULL) { if (vdev->ops[i].cfgread(vdev, offset, bytes, val) == 0) { handled = true; } } } /* Not handled by any handlers. Passthru to physical device */ if (!handled) { *val = pci_pdev_read_cfg(vdev->pdev.bdf, offset, bytes); } } } static void sharing_mode_cfgwrite(__unused struct acrn_vpci *vpci, union pci_bdf bdf, uint32_t offset, uint32_t bytes, uint32_t val) { struct pci_vdev *vdev; bool handled = false; uint32_t i; if ((bytes == 1U) || (bytes == 2U) || (bytes == 4U)) { vdev = sharing_mode_find_vdev(bdf); if (vdev != NULL) { for (i = 0U; (i < vdev->nr_ops) && !handled; i++) { if (vdev->ops[i].cfgwrite != NULL) { if (vdev->ops[i].cfgwrite(vdev, offset, bytes, val) == 0) { handled = true; } } } /* Not handled by any handlers. Passthru to physical device */ if (!handled) { pci_pdev_write_cfg(vdev->pdev.bdf, offset, bytes, val); } } } } static struct pci_vdev *alloc_pci_vdev(const struct acrn_vm *vm, union pci_bdf bdf) { struct pci_vdev *vdev; if (num_pci_vdev < CONFIG_MAX_PCI_DEV_NUM) { vdev = &sharing_mode_vdev_array[num_pci_vdev]; num_pci_vdev++; /* vbdf equals to pbdf otherwise remapped */ vdev->vbdf = bdf; vdev->vpci = &vm->vpci; vdev->pdev.bdf = bdf; } else { vdev = NULL; } return vdev; } static void enumerate_pci_dev(uint16_t pbdf, const void *cb_data) { const struct acrn_vm *vm = (const struct acrn_vm *)cb_data; struct pci_vdev *vdev; vdev = alloc_pci_vdev(vm, (union pci_bdf)pbdf); if (vdev != NULL) { populate_msi_struct(vdev); } } static int32_t sharing_mode_vpci_init(const struct acrn_vm *vm) { struct pci_vdev *vdev; uint32_t i, j; int32_t ret; /* * Only setup IO bitmap for SOS. * IO/MMIO requests from non-vm0 guests will be injected to device model. */ if (!is_vm0(vm)) { ret = -ENODEV; } else { /* Initialize PCI vdev array */ num_pci_vdev = 0U; (void)memset((void *)sharing_mode_vdev_array, 0U, sizeof(sharing_mode_vdev_array)); /* build up vdev array for vm0 */ pci_scan_bus(enumerate_pci_dev, vm); for (i = 0U; i < num_pci_vdev; i++) { vdev = &sharing_mode_vdev_array[i]; for (j = 0U; j < vdev->nr_ops; j++) { if (vdev->ops[j].init != NULL) { (void)vdev->ops[j].init(vdev); } } } ret = 0; } return ret; } static void sharing_mode_vpci_deinit(__unused const struct acrn_vm *vm) { struct pci_vdev *vdev; uint32_t i, j; if (is_vm0(vm)) { for (i = 0U; i < num_pci_vdev; i++) { vdev = &sharing_mode_vdev_array[i]; for (j = 0U; j < vdev->nr_ops; j++) { if (vdev->ops[j].deinit != NULL) { (void)vdev->ops[j].deinit(vdev); } } } } } void add_vdev_handler(struct pci_vdev *vdev, const struct pci_vdev_ops *ops) { if (vdev->nr_ops >= (MAX_VPCI_DEV_OPS - 1U)) { pr_err("%s, adding too many handlers", __func__); } else { vdev->ops[vdev->nr_ops] = *ops; vdev->nr_ops++; } } const struct vpci_ops sharing_mode_vpci_ops = { .init = sharing_mode_vpci_init, .deinit = sharing_mode_vpci_deinit, .cfgread = sharing_mode_cfgread, .cfgwrite = sharing_mode_cfgwrite, }; void vpci_set_ptdev_intr_info(const struct acrn_vm *target_vm, uint16_t vbdf, uint16_t pbdf) { struct pci_vdev *vdev; vdev = sharing_mode_find_vdev((union pci_bdf)pbdf); if (vdev == NULL) { pr_err("%s, can't find PCI device for vm%d, vbdf (0x%x) pbdf (0x%x)", __func__, target_vm->vm_id, vbdf, pbdf); } else { /* UOS may do BDF mapping */ vdev->vpci = (struct acrn_vpci *)&(target_vm->vpci); vdev->vbdf.value = vbdf; vdev->pdev.bdf.value = pbdf; } } void vpci_reset_ptdev_intr_info(const struct acrn_vm *target_vm, uint16_t vbdf, uint16_t pbdf) { struct pci_vdev *vdev; struct acrn_vm *vm; vdev = sharing_mode_find_vdev((union pci_bdf)pbdf); if (vdev == NULL) { pr_err("%s, can't find PCI device for vm%d, vbdf (0x%x) pbdf (0x%x)", __func__, target_vm->vm_id, vbdf, pbdf); } else { /* Return this PCI device to SOS */ if (vdev->vpci->vm == target_vm) { vm = get_vm_from_vmid(0U); if (vm != NULL) { vdev->vpci = &vm->vpci; } } } }