mirror of
https://github.com/projectacrn/acrn-hypervisor.git
synced 2025-05-01 05:03:55 +00:00
In this case, we could handle all the passthrough PCI devices in ACRN hypervisor. But we still need DM to initialize BAR resources and Intx for passthrough PCI device for post-launched VM since these informations should been filled into ACPI tables. So 1. we add a HC vm_assign_pcidev to pass the extra informations to replace the old vm_assign_ptdev. 2. we saso remove HC vm_set_ptdev_msix_info since it could been setted by the post-launched VM now same as SOS. 3. remove vm_map_ptdev_mmio call for PTDev in DM since ACRN hypervisor will handle these BAR access. 4. the most important thing is to trap PCI configure space access for PTDev in HV for post-launched VM and bypass the virtual PCI device configure space access to DM. This patch doesn't do the clean work. Will do it in the next patch. Tracked-On: #4371 Signed-off-by: Li Fei1 <fei1.li@intel.com>
168 lines
4.2 KiB
C
168 lines
4.2 KiB
C
/*
|
|
* Copyright (c) 2011 NetApp, Inc.
|
|
* Copyright (c) 2018 Intel Corporation
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#include <vm.h>
|
|
#include "vpci_priv.h"
|
|
#include <ept.h>
|
|
#include <logmsg.h>
|
|
|
|
/**
|
|
* @pre vdev != NULL
|
|
*/
|
|
uint32_t pci_vdev_read_cfg(const struct pci_vdev *vdev, uint32_t offset, uint32_t bytes)
|
|
{
|
|
uint32_t val;
|
|
|
|
switch (bytes) {
|
|
case 1U:
|
|
val = pci_vdev_read_cfg_u8(vdev, offset);
|
|
break;
|
|
case 2U:
|
|
val = pci_vdev_read_cfg_u16(vdev, offset);
|
|
break;
|
|
default:
|
|
val = pci_vdev_read_cfg_u32(vdev, offset);
|
|
break;
|
|
}
|
|
|
|
return val;
|
|
}
|
|
|
|
/**
|
|
* @pre vdev != NULL
|
|
*/
|
|
void pci_vdev_write_cfg(struct pci_vdev *vdev, uint32_t offset, uint32_t bytes, uint32_t val)
|
|
{
|
|
switch (bytes) {
|
|
case 1U:
|
|
pci_vdev_write_cfg_u8(vdev, offset, (uint8_t)val);
|
|
break;
|
|
case 2U:
|
|
pci_vdev_write_cfg_u16(vdev, offset, (uint16_t)val);
|
|
break;
|
|
default:
|
|
pci_vdev_write_cfg_u32(vdev, offset, val);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @pre vpci != NULL
|
|
* @pre vpci->pci_vdev_cnt <= CONFIG_MAX_PCI_DEV_NUM
|
|
*/
|
|
struct pci_vdev *pci_find_vdev(struct acrn_vpci *vpci, union pci_bdf vbdf)
|
|
{
|
|
struct pci_vdev *vdev, *tmp;
|
|
uint32_t i;
|
|
|
|
vdev = NULL;
|
|
for (i = 0U; i < vpci->pci_vdev_cnt; i++) {
|
|
tmp = &(vpci->pci_vdevs[i]);
|
|
|
|
if (bdf_is_equal(tmp->bdf, vbdf)) {
|
|
vdev = tmp;
|
|
break;
|
|
}
|
|
}
|
|
|
|
return vdev;
|
|
}
|
|
|
|
uint32_t pci_vdev_read_bar(const struct pci_vdev *vdev, uint32_t idx)
|
|
{
|
|
uint32_t bar, offset;
|
|
|
|
offset = pci_bar_offset(idx);
|
|
bar = pci_vdev_read_cfg_u32(vdev, offset);
|
|
/* Sizing BAR */
|
|
if (bar == ~0U) {
|
|
bar = vdev->vbars[idx].mask | vdev->vbars[idx].fixed;
|
|
}
|
|
return bar;
|
|
}
|
|
|
|
static void pci_vdev_update_bar_base(struct pci_vdev *vdev, uint32_t idx)
|
|
{
|
|
struct pci_vbar *vbar;
|
|
enum pci_bar_type type;
|
|
uint64_t base = 0UL;
|
|
uint32_t lo, hi, offset;
|
|
struct acrn_vm *vm = vdev->vpci->vm;
|
|
|
|
vbar = &vdev->vbars[idx];
|
|
offset = pci_bar_offset(idx);
|
|
lo = pci_vdev_read_cfg_u32(vdev, offset);
|
|
if ((vbar->type != PCIBAR_NONE) && (lo != ~0U)) {
|
|
type = vbar->type;
|
|
base = lo & vbar->mask;
|
|
|
|
if (vbar->type == PCIBAR_MEM64) {
|
|
vbar = &vdev->vbars[idx + 1U];
|
|
hi = pci_vdev_read_cfg_u32(vdev, offset + 4U);
|
|
if (hi != ~0U) {
|
|
hi &= vbar->mask;
|
|
base |= ((uint64_t)hi << 32U);
|
|
} else {
|
|
base = 0UL;
|
|
}
|
|
}
|
|
if (type == PCIBAR_IO_SPACE) {
|
|
base &= 0xffffUL;
|
|
}
|
|
}
|
|
|
|
if ((base != 0UL) && !ept_is_mr_valid(vm, base, vdev->vbars[idx].size)) {
|
|
pr_fatal("%s, %x:%x.%x set invalid bar[%d] base: 0x%lx, size: 0x%lx\n", __func__,
|
|
vdev->bdf.bits.b, vdev->bdf.bits.d, vdev->bdf.bits.f, idx, base, vdev->vbars[idx].size);
|
|
/* If guest set a invalid GPA, ignore it temporarily */
|
|
base = 0UL;
|
|
}
|
|
|
|
vdev->vbars[idx].base = base;
|
|
}
|
|
|
|
void pci_vdev_write_bar(struct pci_vdev *vdev, uint32_t idx, uint32_t val)
|
|
{
|
|
struct pci_vbar *vbar;
|
|
uint32_t bar, offset;
|
|
uint32_t update_idx = idx;
|
|
|
|
vbar = &vdev->vbars[idx];
|
|
bar = val & vbar->mask;
|
|
bar |= vbar->fixed;
|
|
offset = pci_bar_offset(idx);
|
|
pci_vdev_write_cfg_u32(vdev, offset, bar);
|
|
|
|
if (vbar->type == PCIBAR_MEM64HI) {
|
|
update_idx -= 1U;
|
|
}
|
|
|
|
pci_vdev_update_bar_base(vdev, update_idx);
|
|
}
|