acrn-hypervisor/hypervisor/arch/x86/guest/optee.c
Geoffroy Van Cutsem 8b16be9185 Remove "All rights reserved" string headers
Many of the license and Intel copyright headers include the "All rights
reserved" string. It is not relevant in the context of the BSD-3-Clause
license that the code is released under. This patch removes those strings
throughout the code (hypervisor, devicemodel and misc).

Tracked-On: #7254
Signed-off-by: Geoffroy Van Cutsem <geoffroy.vancutsem@intel.com>
2022-04-06 13:21:02 +08:00

220 lines
6.4 KiB
C

/*
* Copyright (C) 2021 Intel Corporation.
*
* SPDX-License-Identifier: BSD-3-Clause
*/
#include <asm/guest/vm.h>
#include <asm/guest/ept.h>
#include <asm/vm_config.h>
#include <asm/mmu.h>
#include <asm/guest/optee.h>
#include <asm/trampoline.h>
#include <asm/guest/vlapic.h>
#include <asm/guest/virq.h>
#include <asm/lapic.h>
#include <reloc.h>
#include <hypercall.h>
#include <logmsg.h>
int is_tee_vm(struct acrn_vm *vm)
{
return (get_vm_config(vm->vm_id)->guest_flags & GUEST_FLAG_TEE) != 0;
}
int is_ree_vm(struct acrn_vm *vm)
{
return (get_vm_config(vm->vm_id)->guest_flags & GUEST_FLAG_REE) != 0;
}
void prepare_tee_vm_memmap(struct acrn_vm *vm, const struct acrn_vm_config *vm_config)
{
uint64_t hv_hpa;
/*
* Only need to map following things, let init_vpci to map the secure devices
* if any.
*
* 1. go through physical e820 table, to ept add all system memory entries.
* 2. remove hv owned memory.
*/
if ((vm_config->guest_flags & GUEST_FLAG_TEE) != 0U) {
vm->e820_entry_num = get_e820_entries_count();
vm->e820_entries = (struct e820_entry *)get_e820_entry();
prepare_vm_identical_memmap(vm, E820_TYPE_RAM, EPT_WB | EPT_RWX);
hv_hpa = hva2hpa((void *)(get_hv_image_base()));
ept_del_mr(vm, (uint64_t *)vm->arch_vm.nworld_eptp, hv_hpa, get_hv_ram_size());
}
}
static struct acrn_vm *get_companion_vm(struct acrn_vm *vm)
{
return get_vm_from_vmid(get_vm_config(vm->vm_id)->companion_vm_id);
}
static int32_t tee_switch_to_ree(struct acrn_vcpu *vcpu)
{
uint64_t rdi, rsi, rdx, rbx;
struct acrn_vm *ree_vm;
struct acrn_vcpu *ree_vcpu;
uint32_t pending_intr;
int32_t ret = -EINVAL;
rdi = vcpu_get_gpreg(vcpu, CPU_REG_RDI);
rsi = vcpu_get_gpreg(vcpu, CPU_REG_RSI);
rdx = vcpu_get_gpreg(vcpu, CPU_REG_RDX);
rbx = vcpu_get_gpreg(vcpu, CPU_REG_RBX);
ree_vm = get_companion_vm(vcpu->vm);
ree_vcpu = vcpu_from_pid(ree_vm, get_pcpu_id());
if (ree_vcpu != NULL) {
/*
* We should avoid copy any values to REE registers,
* If this is a FIQ return.
*/
if (rdi != OPTEE_RETURN_FIQ_DONE) {
vcpu_set_gpreg(ree_vcpu, CPU_REG_RDI, rdi);
vcpu_set_gpreg(ree_vcpu, CPU_REG_RSI, rsi);
vcpu_set_gpreg(ree_vcpu, CPU_REG_RDX, rdx);
vcpu_set_gpreg(ree_vcpu, CPU_REG_RBX, rbx);
}
pending_intr = vlapic_get_next_pending_intr(vcpu);
if (prio(pending_intr) > prio(TEE_FIXED_NONSECURE_VECTOR)) {
/* For TEE, all non-secure interrupts are represented as
* TEE_FIXED_NONSECURE_VECTOR that has lower priority than all
* secure interrupts.
*
* If there are secure interrupts pending, we inject TEE's PI
* ANV and schedules REE. This way REE gets trapped immediately
* after VM Entry and will go through the secure interrupt handling
* flow in handle_x86_tee_int.
*/
send_single_ipi(pcpuid_from_vcpu(ree_vcpu),
(uint32_t)(vcpu->arch.pid.control.bits.nv));
} else if (prio(pending_intr) == prio(TEE_FIXED_NONSECURE_VECTOR)) {
/* The TEE_FIXED_NONSECURE_VECTOR needs to be cleared as the
* pending non-secure interrupts will be handled immediately
* after resuming to REE. On ARM this is automatically done
* by hardware and ACRN emulates this behavior.
*/
vlapic_clear_pending_intr(vcpu, TEE_FIXED_NONSECURE_VECTOR);
}
sleep_thread(&vcpu->thread_obj);
ret = 0;
} else {
pr_fatal("No REE vCPU running on this pCPU%u, \n", get_pcpu_id());
}
return ret;
}
static int32_t ree_switch_to_tee(struct acrn_vcpu *vcpu)
{
uint64_t rax, rdi, rsi, rdx, rbx, rcx;
struct acrn_vm *tee_vm;
struct acrn_vcpu *tee_vcpu;
int32_t ret = -EINVAL;
rax = vcpu_get_gpreg(vcpu, CPU_REG_RAX);
rdi = vcpu_get_gpreg(vcpu, CPU_REG_RDI);
rsi = vcpu_get_gpreg(vcpu, CPU_REG_RSI);
rdx = vcpu_get_gpreg(vcpu, CPU_REG_RDX);
rbx = vcpu_get_gpreg(vcpu, CPU_REG_RBX);
rcx = vcpu_get_gpreg(vcpu, CPU_REG_RCX);
tee_vm = get_companion_vm(vcpu->vm);
tee_vcpu = vcpu_from_pid(tee_vm, get_pcpu_id());
if (tee_vcpu != NULL) {
vcpu_set_gpreg(tee_vcpu, CPU_REG_RAX, rax);
vcpu_set_gpreg(tee_vcpu, CPU_REG_RDI, rdi);
vcpu_set_gpreg(tee_vcpu, CPU_REG_RSI, rsi);
vcpu_set_gpreg(tee_vcpu, CPU_REG_RDX, rdx);
vcpu_set_gpreg(tee_vcpu, CPU_REG_RBX, rbx);
vcpu_set_gpreg(tee_vcpu, CPU_REG_RCX, rcx);
wake_thread(&tee_vcpu->thread_obj);
ret = 0;
} else {
pr_fatal("No TEE vCPU running on this pCPU%u, \n", get_pcpu_id());
}
return ret;
}
int32_t hcall_handle_tee_vcpu_boot_done(struct acrn_vcpu *vcpu, __unused struct acrn_vm *target_vm,
__unused uint64_t param1, __unused uint64_t param2)
{
struct acrn_vm *ree_vm;
uint64_t rdi;
/*
* The (RDI == 1) indicates to start REE VM, otherwise only need
* to sleep the corresponding TEE vCPU.
*/
rdi = vcpu_get_gpreg(vcpu, CPU_REG_RDI);
if (rdi == 1UL) {
ree_vm = get_companion_vm(vcpu->vm);
start_vm(ree_vm);
}
sleep_thread(&vcpu->thread_obj);
return 0;
}
int32_t hcall_switch_ee(struct acrn_vcpu *vcpu, __unused struct acrn_vm *target_vm,
__unused uint64_t param1, __unused uint64_t param2)
{
int32_t ret = 0;
if (is_tee_vm(vcpu->vm)) {
ret = tee_switch_to_ree(vcpu);
} else if (is_ree_vm(vcpu->vm)) {
ret = ree_switch_to_tee(vcpu);
}
return ret;
}
void handle_x86_tee_int(struct ptirq_remapping_info *entry, uint16_t pcpu_id)
{
struct acrn_vcpu *tee_vcpu;
struct acrn_vcpu *curr_vcpu = get_running_vcpu(pcpu_id);
if (is_ree_vm(entry->vm) && is_tee_vm(curr_vcpu->vm)) {
/*
* Non-Secure interrupt (interrupt belongs to REE) comes
* when REE vcpu is running, the interrupt will be injected
* to REE directly. But when TEE vcpu is running at that time,
* we need to inject a predefined vector to TEE for notification
* and continue to switch back to TEE for running.
*/
tee_vcpu = vcpu_from_pid(get_companion_vm(entry->vm), pcpu_id);
vlapic_set_intr(tee_vcpu, TEE_FIXED_NONSECURE_VECTOR, LAPIC_TRIG_EDGE);
} else if (is_tee_vm(entry->vm) && is_ree_vm(curr_vcpu->vm)) {
/*
* Secure interrupt (interrupt belongs to TEE) comes
* when TEE vcpu is running, the interrupt will be
* injected to TEE directly. But when REE vcpu is running
* at that time, we need to switch to TEE for handling,
* and copy 0xB20000FF to RDI to notify OPTEE about this.
*/
tee_vcpu = vcpu_from_pid(entry->vm, pcpu_id);
/*
* Copy 0xB20000FF to RDI to indicate the switch is from secure interrupt
* This is the contract with OPTEE.
*/
vcpu_set_gpreg(tee_vcpu, CPU_REG_RDI, OPTEE_FIQ_ENTRY);
wake_thread(&tee_vcpu->thread_obj);
} else {
/* Nothing need to do for this moment */
}
}