mirror of
https://github.com/projectacrn/acrn-hypervisor.git
synced 2025-05-12 10:24:21 +00:00
take use of VCPU_NOTIFY vector, add smp_call_function support. added a per_cpu field smp_call_info, and make each smp_call_function is not re-entered, and the caller CPU is returned when all the target CPUs complete the call. v4: - remove global lock - take use of wait_sync_change function to do the sequence sync v3: - remove per_cpu lock in smp_call_info - use a global lock to ensure smp_call_function sequence - use pcpu_sync_sleep to wait IPI complete v2: - after new smp function come, if old one exist, changed from overwirte with the new one to ignore the new one. Signed-off-by: Jason Chen CJ <jason.cj.chen@intel.com>
109 lines
2.6 KiB
C
109 lines
2.6 KiB
C
/*
|
|
* Copyright (C) 2018 Intel Corporation. All rights reserved.
|
|
*
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
*/
|
|
|
|
#include <hypervisor.h>
|
|
|
|
static uint32_t notification_irq = IRQ_INVALID;
|
|
|
|
static volatile uint64_t smp_call_mask = 0UL;
|
|
|
|
/* run in interrupt context */
|
|
static int kick_notification(__unused uint32_t irq, __unused void *data)
|
|
{
|
|
/* Notification vector is used to kick taget cpu out of non-root mode.
|
|
* And it also serves for smp call.
|
|
*/
|
|
uint16_t pcpu_id = get_cpu_id();
|
|
|
|
if (bitmap_test(pcpu_id, &smp_call_mask)) {
|
|
struct smp_call_info_data *smp_call =
|
|
&per_cpu(smp_call_info, pcpu_id);
|
|
|
|
if (smp_call->func)
|
|
smp_call->func(smp_call->data);
|
|
bitmap_clear_nolock(pcpu_id, &smp_call_mask);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void smp_call_function(uint64_t mask, smp_call_func_t func, void *data)
|
|
{
|
|
uint16_t pcpu_id;
|
|
struct smp_call_info_data *smp_call;
|
|
|
|
/* wait for previous smp call complete, which may run on other cpus */
|
|
while (atomic_cmpxchg64(&smp_call_mask, 0UL, mask & INVALID_BIT_INDEX));
|
|
while ((pcpu_id = ffs64(mask)) != INVALID_BIT_INDEX) {
|
|
bitmap_clear_nolock(pcpu_id, &mask);
|
|
if (bitmap_test(pcpu_id, &pcpu_active_bitmap)) {
|
|
smp_call = &per_cpu(smp_call_info, pcpu_id);
|
|
smp_call->func = func;
|
|
smp_call->data = data;
|
|
} else {
|
|
/* pcpu is not in active, print error */
|
|
pr_err("pcpu_id %d not in active!", pcpu_id);
|
|
bitmap_clear_nolock(pcpu_id, &smp_call_mask);
|
|
}
|
|
}
|
|
send_dest_ipi(smp_call_mask, VECTOR_NOTIFY_VCPU,
|
|
INTR_LAPIC_ICR_LOGICAL);
|
|
/* wait for current smp call complete */
|
|
wait_sync_change(&smp_call_mask, 0UL);
|
|
}
|
|
|
|
static int request_notification_irq(irq_action_t func, void *data,
|
|
const char *name)
|
|
{
|
|
int32_t retval;
|
|
|
|
if (notification_irq != IRQ_INVALID) {
|
|
pr_info("%s, Notification vector already allocated on this CPU",
|
|
__func__);
|
|
return -EBUSY;
|
|
}
|
|
|
|
/* all cpu register the same notification vector */
|
|
retval = request_irq(NOTIFY_IRQ, func, data, name);
|
|
if (retval < 0) {
|
|
pr_err("Failed to add notify isr");
|
|
return -ENODEV;
|
|
}
|
|
|
|
notification_irq = (uint32_t)retval;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void setup_notification(void)
|
|
{
|
|
uint16_t cpu;
|
|
char name[32] = {0};
|
|
|
|
cpu = get_cpu_id();
|
|
if (cpu > 0U) {
|
|
return;
|
|
}
|
|
|
|
/* support IPI notification, VM0 will register all CPU */
|
|
snprintf(name, 32, "NOTIFY_ISR%d", cpu);
|
|
if (request_notification_irq(kick_notification, NULL, name) < 0) {
|
|
pr_err("Failed to setup notification");
|
|
return;
|
|
}
|
|
|
|
dev_dbg(ACRN_DBG_PTIRQ, "NOTIFY: irq[%d] setup vector %x",
|
|
notification_irq, irq_to_vector(notification_irq));
|
|
}
|
|
|
|
static void cleanup_notification(void)
|
|
{
|
|
if (notification_irq != IRQ_INVALID) {
|
|
free_irq(notification_irq);
|
|
}
|
|
notification_irq = IRQ_INVALID;
|
|
}
|