acrn-hypervisor/hypervisor/common/schedule.c
Wu Zhou 064be1e3e6 hv: support halt in hv idle
When all vCPU threads on one pCPU are put to sleep (e.g., when all
guests execute HLT), hv would schedule to idle thread. Currently the
idle thread executes PAUSE which does not enter any c-state and consumes
a lot of power. This patch is to support HLT in the idle thread.

When we switch to HLT, we have to make sure events that would wake a
vCPU must also be able to wake the pCPU. Those events are either
generated by local interrupt or issued by other pCPUs followed by an
ipi kick.

Each of them have an interrupt involved, so they are also able to wake
the halted pCPU. Except when the pCPU has just scheduled to idle thread
but not yet halted, interrupts could be missed.

sleep-------schedule to idle------IRQ ON---HLT--(kick missed)
                                         ^
                              wake---kick|

This areas should be protected. This is done by a safe halt
mechanism leveraging STI instruction’s delay effect (same as Linux).

vCPUs with lapic_pt or hv with CONFIG_KEEP_IRQ_DISABLED=y does not allow
interrupts in root mode, so they could never wake from HLT (INIT kick
does not wake HLT in root mode either). They should continue using PAUSE
in idle.

Tracked-On: #8507
Signed-off-by: Wu Zhou <wu.zhou@intel.com>
Reviewed-by: Junjie Mao <junjie.mao@intel.com>
2023-09-15 11:52:40 +08:00

254 lines
5.7 KiB
C

/*
* Copyright (C) 2018-2022 Intel Corporation.
*
* SPDX-License-Identifier: BSD-3-Clause
*/
#include <rtl.h>
#include <list.h>
#include <asm/lib/bits.h>
#include <asm/cpu.h>
#include <asm/per_cpu.h>
#include <asm/lapic.h>
#include <schedule.h>
#include <sprintf.h>
#include <asm/irq.h>
bool is_idle_thread(const struct thread_object *obj)
{
uint16_t pcpu_id = obj->pcpu_id;
return (obj == &per_cpu(idle, pcpu_id));
}
static inline bool is_blocked(const struct thread_object *obj)
{
return obj->status == THREAD_STS_BLOCKED;
}
static inline bool is_running(const struct thread_object *obj)
{
return obj->status == THREAD_STS_RUNNING;
}
static inline void set_thread_status(struct thread_object *obj, enum thread_object_state status)
{
obj->status = status;
}
void obtain_schedule_lock(uint16_t pcpu_id, uint64_t *rflag)
{
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
spinlock_irqsave_obtain(&ctl->scheduler_lock, rflag);
}
void release_schedule_lock(uint16_t pcpu_id, uint64_t rflag)
{
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
spinlock_irqrestore_release(&ctl->scheduler_lock, rflag);
}
static struct acrn_scheduler *get_scheduler(uint16_t pcpu_id)
{
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
return ctl->scheduler;
}
/**
* @pre obj != NULL
*/
uint16_t sched_get_pcpuid(const struct thread_object *obj)
{
return obj->pcpu_id;
}
void init_sched(uint16_t pcpu_id)
{
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
per_cpu(mode_to_idle, pcpu_id) = IDLE_MODE_HLT;
per_cpu(mode_to_kick_pcpu, pcpu_id) = DEL_MODE_IPI;
spinlock_init(&ctl->scheduler_lock);
ctl->flags = 0UL;
ctl->curr_obj = NULL;
ctl->pcpu_id = pcpu_id;
#ifdef CONFIG_SCHED_NOOP
ctl->scheduler = &sched_noop;
#endif
#ifdef CONFIG_SCHED_IORR
ctl->scheduler = &sched_iorr;
#endif
#ifdef CONFIG_SCHED_BVT
ctl->scheduler = &sched_bvt;
#endif
#ifdef CONFIG_SCHED_PRIO
ctl->scheduler = &sched_prio;
#endif
if (ctl->scheduler->init != NULL) {
ctl->scheduler->init(ctl);
}
}
void deinit_sched(uint16_t pcpu_id)
{
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
if (ctl->scheduler->deinit != NULL) {
ctl->scheduler->deinit(ctl);
}
}
void init_thread_data(struct thread_object *obj)
{
struct acrn_scheduler *scheduler = get_scheduler(obj->pcpu_id);
uint64_t rflag;
obtain_schedule_lock(obj->pcpu_id, &rflag);
if (scheduler->init_data != NULL) {
scheduler->init_data(obj);
}
/* initial as BLOCKED status, so we can wake it up to run */
set_thread_status(obj, THREAD_STS_BLOCKED);
release_schedule_lock(obj->pcpu_id, rflag);
}
void deinit_thread_data(struct thread_object *obj)
{
struct acrn_scheduler *scheduler = get_scheduler(obj->pcpu_id);
if (scheduler->deinit_data != NULL) {
scheduler->deinit_data(obj);
}
}
struct thread_object *sched_get_current(uint16_t pcpu_id)
{
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
return ctl->curr_obj;
}
/**
* @pre delmode == DEL_MODE_IPI || delmode == DEL_MODE_INIT
*/
void make_reschedule_request(uint16_t pcpu_id)
{
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
bitmap_set_lock(NEED_RESCHEDULE, &ctl->flags);
if (get_pcpu_id() != pcpu_id) {
kick_pcpu(pcpu_id);
}
}
bool need_reschedule(uint16_t pcpu_id)
{
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
return bitmap_test(NEED_RESCHEDULE, &ctl->flags);
}
void schedule(void)
{
uint16_t pcpu_id = get_pcpu_id();
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
struct thread_object *next = &per_cpu(idle, pcpu_id);
struct thread_object *prev = ctl->curr_obj;
uint64_t rflag;
obtain_schedule_lock(pcpu_id, &rflag);
if (ctl->scheduler->pick_next != NULL) {
next = ctl->scheduler->pick_next(ctl);
}
bitmap_clear_lock(NEED_RESCHEDULE, &ctl->flags);
/* If we picked different sched object, switch context */
if (prev != next) {
if (prev != NULL) {
if (prev->switch_out != NULL) {
prev->switch_out(prev);
}
set_thread_status(prev, prev->be_blocking ? THREAD_STS_BLOCKED : THREAD_STS_RUNNABLE);
prev->be_blocking = false;
}
if (next->switch_in != NULL) {
next->switch_in(next);
}
set_thread_status(next, THREAD_STS_RUNNING);
ctl->curr_obj = next;
release_schedule_lock(pcpu_id, rflag);
arch_switch_to(&prev->host_sp, &next->host_sp);
} else {
release_schedule_lock(pcpu_id, rflag);
}
}
void sleep_thread(struct thread_object *obj)
{
uint16_t pcpu_id = obj->pcpu_id;
struct acrn_scheduler *scheduler = get_scheduler(pcpu_id);
uint64_t rflag;
obtain_schedule_lock(pcpu_id, &rflag);
if (scheduler->sleep != NULL) {
scheduler->sleep(obj);
}
if (is_running(obj)) {
make_reschedule_request(pcpu_id);
obj->be_blocking = true;
} else {
set_thread_status(obj, THREAD_STS_BLOCKED);
}
release_schedule_lock(pcpu_id, rflag);
}
void sleep_thread_sync(struct thread_object *obj)
{
sleep_thread(obj);
while (!is_blocked(obj)) {
asm_pause();
}
}
void wake_thread(struct thread_object *obj)
{
uint16_t pcpu_id = obj->pcpu_id;
struct acrn_scheduler *scheduler;
uint64_t rflag;
obtain_schedule_lock(pcpu_id, &rflag);
if (is_blocked(obj) || obj->be_blocking) {
scheduler = get_scheduler(pcpu_id);
if (scheduler->wake != NULL) {
scheduler->wake(obj);
}
if (is_blocked(obj)) {
set_thread_status(obj, THREAD_STS_RUNNABLE);
make_reschedule_request(pcpu_id);
}
obj->be_blocking = false;
}
release_schedule_lock(pcpu_id, rflag);
}
void yield_current(void)
{
make_reschedule_request(get_pcpu_id());
}
void run_thread(struct thread_object *obj)
{
uint64_t rflag;
init_thread_data(obj);
obtain_schedule_lock(obj->pcpu_id, &rflag);
get_cpu_var(sched_ctl).curr_obj = obj;
set_thread_status(obj, THREAD_STS_RUNNING);
release_schedule_lock(obj->pcpu_id, rflag);
if (obj->thread_entry != NULL) {
obj->thread_entry(obj);
}
}