Message ID | 20170206153113.27729-21-alex.bennee@linaro.org |
---|---|
State | New |
Headers | show |
Series | MTTCG Base enabling patches with ARM enablement | expand |
On 6 February 2017 at 15:31, Alex Bennée <alex.bennee@linaro.org> wrote: > When switching a new vCPU on we want to complete a bunch of the setup > work before we start scheduling the vCPU thread. To do this cleanly we > defer vCPU setup to async work which will run the vCPUs execution > context as the thread is woken up. The scheduling of the work will kick > the vCPU awake. > > This avoids potential races in MTTCG system emulation. > > Signed-off-by: Alex Bennée <alex.bennee@linaro.org> > Reviewed-by: Richard Henderson <rth@twiddle.net> > @@ -77,7 +159,7 @@ int arm_set_cpu_on(uint64_t cpuid, uint64_t entry, uint64_t context_id, > } > > target_cpu = ARM_CPU(target_cpu_state); > - if (!target_cpu->powered_off) { > + if (!atomic_mb_read(&target_cpu->powered_off)) { > qemu_log_mask(LOG_GUEST_ERROR, > "[ARM]%s: CPU %" PRId64 " is already on\n", > __func__, cpuid); > + /* > + * If another CPU has powered the target on we are in the state > + * ON_PENDING and additional attempts to power on the CPU should > + * fail (see 6.6 Implementation CPU_ON/CPU_OFF races in the PSCI > + * spec) > + */ > + if (atomic_cmpxchg(&target_cpu->powering_on, false, true)) { > + qemu_log_mask(LOG_GUEST_ERROR, > + "[ARM]%s: CPU %" PRId64 " is already powering on\n", > + __func__, cpuid); > + return QEMU_ARM_POWERCTL_ON_PENDING; > } I find this too hard to reason about :-( We store the current CPU state in not one but two CPU structure fields, and then we check and manipulate it not by doing "take lock; access and update; drop lock" but by doing successive atomic operations on the various different flags. This seems unnecessarily tricky, and the patch doesn't provide any documentation about what's going on and what the constraints are to avoid races. > > - /* Start the new CPU at the requested address */ > - cpu_set_pc(target_cpu_state, entry); > + /* To avoid racing with a CPU we are just kicking off we do the > + * final bit of preparation for the work in the target CPUs > + * context. > + */ > + info = g_new(struct cpu_on_info, 1); > + info->entry = entry; > + info->context_id = context_id; > + info->target_el = target_el; > + info->target_aa64 = target_aa64; > > - qemu_cpu_kick(target_cpu_state); > + async_run_on_cpu(target_cpu_state, arm_set_cpu_on_async_work, > + RUN_ON_CPU_HOST_PTR(info)); > > /* We are good to go */ > return QEMU_ARM_POWERCTL_RET_SUCCESS; > } > @@ -221,8 +284,9 @@ int arm_reset_cpu(uint64_t cpuid) > return QEMU_ARM_POWERCTL_IS_OFF; > } > > - /* Reset the cpu */ > - cpu_reset(target_cpu_state); > + /* Queue work to run under the target vCPUs context */ > + async_run_on_cpu(target_cpu_state, arm_reset_cpu_async_work, > + RUN_ON_CPU_NULL); The imx6 datasheet says that the RST bit in the SCR register is set by the guest to trigger a reset, and then the bit remains set until the reset has finished (at which point it self-clears). At the moment the imx6_src.c code does this: if (EXTRACT(change_mask, CORE0_RST)) { arm_reset_cpu(0); clear_bit(CORE0_RST_SHIFT, ¤t_value); } ie it assumes that arm_reset_cpu() is synchronous. PS: if the code does an arm_set_cpu_on() and then arm_reset_cpu() do we do the two lots of async work in the right order? > return QEMU_ARM_POWERCTL_RET_SUCCESS; > } > diff --git a/target/arm/arm-powerctl.h b/target/arm/arm-powerctl.h > index 98ee04989b..04353923c0 100644 > --- a/target/arm/arm-powerctl.h > +++ b/target/arm/arm-powerctl.h > @@ -17,6 +17,7 @@ > #define QEMU_ARM_POWERCTL_INVALID_PARAM QEMU_PSCI_RET_INVALID_PARAMS > #define QEMU_ARM_POWERCTL_ALREADY_ON QEMU_PSCI_RET_ALREADY_ON > #define QEMU_ARM_POWERCTL_IS_OFF QEMU_PSCI_RET_DENIED > +#define QEMU_ARM_POWERCTL_ON_PENDING QEMU_PSCI_RET_ON_PENDING > > /* > * arm_get_cpu_by_id: > @@ -43,6 +44,7 @@ CPUState *arm_get_cpu_by_id(uint64_t cpuid); > * Returns: QEMU_ARM_POWERCTL_RET_SUCCESS on success. > * QEMU_ARM_POWERCTL_INVALID_PARAM if bad parameters are provided. > * QEMU_ARM_POWERCTL_ALREADY_ON if the CPU was already started. > + * QEMU_ARM_POWERCTL_ON_PENDING if the CPU is still powering up > */ > int arm_set_cpu_on(uint64_t cpuid, uint64_t entry, uint64_t context_id, > uint32_t target_el, bool target_aa64); > diff --git a/target/arm/cpu.h b/target/arm/cpu.h > index 39bff86daf..b8f82d5d20 100644 > --- a/target/arm/cpu.h > +++ b/target/arm/cpu.h > @@ -582,8 +582,16 @@ struct ARMCPU { > > /* Should CPU start in PSCI powered-off state? */ > bool start_powered_off; > - /* CPU currently in PSCI powered-off state */ > + /* CPU PSCI state. > + * > + * For TCG these can be cross-vCPU accesses and should be done > + * atomically to avoid races. > + * > + * - powered_off indicates the vCPU state > + * - powering_on true if the vCPU has had a CPU_ON but not yet up > + */ > bool powered_off; > + bool powering_on; /* PSCI ON_PENDING */ > /* CPU has virtualization extension */ > bool has_el2; > /* CPU has security extension */ > -- > 2.11.0 > thanks -- PMM
Peter Maydell <peter.maydell@linaro.org> writes: > On 6 February 2017 at 15:31, Alex Bennée <alex.bennee@linaro.org> wrote: >> When switching a new vCPU on we want to complete a bunch of the setup >> work before we start scheduling the vCPU thread. To do this cleanly we >> defer vCPU setup to async work which will run the vCPUs execution >> context as the thread is woken up. The scheduling of the work will kick >> the vCPU awake. >> >> This avoids potential races in MTTCG system emulation. >> >> Signed-off-by: Alex Bennée <alex.bennee@linaro.org> >> Reviewed-by: Richard Henderson <rth@twiddle.net> > >> @@ -77,7 +159,7 @@ int arm_set_cpu_on(uint64_t cpuid, uint64_t entry, uint64_t context_id, >> } >> >> target_cpu = ARM_CPU(target_cpu_state); >> - if (!target_cpu->powered_off) { >> + if (!atomic_mb_read(&target_cpu->powered_off)) { >> qemu_log_mask(LOG_GUEST_ERROR, >> "[ARM]%s: CPU %" PRId64 " is already on\n", >> __func__, cpuid); > >> + /* >> + * If another CPU has powered the target on we are in the state >> + * ON_PENDING and additional attempts to power on the CPU should >> + * fail (see 6.6 Implementation CPU_ON/CPU_OFF races in the PSCI >> + * spec) >> + */ >> + if (atomic_cmpxchg(&target_cpu->powering_on, false, true)) { >> + qemu_log_mask(LOG_GUEST_ERROR, >> + "[ARM]%s: CPU %" PRId64 " is already powering on\n", >> + __func__, cpuid); >> + return QEMU_ARM_POWERCTL_ON_PENDING; >> } > > I find this too hard to reason about :-( We store the current > CPU state in not one but two CPU structure fields, and then > we check and manipulate it not by doing "take lock; access > and update; drop lock" but by doing successive atomic > operations on the various different flags. This seems unnecessarily > tricky, and the patch doesn't provide any documentation about > what's going on and what the constraints are to avoid races. I was trying to avoid changing the powered_off state to a new variable (maybe pcsi_power_state?) because it meant I didn't have to touch the kvm code that treated powered_off as a simple bool. However I can unify the state into one variable if you wish. As for locking do we have a handy per-CPU lock already or would it be enough to use the BQL for this? Although that may be problematic as the cpu_reset/on can come from two places (the initial realization and the runtime switching of power state.) > >> >> - /* Start the new CPU at the requested address */ >> - cpu_set_pc(target_cpu_state, entry); >> + /* To avoid racing with a CPU we are just kicking off we do the >> + * final bit of preparation for the work in the target CPUs >> + * context. >> + */ >> + info = g_new(struct cpu_on_info, 1); >> + info->entry = entry; >> + info->context_id = context_id; >> + info->target_el = target_el; >> + info->target_aa64 = target_aa64; >> >> - qemu_cpu_kick(target_cpu_state); >> + async_run_on_cpu(target_cpu_state, arm_set_cpu_on_async_work, >> + RUN_ON_CPU_HOST_PTR(info)); >> >> /* We are good to go */ >> return QEMU_ARM_POWERCTL_RET_SUCCESS; >> } > >> @@ -221,8 +284,9 @@ int arm_reset_cpu(uint64_t cpuid) >> return QEMU_ARM_POWERCTL_IS_OFF; >> } >> >> - /* Reset the cpu */ >> - cpu_reset(target_cpu_state); >> + /* Queue work to run under the target vCPUs context */ >> + async_run_on_cpu(target_cpu_state, arm_reset_cpu_async_work, >> + RUN_ON_CPU_NULL); > > The imx6 datasheet says that the RST bit in the SCR register > is set by the guest to trigger a reset, and then the bit > remains set until the reset has finished (at which point it > self-clears). At the moment the imx6_src.c code does this: > > if (EXTRACT(change_mask, CORE0_RST)) { > arm_reset_cpu(0); > clear_bit(CORE0_RST_SHIFT, ¤t_value); > } > > ie it assumes that arm_reset_cpu() is synchronous. Well for this case the imx6 could queue the "clear_bit" function as async work and then it will be correctly reported after the reset has occurred. > > PS: if the code does an arm_set_cpu_on() and then > arm_reset_cpu() do we do the two lots of async work > in the right order? Yes. The actual run queue is protected by a lock and run on a FIFO basis. > >> return QEMU_ARM_POWERCTL_RET_SUCCESS; >> } >> diff --git a/target/arm/arm-powerctl.h b/target/arm/arm-powerctl.h >> index 98ee04989b..04353923c0 100644 >> --- a/target/arm/arm-powerctl.h >> +++ b/target/arm/arm-powerctl.h >> @@ -17,6 +17,7 @@ >> #define QEMU_ARM_POWERCTL_INVALID_PARAM QEMU_PSCI_RET_INVALID_PARAMS >> #define QEMU_ARM_POWERCTL_ALREADY_ON QEMU_PSCI_RET_ALREADY_ON >> #define QEMU_ARM_POWERCTL_IS_OFF QEMU_PSCI_RET_DENIED >> +#define QEMU_ARM_POWERCTL_ON_PENDING QEMU_PSCI_RET_ON_PENDING >> >> /* >> * arm_get_cpu_by_id: >> @@ -43,6 +44,7 @@ CPUState *arm_get_cpu_by_id(uint64_t cpuid); >> * Returns: QEMU_ARM_POWERCTL_RET_SUCCESS on success. >> * QEMU_ARM_POWERCTL_INVALID_PARAM if bad parameters are provided. >> * QEMU_ARM_POWERCTL_ALREADY_ON if the CPU was already started. >> + * QEMU_ARM_POWERCTL_ON_PENDING if the CPU is still powering up >> */ >> int arm_set_cpu_on(uint64_t cpuid, uint64_t entry, uint64_t context_id, >> uint32_t target_el, bool target_aa64); >> diff --git a/target/arm/cpu.h b/target/arm/cpu.h >> index 39bff86daf..b8f82d5d20 100644 >> --- a/target/arm/cpu.h >> +++ b/target/arm/cpu.h >> @@ -582,8 +582,16 @@ struct ARMCPU { >> >> /* Should CPU start in PSCI powered-off state? */ >> bool start_powered_off; >> - /* CPU currently in PSCI powered-off state */ >> + /* CPU PSCI state. >> + * >> + * For TCG these can be cross-vCPU accesses and should be done >> + * atomically to avoid races. >> + * >> + * - powered_off indicates the vCPU state >> + * - powering_on true if the vCPU has had a CPU_ON but not yet up >> + */ >> bool powered_off; >> + bool powering_on; /* PSCI ON_PENDING */ >> /* CPU has virtualization extension */ >> bool has_el2; >> /* CPU has security extension */ >> -- >> 2.11.0 >> > > thanks > -- PMM -- Alex Bennée
On 7 February 2017 at 16:52, Alex Bennée <alex.bennee@linaro.org> wrote: > > Peter Maydell <peter.maydell@linaro.org> writes: >> I find this too hard to reason about :-( We store the current >> CPU state in not one but two CPU structure fields, and then >> we check and manipulate it not by doing "take lock; access >> and update; drop lock" but by doing successive atomic >> operations on the various different flags. This seems unnecessarily >> tricky, and the patch doesn't provide any documentation about >> what's going on and what the constraints are to avoid races. > > I was trying to avoid changing the powered_off state to a new variable > (maybe pcsi_power_state?) because it meant I didn't have to touch the > kvm code that treated powered_off as a simple bool. However I can unify > the state into one variable if you wish. Changing it also makes migration compatibility awkward. (Speaking of which, your new flag needs a migration subsection and appropriate needed function to only transfer it if it's set.) > As for locking do we have a handy per-CPU lock already or would it be > enough to use the BQL for this? Although that may be problematic as the > cpu_reset/on can come from two places (the initial realization and the > runtime switching of power state.) Nobody's going to put powering CPUs up and down in their critical path, so I think the BQL would be fine. >>> - /* Start the new CPU at the requested address */ >>> - cpu_set_pc(target_cpu_state, entry); >>> + /* To avoid racing with a CPU we are just kicking off we do the >>> + * final bit of preparation for the work in the target CPUs >>> + * context. >>> + */ >>> + info = g_new(struct cpu_on_info, 1); >>> + info->entry = entry; >>> + info->context_id = context_id; >>> + info->target_el = target_el; >>> + info->target_aa64 = target_aa64; >>> >>> - qemu_cpu_kick(target_cpu_state); >>> + async_run_on_cpu(target_cpu_state, arm_set_cpu_on_async_work, >>> + RUN_ON_CPU_HOST_PTR(info)); >>> >>> /* We are good to go */ >>> return QEMU_ARM_POWERCTL_RET_SUCCESS; >>> } >> >>> @@ -221,8 +284,9 @@ int arm_reset_cpu(uint64_t cpuid) >>> return QEMU_ARM_POWERCTL_IS_OFF; >>> } >>> >>> - /* Reset the cpu */ >>> - cpu_reset(target_cpu_state); >>> + /* Queue work to run under the target vCPUs context */ >>> + async_run_on_cpu(target_cpu_state, arm_reset_cpu_async_work, >>> + RUN_ON_CPU_NULL); >> >> The imx6 datasheet says that the RST bit in the SCR register >> is set by the guest to trigger a reset, and then the bit >> remains set until the reset has finished (at which point it >> self-clears). At the moment the imx6_src.c code does this: >> >> if (EXTRACT(change_mask, CORE0_RST)) { >> arm_reset_cpu(0); >> clear_bit(CORE0_RST_SHIFT, ¤t_value); >> } >> >> ie it assumes that arm_reset_cpu() is synchronous. > > Well for this case the imx6 could queue the "clear_bit" function as > async work and then it will be correctly reported after the reset has > occurred. It could, but as of this patchset it doesn't... thanks -- PMM
diff --git a/target/arm/arm-powerctl.c b/target/arm/arm-powerctl.c index fbb7a15daa..bf14740852 100644 --- a/target/arm/arm-powerctl.c +++ b/target/arm/arm-powerctl.c @@ -48,11 +48,93 @@ CPUState *arm_get_cpu_by_id(uint64_t id) return NULL; } +struct cpu_on_info { + uint64_t entry; + uint64_t context_id; + uint32_t target_el; + bool target_aa64; +}; + + +static void arm_set_cpu_on_async_work(CPUState *target_cpu_state, + run_on_cpu_data data) +{ + ARMCPU *target_cpu = ARM_CPU(target_cpu_state); + struct cpu_on_info *info = (struct cpu_on_info *) data.host_ptr; + bool old_powering_on; + + /* Initialize the cpu we are turning on */ + cpu_reset(target_cpu_state); + target_cpu_state->halted = 0; + + if (info->target_aa64) { + if ((info->target_el < 3) && arm_feature(&target_cpu->env, + ARM_FEATURE_EL3)) { + /* + * As target mode is AArch64, we need to set lower + * exception level (the requested level 2) to AArch64 + */ + target_cpu->env.cp15.scr_el3 |= SCR_RW; + } + + if ((info->target_el < 2) && arm_feature(&target_cpu->env, + ARM_FEATURE_EL2)) { + /* + * As target mode is AArch64, we need to set lower + * exception level (the requested level 1) to AArch64 + */ + target_cpu->env.cp15.hcr_el2 |= HCR_RW; + } + + target_cpu->env.pstate = aarch64_pstate_mode(info->target_el, true); + } else { + /* We are requested to boot in AArch32 mode */ + static const uint32_t mode_for_el[] = { 0, + ARM_CPU_MODE_SVC, + ARM_CPU_MODE_HYP, + ARM_CPU_MODE_SVC }; + + cpsr_write(&target_cpu->env, mode_for_el[info->target_el], CPSR_M, + CPSRWriteRaw); + } + + if (info->target_el == 3) { + /* Processor is in secure mode */ + target_cpu->env.cp15.scr_el3 &= ~SCR_NS; + } else { + /* Processor is not in secure mode */ + target_cpu->env.cp15.scr_el3 |= SCR_NS; + } + + /* We check if the started CPU is now at the correct level */ + assert(info->target_el == arm_current_el(&target_cpu->env)); + + if (info->target_aa64) { + target_cpu->env.xregs[0] = info->context_id; + target_cpu->env.thumb = false; + } else { + target_cpu->env.regs[0] = info->context_id; + target_cpu->env.thumb = info->entry & 1; + info->entry &= 0xfffffffe; + } + + /* Start the new CPU at the requested address */ + cpu_set_pc(target_cpu_state, info->entry); + + g_free(info); + + /* Finally set the power status */ + atomic_mb_set(&target_cpu->powered_off, false); + old_powering_on = atomic_cmpxchg(&target_cpu->powering_on, true, false); + assert(old_powering_on); +} + int arm_set_cpu_on(uint64_t cpuid, uint64_t entry, uint64_t context_id, uint32_t target_el, bool target_aa64) { CPUState *target_cpu_state; ARMCPU *target_cpu; + struct cpu_on_info *info; DPRINTF("cpu %" PRId64 " (EL %d, %s) @ 0x%" PRIx64 " with R0 = 0x%" PRIx64 "\n", cpuid, target_el, target_aa64 ? "aarch64" : "aarch32", entry, @@ -77,7 +159,7 @@ int arm_set_cpu_on(uint64_t cpuid, uint64_t entry, uint64_t context_id, } target_cpu = ARM_CPU(target_cpu_state); - if (!target_cpu->powered_off) { + if (!atomic_mb_read(&target_cpu->powered_off)) { qemu_log_mask(LOG_GUEST_ERROR, "[ARM]%s: CPU %" PRId64 " is already on\n", __func__, cpuid); @@ -109,69 +191,45 @@ int arm_set_cpu_on(uint64_t cpuid, uint64_t entry, uint64_t context_id, return QEMU_ARM_POWERCTL_INVALID_PARAM; } - /* Initialize the cpu we are turning on */ - cpu_reset(target_cpu_state); - target_cpu->powered_off = false; - target_cpu_state->halted = 0; - - if (target_aa64) { - if ((target_el < 3) && arm_feature(&target_cpu->env, ARM_FEATURE_EL3)) { - /* - * As target mode is AArch64, we need to set lower - * exception level (the requested level 2) to AArch64 - */ - target_cpu->env.cp15.scr_el3 |= SCR_RW; - } - - if ((target_el < 2) && arm_feature(&target_cpu->env, ARM_FEATURE_EL2)) { - /* - * As target mode is AArch64, we need to set lower - * exception level (the requested level 1) to AArch64 - */ - target_cpu->env.cp15.hcr_el2 |= HCR_RW; - } - - target_cpu->env.pstate = aarch64_pstate_mode(target_el, true); - } else { - /* We are requested to boot in AArch32 mode */ - static uint32_t mode_for_el[] = { 0, - ARM_CPU_MODE_SVC, - ARM_CPU_MODE_HYP, - ARM_CPU_MODE_SVC }; - - cpsr_write(&target_cpu->env, mode_for_el[target_el], CPSR_M, - CPSRWriteRaw); - } - - if (target_el == 3) { - /* Processor is in secure mode */ - target_cpu->env.cp15.scr_el3 &= ~SCR_NS; - } else { - /* Processor is not in secure mode */ - target_cpu->env.cp15.scr_el3 |= SCR_NS; - } - - /* We check if the started CPU is now at the correct level */ - assert(target_el == arm_current_el(&target_cpu->env)); - - if (target_aa64) { - target_cpu->env.xregs[0] = context_id; - target_cpu->env.thumb = false; - } else { - target_cpu->env.regs[0] = context_id; - target_cpu->env.thumb = entry & 1; - entry &= 0xfffffffe; + /* + * If another CPU has powered the target on we are in the state + * ON_PENDING and additional attempts to power on the CPU should + * fail (see 6.6 Implementation CPU_ON/CPU_OFF races in the PSCI + * spec) + */ + if (atomic_cmpxchg(&target_cpu->powering_on, false, true)) { + qemu_log_mask(LOG_GUEST_ERROR, + "[ARM]%s: CPU %" PRId64 " is already powering on\n", + __func__, cpuid); + return QEMU_ARM_POWERCTL_ON_PENDING; } - /* Start the new CPU at the requested address */ - cpu_set_pc(target_cpu_state, entry); + /* To avoid racing with a CPU we are just kicking off we do the + * final bit of preparation for the work in the target CPUs + * context. + */ + info = g_new(struct cpu_on_info, 1); + info->entry = entry; + info->context_id = context_id; + info->target_el = target_el; + info->target_aa64 = target_aa64; - qemu_cpu_kick(target_cpu_state); + async_run_on_cpu(target_cpu_state, arm_set_cpu_on_async_work, + RUN_ON_CPU_HOST_PTR(info)); /* We are good to go */ return QEMU_ARM_POWERCTL_RET_SUCCESS; } +static void arm_set_cpu_off_async_work(CPUState *target_cpu_state, + run_on_cpu_data data) +{ + ARMCPU *target_cpu = ARM_CPU(target_cpu_state); + atomic_mb_set(&target_cpu->powered_off, true); + target_cpu_state->halted = 1; + target_cpu_state->exception_index = EXCP_HLT; +} + int arm_set_cpu_off(uint64_t cpuid) { CPUState *target_cpu_state; @@ -185,22 +243,27 @@ int arm_set_cpu_off(uint64_t cpuid) return QEMU_ARM_POWERCTL_INVALID_PARAM; } target_cpu = ARM_CPU(target_cpu_state); - if (target_cpu->powered_off) { + if (atomic_mb_read(&target_cpu->powered_off)) { qemu_log_mask(LOG_GUEST_ERROR, "[ARM]%s: CPU %" PRId64 " is already off\n", __func__, cpuid); return QEMU_ARM_POWERCTL_IS_OFF; } - target_cpu->powered_off = true; - target_cpu_state->halted = 1; - target_cpu_state->exception_index = EXCP_HLT; - cpu_loop_exit(target_cpu_state); - /* notreached */ + /* Queue work to run under the target vCPUs context */ + async_run_on_cpu(target_cpu_state, arm_set_cpu_off_async_work, + RUN_ON_CPU_NULL); return QEMU_ARM_POWERCTL_RET_SUCCESS; } +static void arm_reset_cpu_async_work(CPUState *target_cpu_state, + run_on_cpu_data data) +{ + /* Reset the cpu */ + cpu_reset(target_cpu_state); +} + int arm_reset_cpu(uint64_t cpuid) { CPUState *target_cpu_state; @@ -221,8 +284,9 @@ int arm_reset_cpu(uint64_t cpuid) return QEMU_ARM_POWERCTL_IS_OFF; } - /* Reset the cpu */ - cpu_reset(target_cpu_state); + /* Queue work to run under the target vCPUs context */ + async_run_on_cpu(target_cpu_state, arm_reset_cpu_async_work, + RUN_ON_CPU_NULL); return QEMU_ARM_POWERCTL_RET_SUCCESS; } diff --git a/target/arm/arm-powerctl.h b/target/arm/arm-powerctl.h index 98ee04989b..04353923c0 100644 --- a/target/arm/arm-powerctl.h +++ b/target/arm/arm-powerctl.h @@ -17,6 +17,7 @@ #define QEMU_ARM_POWERCTL_INVALID_PARAM QEMU_PSCI_RET_INVALID_PARAMS #define QEMU_ARM_POWERCTL_ALREADY_ON QEMU_PSCI_RET_ALREADY_ON #define QEMU_ARM_POWERCTL_IS_OFF QEMU_PSCI_RET_DENIED +#define QEMU_ARM_POWERCTL_ON_PENDING QEMU_PSCI_RET_ON_PENDING /* * arm_get_cpu_by_id: @@ -43,6 +44,7 @@ CPUState *arm_get_cpu_by_id(uint64_t cpuid); * Returns: QEMU_ARM_POWERCTL_RET_SUCCESS on success. * QEMU_ARM_POWERCTL_INVALID_PARAM if bad parameters are provided. * QEMU_ARM_POWERCTL_ALREADY_ON if the CPU was already started. + * QEMU_ARM_POWERCTL_ON_PENDING if the CPU is still powering up */ int arm_set_cpu_on(uint64_t cpuid, uint64_t entry, uint64_t context_id, uint32_t target_el, bool target_aa64); diff --git a/target/arm/cpu.h b/target/arm/cpu.h index 39bff86daf..b8f82d5d20 100644 --- a/target/arm/cpu.h +++ b/target/arm/cpu.h @@ -582,8 +582,16 @@ struct ARMCPU { /* Should CPU start in PSCI powered-off state? */ bool start_powered_off; - /* CPU currently in PSCI powered-off state */ + /* CPU PSCI state. + * + * For TCG these can be cross-vCPU accesses and should be done + * atomically to avoid races. + * + * - powered_off indicates the vCPU state + * - powering_on true if the vCPU has had a CPU_ON but not yet up + */ bool powered_off; + bool powering_on; /* PSCI ON_PENDING */ /* CPU has virtualization extension */ bool has_el2; /* CPU has security extension */