Message ID | 20181203203839.757-9-richard.henderson@linaro.org |
---|---|
State | New |
Headers | show |
Series | target/arm: LOR, HPD, AA32HPD | expand |
On Mon, 3 Dec 2018 at 20:38, Richard Henderson <richard.henderson@linaro.org> wrote: > > Provide a trivial implementation with zero limited ordering regions, > which causes the LDLAR and STLLR instructions to devolve into the > LDAR and STLR instructions from the base ARMv8.0 instruction set. > > Signed-off-by: Richard Henderson <richard.henderson@linaro.org> > > --- > v2: Mark LORID_EL1 read-only. > Add TLOR access checks. > Conditionally unmask TLOR in hcr/scr_write. > --- > target/arm/cpu.h | 5 +++ > target/arm/cpu64.c | 4 ++ > target/arm/helper.c | 91 ++++++++++++++++++++++++++++++++++++++ > target/arm/translate-a64.c | 12 +++++ > 4 files changed, 112 insertions(+) > > diff --git a/target/arm/cpu.h b/target/arm/cpu.h > index a84101efa9..ba0c368292 100644 > --- a/target/arm/cpu.h > +++ b/target/arm/cpu.h > @@ -3285,6 +3285,11 @@ static inline bool isar_feature_aa64_atomics(const ARMISARegisters *id) > return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, ATOMIC) != 0; > } > > +static inline bool isar_feature_aa64_lor(const ARMISARegisters *id) > +{ > + return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR1, LO) != 0; You're testing the wrong ID register here... > +} > + > +static CPAccessResult access_lor_other(CPUARMState *env, > + const ARMCPRegInfo *ri, bool isread) > +{ > + int el = arm_current_el(env); > + > + if (arm_is_secure_below_el3(env)) { > + /* Access denied in secure mode. */ > + return CP_ACCESS_TRAP; > + } > + if (el < 2 && arm_el_is_aa64(env, 2)) { You don't need to explicitly check if EL2 is AArch64 -- these registers are AArch64 only so if we accessed them from a lower EL then that EL is AArch64 and so all the ELs above it must be too. > + uint64_t hcr = arm_hcr_el2_eff(env); > + if (hcr & HCR_E2H) { > + hcr &= HCR_TLOR; > + } else { > + hcr &= HCR_TGE | HCR_TLOR; This doesn't make sense to me: I think TGE can't be 1 here. We know (.access flag) that we're not in EL0. We know from the el < 2 that we're in EL1, and we've already ruled out Secure EL1. So this is NS EL1. If E2H == 0 then TGE can't be set, because E2H,TGE={0,1} means the CPU can never be in NS EL1. (these remarks apply also to the access_lorid function) > + } > + if (hcr == HCR_TLOR) { > + return CP_ACCESS_TRAP_EL2; > + } > + } > + if (el < 3 && (env->cp15.scr_el3 & SCR_TLOR)) { > + return CP_ACCESS_TRAP_EL3; > + } > + return CP_ACCESS_OK; > +} > + > void register_cp_regs_for_features(ARMCPU *cpu) > { > /* Register all the coprocessor registers based on feature bits */ > @@ -5741,6 +5800,38 @@ void register_cp_regs_for_features(ARMCPU *cpu) > define_one_arm_cp_reg(cpu, &sctlr); > } > > + if (cpu_isar_feature(aa64_lor, cpu)) { > + /* > + * A trivial implementation of ARMv8.1-LOR leaves all of these > + * registers fixed at 0, which indicates that there are zero > + * supported Limited Ordering regions. > + */ > + static const ARMCPRegInfo lor_reginfo[] = { > + { .name = "LORSA_EL1", .state = ARM_CP_STATE_AA64, > + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 0, > + .access = PL1_RW, .accessfn = access_lor_other, > + .type = ARM_CP_CONST, .resetvalue = 0 }, > + { .name = "LOREA_EL1", .state = ARM_CP_STATE_AA64, > + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 1, > + .access = PL1_RW, .accessfn = access_lor_other, > + .type = ARM_CP_CONST, .resetvalue = 0 }, > + { .name = "LORN_EL1", .state = ARM_CP_STATE_AA64, > + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 2, > + .access = PL1_RW, .accessfn = access_lor_other, > + .type = ARM_CP_CONST, .resetvalue = 0 }, > + { .name = "LORC_EL1", .state = ARM_CP_STATE_AA64, > + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 3, > + .access = PL1_RW, .accessfn = access_lor_other, > + .type = ARM_CP_CONST, .resetvalue = 0 }, > + { .name = "LORID_EL1", .state = ARM_CP_STATE_AA64, > + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 7, > + .access = PL1_R, .accessfn = access_lorid, > + .type = ARM_CP_CONST, .resetvalue = 0 }, > + REGINFO_SENTINEL > + }; > + define_arm_cp_regs(cpu, lor_reginfo); > + } > + > if (cpu_isar_feature(aa64_sve, cpu)) { > define_one_arm_cp_reg(cpu, &zcr_el1_reginfo); > if (arm_feature(env, ARM_FEATURE_EL2)) { > diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c > index fd36425f1a..5952a9d1cc 100644 > --- a/target/arm/translate-a64.c > +++ b/target/arm/translate-a64.c > @@ -2290,6 +2290,12 @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn) > } > return; > > + case 0x8: /* STLLR */ > + if (!dc_isar_feature(aa64_lor, s)) { > + break; > + } > + /* StoreLORelease is the same as Store-Release for QEMU. */ > + /* fallthru */ We are as usual a bit inconsistent, but we seem to use the spelling "fall through" for this linter-hint more often than any of the other variations. > case 0x9: /* STLR */ > /* Generate ISS for non-exclusive accesses including LASR. */ > if (rn == 31) { > @@ -2301,6 +2307,12 @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn) > disas_ldst_compute_iss_sf(size, false, 0), is_lasr); > return; > > + case 0xc: /* LDLAR */ > + if (!dc_isar_feature(aa64_lor, s)) { > + break; > + } > + /* LoadLOAcquire is the same as Load-Acquire for QEMU. */ > + /* fallthru */ > case 0xd: /* LDAR */ > /* Generate ISS for non-exclusive accesses including LASR. */ > if (rn == 31) { thanks -- PMM
On 12/6/18 7:49 AM, Peter Maydell wrote: >> +static inline bool isar_feature_aa64_lor(const ARMISARegisters *id) >> +{ >> + return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR1, LO) != 0; > > You're testing the wrong ID register here... Oops, I thought I fixed that... >> +static CPAccessResult access_lor_other(CPUARMState *env, >> + const ARMCPRegInfo *ri, bool isread) >> +{ >> + int el = arm_current_el(env); >> + >> + if (arm_is_secure_below_el3(env)) { >> + /* Access denied in secure mode. */ >> + return CP_ACCESS_TRAP; >> + } >> + if (el < 2 && arm_el_is_aa64(env, 2)) { > > You don't need to explicitly check if EL2 is AArch64 -- > these registers are AArch64 only so if we accessed them > from a lower EL then that EL is AArch64 and so all the > ELs above it must be too. Ok. > >> + uint64_t hcr = arm_hcr_el2_eff(env); >> + if (hcr & HCR_E2H) { >> + hcr &= HCR_TLOR; >> + } else { >> + hcr &= HCR_TGE | HCR_TLOR; > > This doesn't make sense to me: I think TGE can't be 1 here. > We know (.access flag) that we're not in EL0. We know from > the el < 2 that we're in EL1, and we've already ruled out > Secure EL1. So this is NS EL1. If E2H == 0 then TGE can't > be set, because E2H,TGE={0,1} means the CPU can never be > in NS EL1. Ok. Perhaps I was too blindly following the access cases listed in the ARM and not applying any extra thought. > (these remarks apply also to the access_lorid function) I think I will split out a shared subroutine for these. >> + case 0x8: /* STLLR */ >> + if (!dc_isar_feature(aa64_lor, s)) { >> + break; >> + } >> + /* StoreLORelease is the same as Store-Release for QEMU. */ >> + /* fallthru */ > > We are as usual a bit inconsistent, but we seem to use the > spelling "fall through" for this linter-hint more often than > any of the other variations. I was going to say something about old habits, but I do note that I've somehow migrated from the witheringly old FALLTHRU. So perhaps I can learn eventually... r~
On 12/6/18 7:49 AM, Peter Maydell wrote: >> + uint64_t hcr = arm_hcr_el2_eff(env); >> + if (hcr & HCR_E2H) { >> + hcr &= HCR_TLOR; >> + } else { >> + hcr &= HCR_TGE | HCR_TLOR; > This doesn't make sense to me The logic is backward. What I was after was if (hcr & HCR_E2H) { hcr &= HCR_TGE | HCR_TLOR; } else { hcr &= HCR_TLOR; } if (hcr == HCR_TLOR) { trap to el2. } I.e. swap the then and else condition. This takes care of the two rules -- If (SCR_EL3.NS == 1 || SCR_EL3.EEL2 == 1) && IsUsingAArch64(EL2) && HCR_EL2.E2H == 0 && HCR_EL2.TLOR == 1, then accesses at EL1 are trapped to EL2. -- If (SCR_EL3.NS == 1 || SCR_EL3.EEL2 == 1) && IsUsingAArch64(EL2) && HCR_EL2.E2H == 1 && HCR_EL2.TGE == 0 && HCR_EL2.TLOR == 1, then accesses at EL1 are trapped to EL2. r~
diff --git a/target/arm/cpu.h b/target/arm/cpu.h index a84101efa9..ba0c368292 100644 --- a/target/arm/cpu.h +++ b/target/arm/cpu.h @@ -3285,6 +3285,11 @@ static inline bool isar_feature_aa64_atomics(const ARMISARegisters *id) return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, ATOMIC) != 0; } +static inline bool isar_feature_aa64_lor(const ARMISARegisters *id) +{ + return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR1, LO) != 0; +} + static inline bool isar_feature_aa64_rdm(const ARMISARegisters *id) { return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, RDM) != 0; diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c index 0babe483ac..aac6283018 100644 --- a/target/arm/cpu64.c +++ b/target/arm/cpu64.c @@ -324,6 +324,10 @@ static void aarch64_max_initfn(Object *obj) t = FIELD_DP64(t, ID_AA64PFR0, ADVSIMD, 1); cpu->isar.id_aa64pfr0 = t; + t = cpu->isar.id_aa64mmfr1; + t = FIELD_DP64(t, ID_AA64MMFR1, LO, 1); + cpu->isar.id_aa64mmfr1 = t; + /* Replicate the same data to the 32-bit id registers. */ u = cpu->isar.id_isar5; u = FIELD_DP32(u, ID_ISAR5, AES, 2); /* AES + PMULL */ diff --git a/target/arm/helper.c b/target/arm/helper.c index faf7f922bf..a0ee1fbc5a 100644 --- a/target/arm/helper.c +++ b/target/arm/helper.c @@ -1281,6 +1281,7 @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value) { /* Begin with base v8.0 state. */ uint32_t valid_mask = 0x3fff; + ARMCPU *cpu = arm_env_get_cpu(env); if (arm_el_is_aa64(env, 3)) { value |= SCR_FW | SCR_AW; /* these two bits are RES1. */ @@ -1303,6 +1304,9 @@ static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value) valid_mask &= ~SCR_SMD; } } + if (cpu_isar_feature(aa64_lor, cpu)) { + valid_mask |= SCR_TLOR; + } /* Clear all-context RES0 bits. */ value &= valid_mask; @@ -3950,6 +3954,9 @@ static void hcr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value) */ valid_mask &= ~HCR_TSC; } + if (cpu_isar_feature(aa64_lor, cpu)) { + valid_mask |= HCR_TLOR; + } /* Clear RES0 bits. */ value &= valid_mask; @@ -5004,6 +5011,58 @@ static uint64_t id_aa64pfr0_read(CPUARMState *env, const ARMCPRegInfo *ri) return pfr0; } +static CPAccessResult access_lorid(CPUARMState *env, const ARMCPRegInfo *ri, + bool isread) +{ + int el = arm_current_el(env); + + if (arm_is_secure_below_el3(env)) { + /* Access ok in secure mode. */ + return CP_ACCESS_OK; + } + if (el < 2 && arm_el_is_aa64(env, 2)) { + uint64_t hcr = arm_hcr_el2_eff(env); + if (hcr & HCR_E2H) { + hcr &= HCR_TLOR; + } else { + hcr &= HCR_TGE | HCR_TLOR; + } + if (hcr == HCR_TLOR) { + return CP_ACCESS_TRAP_EL2; + } + } + if (el < 3 && (env->cp15.scr_el3 & SCR_TLOR)) { + return CP_ACCESS_TRAP_EL3; + } + return CP_ACCESS_OK; +} + +static CPAccessResult access_lor_other(CPUARMState *env, + const ARMCPRegInfo *ri, bool isread) +{ + int el = arm_current_el(env); + + if (arm_is_secure_below_el3(env)) { + /* Access denied in secure mode. */ + return CP_ACCESS_TRAP; + } + if (el < 2 && arm_el_is_aa64(env, 2)) { + uint64_t hcr = arm_hcr_el2_eff(env); + if (hcr & HCR_E2H) { + hcr &= HCR_TLOR; + } else { + hcr &= HCR_TGE | HCR_TLOR; + } + if (hcr == HCR_TLOR) { + return CP_ACCESS_TRAP_EL2; + } + } + if (el < 3 && (env->cp15.scr_el3 & SCR_TLOR)) { + return CP_ACCESS_TRAP_EL3; + } + return CP_ACCESS_OK; +} + void register_cp_regs_for_features(ARMCPU *cpu) { /* Register all the coprocessor registers based on feature bits */ @@ -5741,6 +5800,38 @@ void register_cp_regs_for_features(ARMCPU *cpu) define_one_arm_cp_reg(cpu, &sctlr); } + if (cpu_isar_feature(aa64_lor, cpu)) { + /* + * A trivial implementation of ARMv8.1-LOR leaves all of these + * registers fixed at 0, which indicates that there are zero + * supported Limited Ordering regions. + */ + static const ARMCPRegInfo lor_reginfo[] = { + { .name = "LORSA_EL1", .state = ARM_CP_STATE_AA64, + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 0, + .access = PL1_RW, .accessfn = access_lor_other, + .type = ARM_CP_CONST, .resetvalue = 0 }, + { .name = "LOREA_EL1", .state = ARM_CP_STATE_AA64, + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 1, + .access = PL1_RW, .accessfn = access_lor_other, + .type = ARM_CP_CONST, .resetvalue = 0 }, + { .name = "LORN_EL1", .state = ARM_CP_STATE_AA64, + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 2, + .access = PL1_RW, .accessfn = access_lor_other, + .type = ARM_CP_CONST, .resetvalue = 0 }, + { .name = "LORC_EL1", .state = ARM_CP_STATE_AA64, + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 3, + .access = PL1_RW, .accessfn = access_lor_other, + .type = ARM_CP_CONST, .resetvalue = 0 }, + { .name = "LORID_EL1", .state = ARM_CP_STATE_AA64, + .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 7, + .access = PL1_R, .accessfn = access_lorid, + .type = ARM_CP_CONST, .resetvalue = 0 }, + REGINFO_SENTINEL + }; + define_arm_cp_regs(cpu, lor_reginfo); + } + if (cpu_isar_feature(aa64_sve, cpu)) { define_one_arm_cp_reg(cpu, &zcr_el1_reginfo); if (arm_feature(env, ARM_FEATURE_EL2)) { diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c index fd36425f1a..5952a9d1cc 100644 --- a/target/arm/translate-a64.c +++ b/target/arm/translate-a64.c @@ -2290,6 +2290,12 @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn) } return; + case 0x8: /* STLLR */ + if (!dc_isar_feature(aa64_lor, s)) { + break; + } + /* StoreLORelease is the same as Store-Release for QEMU. */ + /* fallthru */ case 0x9: /* STLR */ /* Generate ISS for non-exclusive accesses including LASR. */ if (rn == 31) { @@ -2301,6 +2307,12 @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn) disas_ldst_compute_iss_sf(size, false, 0), is_lasr); return; + case 0xc: /* LDLAR */ + if (!dc_isar_feature(aa64_lor, s)) { + break; + } + /* LoadLOAcquire is the same as Load-Acquire for QEMU. */ + /* fallthru */ case 0xd: /* LDAR */ /* Generate ISS for non-exclusive accesses including LASR. */ if (rn == 31) {
Provide a trivial implementation with zero limited ordering regions, which causes the LDLAR and STLLR instructions to devolve into the LDAR and STLR instructions from the base ARMv8.0 instruction set. Signed-off-by: Richard Henderson <richard.henderson@linaro.org> --- v2: Mark LORID_EL1 read-only. Add TLOR access checks. Conditionally unmask TLOR in hcr/scr_write. --- target/arm/cpu.h | 5 +++ target/arm/cpu64.c | 4 ++ target/arm/helper.c | 91 ++++++++++++++++++++++++++++++++++++++ target/arm/translate-a64.c | 12 +++++ 4 files changed, 112 insertions(+) -- 2.17.2