diff mbox series

[v2,08/10] target/arm: Implement the ARMv8.1-LOR extension

Message ID 20181203203839.757-9-richard.henderson@linaro.org
State New
Headers show
Series target/arm: LOR, HPD, AA32HPD | expand

Commit Message

Richard Henderson Dec. 3, 2018, 8:38 p.m. UTC
Provide a trivial implementation with zero limited ordering regions,
which causes the LDLAR and STLLR instructions to devolve into the
LDAR and STLR instructions from the base ARMv8.0 instruction set.

Signed-off-by: Richard Henderson <richard.henderson@linaro.org>


---
v2: Mark LORID_EL1 read-only.
    Add TLOR access checks.
    Conditionally unmask TLOR in hcr/scr_write.
---
 target/arm/cpu.h           |  5 +++
 target/arm/cpu64.c         |  4 ++
 target/arm/helper.c        | 91 ++++++++++++++++++++++++++++++++++++++
 target/arm/translate-a64.c | 12 +++++
 4 files changed, 112 insertions(+)

-- 
2.17.2

Comments

Peter Maydell Dec. 6, 2018, 1:49 p.m. UTC | #1
On Mon, 3 Dec 2018 at 20:38, Richard Henderson
<richard.henderson@linaro.org> wrote:
>

> Provide a trivial implementation with zero limited ordering regions,

> which causes the LDLAR and STLLR instructions to devolve into the

> LDAR and STLR instructions from the base ARMv8.0 instruction set.

>

> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>

>

> ---

> v2: Mark LORID_EL1 read-only.

>     Add TLOR access checks.

>     Conditionally unmask TLOR in hcr/scr_write.

> ---

>  target/arm/cpu.h           |  5 +++

>  target/arm/cpu64.c         |  4 ++

>  target/arm/helper.c        | 91 ++++++++++++++++++++++++++++++++++++++

>  target/arm/translate-a64.c | 12 +++++

>  4 files changed, 112 insertions(+)

>

> diff --git a/target/arm/cpu.h b/target/arm/cpu.h

> index a84101efa9..ba0c368292 100644

> --- a/target/arm/cpu.h

> +++ b/target/arm/cpu.h

> @@ -3285,6 +3285,11 @@ static inline bool isar_feature_aa64_atomics(const ARMISARegisters *id)

>      return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, ATOMIC) != 0;

>  }

>

> +static inline bool isar_feature_aa64_lor(const ARMISARegisters *id)

> +{

> +    return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR1, LO) != 0;


You're testing the wrong ID register here...

> +}

> +


> +static CPAccessResult access_lor_other(CPUARMState *env,

> +                                       const ARMCPRegInfo *ri, bool isread)

> +{

> +    int el = arm_current_el(env);

> +

> +    if (arm_is_secure_below_el3(env)) {

> +        /* Access denied in secure mode.  */

> +        return CP_ACCESS_TRAP;

> +    }

> +    if (el < 2 && arm_el_is_aa64(env, 2)) {


You don't need to explicitly check if EL2 is AArch64 --
these registers are AArch64 only so if we accessed them
from a lower EL then that EL is AArch64 and so all the
ELs above it must be too.

> +        uint64_t hcr = arm_hcr_el2_eff(env);

> +        if (hcr & HCR_E2H) {

> +            hcr &= HCR_TLOR;

> +        } else {

> +            hcr &= HCR_TGE | HCR_TLOR;


This doesn't make sense to me: I think TGE can't be 1 here.
We know (.access flag) that we're not in EL0. We know from
the el < 2 that we're in EL1, and we've already ruled out
Secure EL1. So this is NS EL1. If E2H == 0 then TGE can't
be set, because E2H,TGE={0,1} means the CPU can never be
in NS EL1.

(these remarks apply also to the access_lorid function)

> +        }

> +        if (hcr == HCR_TLOR) {

> +            return CP_ACCESS_TRAP_EL2;

> +        }

> +    }

> +    if (el < 3 && (env->cp15.scr_el3 & SCR_TLOR)) {

> +        return CP_ACCESS_TRAP_EL3;

> +    }

> +    return CP_ACCESS_OK;

> +}

> +

>  void register_cp_regs_for_features(ARMCPU *cpu)

>  {

>      /* Register all the coprocessor registers based on feature bits */

> @@ -5741,6 +5800,38 @@ void register_cp_regs_for_features(ARMCPU *cpu)

>          define_one_arm_cp_reg(cpu, &sctlr);

>      }

>

> +    if (cpu_isar_feature(aa64_lor, cpu)) {

> +        /*

> +         * A trivial implementation of ARMv8.1-LOR leaves all of these

> +         * registers fixed at 0, which indicates that there are zero

> +         * supported Limited Ordering regions.

> +         */

> +        static const ARMCPRegInfo lor_reginfo[] = {

> +            { .name = "LORSA_EL1", .state = ARM_CP_STATE_AA64,

> +              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 0,

> +              .access = PL1_RW, .accessfn = access_lor_other,

> +              .type = ARM_CP_CONST, .resetvalue = 0 },

> +            { .name = "LOREA_EL1", .state = ARM_CP_STATE_AA64,

> +              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 1,

> +              .access = PL1_RW, .accessfn = access_lor_other,

> +              .type = ARM_CP_CONST, .resetvalue = 0 },

> +            { .name = "LORN_EL1", .state = ARM_CP_STATE_AA64,

> +              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 2,

> +              .access = PL1_RW, .accessfn = access_lor_other,

> +              .type = ARM_CP_CONST, .resetvalue = 0 },

> +            { .name = "LORC_EL1", .state = ARM_CP_STATE_AA64,

> +              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 3,

> +              .access = PL1_RW, .accessfn = access_lor_other,

> +              .type = ARM_CP_CONST, .resetvalue = 0 },

> +            { .name = "LORID_EL1", .state = ARM_CP_STATE_AA64,

> +              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 7,

> +              .access = PL1_R, .accessfn = access_lorid,

> +              .type = ARM_CP_CONST, .resetvalue = 0 },

> +            REGINFO_SENTINEL

> +        };

> +        define_arm_cp_regs(cpu, lor_reginfo);

> +    }

> +

>      if (cpu_isar_feature(aa64_sve, cpu)) {

>          define_one_arm_cp_reg(cpu, &zcr_el1_reginfo);

>          if (arm_feature(env, ARM_FEATURE_EL2)) {

> diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c

> index fd36425f1a..5952a9d1cc 100644

> --- a/target/arm/translate-a64.c

> +++ b/target/arm/translate-a64.c

> @@ -2290,6 +2290,12 @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)

>          }

>          return;

>

> +    case 0x8: /* STLLR */

> +        if (!dc_isar_feature(aa64_lor, s)) {

> +            break;

> +        }

> +        /* StoreLORelease is the same as Store-Release for QEMU.  */

> +        /* fallthru */


We are as usual a bit inconsistent, but we seem to use the
spelling "fall through" for this linter-hint more often than
any of the other variations.

>      case 0x9: /* STLR */

>          /* Generate ISS for non-exclusive accesses including LASR.  */

>          if (rn == 31) {

> @@ -2301,6 +2307,12 @@ static void disas_ldst_excl(DisasContext *s, uint32_t insn)

>                    disas_ldst_compute_iss_sf(size, false, 0), is_lasr);

>          return;

>

> +    case 0xc: /* LDLAR */

> +        if (!dc_isar_feature(aa64_lor, s)) {

> +            break;

> +        }

> +        /* LoadLOAcquire is the same as Load-Acquire for QEMU.  */

> +        /* fallthru */

>      case 0xd: /* LDAR */

>          /* Generate ISS for non-exclusive accesses including LASR.  */

>          if (rn == 31) {


thanks
-- PMM
Richard Henderson Dec. 6, 2018, 1:58 p.m. UTC | #2
On 12/6/18 7:49 AM, Peter Maydell wrote:
>> +static inline bool isar_feature_aa64_lor(const ARMISARegisters *id)

>> +{

>> +    return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR1, LO) != 0;

> 

> You're testing the wrong ID register here...


Oops, I thought I fixed that...

>> +static CPAccessResult access_lor_other(CPUARMState *env,

>> +                                       const ARMCPRegInfo *ri, bool isread)

>> +{

>> +    int el = arm_current_el(env);

>> +

>> +    if (arm_is_secure_below_el3(env)) {

>> +        /* Access denied in secure mode.  */

>> +        return CP_ACCESS_TRAP;

>> +    }

>> +    if (el < 2 && arm_el_is_aa64(env, 2)) {

> 

> You don't need to explicitly check if EL2 is AArch64 --

> these registers are AArch64 only so if we accessed them

> from a lower EL then that EL is AArch64 and so all the

> ELs above it must be too.


Ok.

> 

>> +        uint64_t hcr = arm_hcr_el2_eff(env);

>> +        if (hcr & HCR_E2H) {

>> +            hcr &= HCR_TLOR;

>> +        } else {

>> +            hcr &= HCR_TGE | HCR_TLOR;

> 

> This doesn't make sense to me: I think TGE can't be 1 here.

> We know (.access flag) that we're not in EL0. We know from

> the el < 2 that we're in EL1, and we've already ruled out

> Secure EL1. So this is NS EL1. If E2H == 0 then TGE can't

> be set, because E2H,TGE={0,1} means the CPU can never be

> in NS EL1.


Ok.  Perhaps I was too blindly following the access cases listed in the ARM and
not applying any extra thought.

> (these remarks apply also to the access_lorid function)


I think I will split out a shared subroutine for these.

>> +    case 0x8: /* STLLR */

>> +        if (!dc_isar_feature(aa64_lor, s)) {

>> +            break;

>> +        }

>> +        /* StoreLORelease is the same as Store-Release for QEMU.  */

>> +        /* fallthru */

> 

> We are as usual a bit inconsistent, but we seem to use the

> spelling "fall through" for this linter-hint more often than

> any of the other variations.


I was going to say something about old habits, but I do note that I've somehow
migrated from the witheringly old FALLTHRU.  So perhaps I can learn eventually...


r~
Richard Henderson Dec. 6, 2018, 4:25 p.m. UTC | #3
On 12/6/18 7:49 AM, Peter Maydell wrote:
>> +        uint64_t hcr = arm_hcr_el2_eff(env);

>> +        if (hcr & HCR_E2H) {

>> +            hcr &= HCR_TLOR;

>> +        } else {

>> +            hcr &= HCR_TGE | HCR_TLOR;

> This doesn't make sense to me


The logic is backward.  What I was after was

  if (hcr & HCR_E2H) {
      hcr &= HCR_TGE | HCR_TLOR;
  } else {
      hcr &= HCR_TLOR;
  }
  if (hcr == HCR_TLOR) {
      trap to el2.
  }

I.e. swap the then and else condition.  This takes care of the two rules

 -- If (SCR_EL3.NS == 1 || SCR_EL3.EEL2 == 1) && IsUsingAArch64(EL2)
    && HCR_EL2.E2H == 0 && HCR_EL2.TLOR == 1, then
    accesses at EL1 are trapped to EL2.
 -- If (SCR_EL3.NS == 1 || SCR_EL3.EEL2 == 1) && IsUsingAArch64(EL2)
    && HCR_EL2.E2H == 1 && HCR_EL2.TGE == 0 && HCR_EL2.TLOR == 1,
    then accesses at EL1 are trapped to EL2.


r~
diff mbox series

Patch

diff --git a/target/arm/cpu.h b/target/arm/cpu.h
index a84101efa9..ba0c368292 100644
--- a/target/arm/cpu.h
+++ b/target/arm/cpu.h
@@ -3285,6 +3285,11 @@  static inline bool isar_feature_aa64_atomics(const ARMISARegisters *id)
     return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, ATOMIC) != 0;
 }
 
+static inline bool isar_feature_aa64_lor(const ARMISARegisters *id)
+{
+    return FIELD_EX64(id->id_aa64mmfr0, ID_AA64MMFR1, LO) != 0;
+}
+
 static inline bool isar_feature_aa64_rdm(const ARMISARegisters *id)
 {
     return FIELD_EX64(id->id_aa64isar0, ID_AA64ISAR0, RDM) != 0;
diff --git a/target/arm/cpu64.c b/target/arm/cpu64.c
index 0babe483ac..aac6283018 100644
--- a/target/arm/cpu64.c
+++ b/target/arm/cpu64.c
@@ -324,6 +324,10 @@  static void aarch64_max_initfn(Object *obj)
         t = FIELD_DP64(t, ID_AA64PFR0, ADVSIMD, 1);
         cpu->isar.id_aa64pfr0 = t;
 
+        t = cpu->isar.id_aa64mmfr1;
+        t = FIELD_DP64(t, ID_AA64MMFR1, LO, 1);
+        cpu->isar.id_aa64mmfr1 = t;
+
         /* Replicate the same data to the 32-bit id registers.  */
         u = cpu->isar.id_isar5;
         u = FIELD_DP32(u, ID_ISAR5, AES, 2); /* AES + PMULL */
diff --git a/target/arm/helper.c b/target/arm/helper.c
index faf7f922bf..a0ee1fbc5a 100644
--- a/target/arm/helper.c
+++ b/target/arm/helper.c
@@ -1281,6 +1281,7 @@  static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
 {
     /* Begin with base v8.0 state.  */
     uint32_t valid_mask = 0x3fff;
+    ARMCPU *cpu = arm_env_get_cpu(env);
 
     if (arm_el_is_aa64(env, 3)) {
         value |= SCR_FW | SCR_AW;   /* these two bits are RES1.  */
@@ -1303,6 +1304,9 @@  static void scr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
             valid_mask &= ~SCR_SMD;
         }
     }
+    if (cpu_isar_feature(aa64_lor, cpu)) {
+        valid_mask |= SCR_TLOR;
+    }
 
     /* Clear all-context RES0 bits.  */
     value &= valid_mask;
@@ -3950,6 +3954,9 @@  static void hcr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value)
          */
         valid_mask &= ~HCR_TSC;
     }
+    if (cpu_isar_feature(aa64_lor, cpu)) {
+        valid_mask |= HCR_TLOR;
+    }
 
     /* Clear RES0 bits.  */
     value &= valid_mask;
@@ -5004,6 +5011,58 @@  static uint64_t id_aa64pfr0_read(CPUARMState *env, const ARMCPRegInfo *ri)
     return pfr0;
 }
 
+static CPAccessResult access_lorid(CPUARMState *env, const ARMCPRegInfo *ri,
+                                   bool isread)
+{
+    int el = arm_current_el(env);
+
+    if (arm_is_secure_below_el3(env)) {
+        /* Access ok in secure mode.  */
+        return CP_ACCESS_OK;
+    }
+    if (el < 2 && arm_el_is_aa64(env, 2)) {
+        uint64_t hcr = arm_hcr_el2_eff(env);
+        if (hcr & HCR_E2H) {
+            hcr &= HCR_TLOR;
+        } else {
+            hcr &= HCR_TGE | HCR_TLOR;
+        }
+        if (hcr == HCR_TLOR) {
+            return CP_ACCESS_TRAP_EL2;
+        }
+    }
+    if (el < 3 && (env->cp15.scr_el3 & SCR_TLOR)) {
+        return CP_ACCESS_TRAP_EL3;
+    }
+    return CP_ACCESS_OK;
+}
+
+static CPAccessResult access_lor_other(CPUARMState *env,
+                                       const ARMCPRegInfo *ri, bool isread)
+{
+    int el = arm_current_el(env);
+
+    if (arm_is_secure_below_el3(env)) {
+        /* Access denied in secure mode.  */
+        return CP_ACCESS_TRAP;
+    }
+    if (el < 2 && arm_el_is_aa64(env, 2)) {
+        uint64_t hcr = arm_hcr_el2_eff(env);
+        if (hcr & HCR_E2H) {
+            hcr &= HCR_TLOR;
+        } else {
+            hcr &= HCR_TGE | HCR_TLOR;
+        }
+        if (hcr == HCR_TLOR) {
+            return CP_ACCESS_TRAP_EL2;
+        }
+    }
+    if (el < 3 && (env->cp15.scr_el3 & SCR_TLOR)) {
+        return CP_ACCESS_TRAP_EL3;
+    }
+    return CP_ACCESS_OK;
+}
+
 void register_cp_regs_for_features(ARMCPU *cpu)
 {
     /* Register all the coprocessor registers based on feature bits */
@@ -5741,6 +5800,38 @@  void register_cp_regs_for_features(ARMCPU *cpu)
         define_one_arm_cp_reg(cpu, &sctlr);
     }
 
+    if (cpu_isar_feature(aa64_lor, cpu)) {
+        /*
+         * A trivial implementation of ARMv8.1-LOR leaves all of these
+         * registers fixed at 0, which indicates that there are zero
+         * supported Limited Ordering regions.
+         */
+        static const ARMCPRegInfo lor_reginfo[] = {
+            { .name = "LORSA_EL1", .state = ARM_CP_STATE_AA64,
+              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 0,
+              .access = PL1_RW, .accessfn = access_lor_other,
+              .type = ARM_CP_CONST, .resetvalue = 0 },
+            { .name = "LOREA_EL1", .state = ARM_CP_STATE_AA64,
+              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 1,
+              .access = PL1_RW, .accessfn = access_lor_other,
+              .type = ARM_CP_CONST, .resetvalue = 0 },
+            { .name = "LORN_EL1", .state = ARM_CP_STATE_AA64,
+              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 2,
+              .access = PL1_RW, .accessfn = access_lor_other,
+              .type = ARM_CP_CONST, .resetvalue = 0 },
+            { .name = "LORC_EL1", .state = ARM_CP_STATE_AA64,
+              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 3,
+              .access = PL1_RW, .accessfn = access_lor_other,
+              .type = ARM_CP_CONST, .resetvalue = 0 },
+            { .name = "LORID_EL1", .state = ARM_CP_STATE_AA64,
+              .opc0 = 3, .opc1 = 0, .crn = 10, .crm = 4, .opc2 = 7,
+              .access = PL1_R, .accessfn = access_lorid,
+              .type = ARM_CP_CONST, .resetvalue = 0 },
+            REGINFO_SENTINEL
+        };
+        define_arm_cp_regs(cpu, lor_reginfo);
+    }
+
     if (cpu_isar_feature(aa64_sve, cpu)) {
         define_one_arm_cp_reg(cpu, &zcr_el1_reginfo);
         if (arm_feature(env, ARM_FEATURE_EL2)) {
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c
index fd36425f1a..5952a9d1cc 100644
--- a/target/arm/translate-a64.c
+++ b/target/arm/translate-a64.c
@@ -2290,6 +2290,12 @@  static void disas_ldst_excl(DisasContext *s, uint32_t insn)
         }
         return;
 
+    case 0x8: /* STLLR */
+        if (!dc_isar_feature(aa64_lor, s)) {
+            break;
+        }
+        /* StoreLORelease is the same as Store-Release for QEMU.  */
+        /* fallthru */
     case 0x9: /* STLR */
         /* Generate ISS for non-exclusive accesses including LASR.  */
         if (rn == 31) {
@@ -2301,6 +2307,12 @@  static void disas_ldst_excl(DisasContext *s, uint32_t insn)
                   disas_ldst_compute_iss_sf(size, false, 0), is_lasr);
         return;
 
+    case 0xc: /* LDLAR */
+        if (!dc_isar_feature(aa64_lor, s)) {
+            break;
+        }
+        /* LoadLOAcquire is the same as Load-Acquire for QEMU.  */
+        /* fallthru */
     case 0xd: /* LDAR */
         /* Generate ISS for non-exclusive accesses including LASR.  */
         if (rn == 31) {