diff mbox series

[12/34] target/ppc: introduce get_avr64() and set_avr64() helpers for VMX register access

Message ID 20181218063911.2112-13-richard.henderson@linaro.org
State New
Headers show
Series tcg, target/ppc vector improvements | expand

Commit Message

Richard Henderson Dec. 18, 2018, 6:38 a.m. UTC
From: Mark Cave-Ayland <mark.cave-ayland@ilande.co.uk>


These helpers allow us to move AVR register values to/from the specified TCGv_i64
argument.

To prevent VMX helpers accessing the cpu_avr{l,h} arrays directly, add extra TCG
temporaries as required.

Signed-off-by: Mark Cave-Ayland <mark.cave-ayland@ilande.co.uk>

Reviewed-by: Richard Henderson <richard.henderson@linaro.org>

Message-Id: <20181217122405.18732-3-mark.cave-ayland@ilande.co.uk>
---
 target/ppc/translate.c              |  10 +++
 target/ppc/translate/vmx-impl.inc.c | 128 ++++++++++++++++++++++------
 2 files changed, 110 insertions(+), 28 deletions(-)

-- 
2.17.2

Comments

David Gibson Dec. 19, 2018, 6:15 a.m. UTC | #1
On Mon, Dec 17, 2018 at 10:38:49PM -0800, Richard Henderson wrote:
> From: Mark Cave-Ayland <mark.cave-ayland@ilande.co.uk>

> 

> These helpers allow us to move AVR register values to/from the specified TCGv_i64

> argument.

> 

> To prevent VMX helpers accessing the cpu_avr{l,h} arrays directly, add extra TCG

> temporaries as required.

> 

> Signed-off-by: Mark Cave-Ayland <mark.cave-ayland@ilande.co.uk>

> Reviewed-by: Richard Henderson <richard.henderson@linaro.org>


Acked-by: David Gibson <david@gibson.dropbear.id.au>


> Message-Id: <20181217122405.18732-3-mark.cave-ayland@ilande.co.uk>

> ---

>  target/ppc/translate.c              |  10 +++

>  target/ppc/translate/vmx-impl.inc.c | 128 ++++++++++++++++++++++------

>  2 files changed, 110 insertions(+), 28 deletions(-)

> 

> diff --git a/target/ppc/translate.c b/target/ppc/translate.c

> index 1d4bf624a3..fa3e8dc114 100644

> --- a/target/ppc/translate.c

> +++ b/target/ppc/translate.c

> @@ -6704,6 +6704,16 @@ static inline void set_fpr(int regno, TCGv_i64 src)

>      tcg_gen_mov_i64(cpu_fpr[regno], src);

>  }

>  

> +static inline void get_avr64(TCGv_i64 dst, int regno, bool high)

> +{

> +    tcg_gen_mov_i64(dst, (high ? cpu_avrh : cpu_avrl)[regno]);

> +}

> +

> +static inline void set_avr64(int regno, TCGv_i64 src, bool high)

> +{

> +    tcg_gen_mov_i64((high ? cpu_avrh : cpu_avrl)[regno], src);

> +}

> +

>  #include "translate/fp-impl.inc.c"

>  

>  #include "translate/vmx-impl.inc.c"

> diff --git a/target/ppc/translate/vmx-impl.inc.c b/target/ppc/translate/vmx-impl.inc.c

> index 3cb6fc2926..30046c6e31 100644

> --- a/target/ppc/translate/vmx-impl.inc.c

> +++ b/target/ppc/translate/vmx-impl.inc.c

> @@ -18,52 +18,66 @@ static inline TCGv_ptr gen_avr_ptr(int reg)

>  static void glue(gen_, name)(DisasContext *ctx)                                       \

>  {                                                                             \

>      TCGv EA;                                                                  \

> +    TCGv_i64 avr;                                                             \

>      if (unlikely(!ctx->altivec_enabled)) {                                    \

>          gen_exception(ctx, POWERPC_EXCP_VPU);                                 \

>          return;                                                               \

>      }                                                                         \

>      gen_set_access_type(ctx, ACCESS_INT);                                     \

> +    avr = tcg_temp_new_i64();                                                 \

>      EA = tcg_temp_new();                                                      \

>      gen_addr_reg_index(ctx, EA);                                              \

>      tcg_gen_andi_tl(EA, EA, ~0xf);                                            \

>      /* We only need to swap high and low halves. gen_qemu_ld64_i64 does       \

>         necessary 64-bit byteswap already. */                                  \

>      if (ctx->le_mode) {                                                       \

> -        gen_qemu_ld64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA);                \

> +        gen_qemu_ld64_i64(ctx, avr, EA);                                      \

> +        set_avr64(rD(ctx->opcode), avr, false);                               \

>          tcg_gen_addi_tl(EA, EA, 8);                                           \

> -        gen_qemu_ld64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA);                \

> +        gen_qemu_ld64_i64(ctx, avr, EA);                                      \

> +        set_avr64(rD(ctx->opcode), avr, true);                                \

>      } else {                                                                  \

> -        gen_qemu_ld64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA);                \

> +        gen_qemu_ld64_i64(ctx, avr, EA);                                      \

> +        set_avr64(rD(ctx->opcode), avr, true);                                \

>          tcg_gen_addi_tl(EA, EA, 8);                                           \

> -        gen_qemu_ld64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA);                \

> +        gen_qemu_ld64_i64(ctx, avr, EA);                                      \

> +        set_avr64(rD(ctx->opcode), avr, false);                               \

>      }                                                                         \

>      tcg_temp_free(EA);                                                        \

> +    tcg_temp_free_i64(avr);                                                   \

>  }

>  

>  #define GEN_VR_STX(name, opc2, opc3)                                          \

>  static void gen_st##name(DisasContext *ctx)                                   \

>  {                                                                             \

>      TCGv EA;                                                                  \

> +    TCGv_i64 avr;                                                             \

>      if (unlikely(!ctx->altivec_enabled)) {                                    \

>          gen_exception(ctx, POWERPC_EXCP_VPU);                                 \

>          return;                                                               \

>      }                                                                         \

>      gen_set_access_type(ctx, ACCESS_INT);                                     \

> +    avr = tcg_temp_new_i64();                                                 \

>      EA = tcg_temp_new();                                                      \

>      gen_addr_reg_index(ctx, EA);                                              \

>      tcg_gen_andi_tl(EA, EA, ~0xf);                                            \

>      /* We only need to swap high and low halves. gen_qemu_st64_i64 does       \

>         necessary 64-bit byteswap already. */                                  \

>      if (ctx->le_mode) {                                                       \

> -        gen_qemu_st64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA);                \

> +        get_avr64(avr, rD(ctx->opcode), false);                               \

> +        gen_qemu_st64_i64(ctx, avr, EA);                                      \

>          tcg_gen_addi_tl(EA, EA, 8);                                           \

> -        gen_qemu_st64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA);                \

> +        get_avr64(avr, rD(ctx->opcode), true);                                \

> +        gen_qemu_st64_i64(ctx, avr, EA);                                      \

>      } else {                                                                  \

> -        gen_qemu_st64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA);                \

> +        get_avr64(avr, rD(ctx->opcode), true);                                \

> +        gen_qemu_st64_i64(ctx, avr, EA);                                      \

>          tcg_gen_addi_tl(EA, EA, 8);                                           \

> -        gen_qemu_st64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA);                \

> +        get_avr64(avr, rD(ctx->opcode), false);                               \

> +        gen_qemu_st64_i64(ctx, avr, EA);                                      \

>      }                                                                         \

>      tcg_temp_free(EA);                                                        \

> +    tcg_temp_free_i64(avr);                                                   \

>  }

>  

>  #define GEN_VR_LVE(name, opc2, opc3, size)                              \

> @@ -159,15 +173,20 @@ static void gen_lvsr(DisasContext *ctx)

>  static void gen_mfvscr(DisasContext *ctx)

>  {

>      TCGv_i32 t;

> +    TCGv_i64 avr;

>      if (unlikely(!ctx->altivec_enabled)) {

>          gen_exception(ctx, POWERPC_EXCP_VPU);

>          return;

>      }

> -    tcg_gen_movi_i64(cpu_avrh[rD(ctx->opcode)], 0);

> +    avr = tcg_temp_new_i64();

> +    tcg_gen_movi_i64(avr, 0);

> +    set_avr64(rD(ctx->opcode), avr, true);

>      t = tcg_temp_new_i32();

>      tcg_gen_ld_i32(t, cpu_env, offsetof(CPUPPCState, vscr));

> -    tcg_gen_extu_i32_i64(cpu_avrl[rD(ctx->opcode)], t);

> +    tcg_gen_extu_i32_i64(avr, t);

> +    set_avr64(rD(ctx->opcode), avr, false);

>      tcg_temp_free_i32(t);

> +    tcg_temp_free_i64(avr);

>  }

>  

>  static void gen_mtvscr(DisasContext *ctx)

> @@ -188,6 +207,7 @@ static void glue(gen_, name)(DisasContext *ctx)                         \

>      TCGv_i64 t0 = tcg_temp_new_i64();                                   \

>      TCGv_i64 t1 = tcg_temp_new_i64();                                   \

>      TCGv_i64 t2 = tcg_temp_new_i64();                                   \

> +    TCGv_i64 avr = tcg_temp_new_i64();                                  \

>      TCGv_i64 ten, z;                                                    \

>                                                                          \

>      if (unlikely(!ctx->altivec_enabled)) {                              \

> @@ -199,26 +219,35 @@ static void glue(gen_, name)(DisasContext *ctx)                         \

>      z = tcg_const_i64(0);                                               \

>                                                                          \

>      if (add_cin) {                                                      \

> -        tcg_gen_mulu2_i64(t0, t1, cpu_avrl[rA(ctx->opcode)], ten);      \

> -        tcg_gen_andi_i64(t2, cpu_avrl[rB(ctx->opcode)], 0xF);           \

> -        tcg_gen_add2_i64(cpu_avrl[rD(ctx->opcode)], t2, t0, t1, t2, z); \

> +        get_avr64(avr, rA(ctx->opcode), false);                         \

> +        tcg_gen_mulu2_i64(t0, t1, avr, ten);                            \

> +        get_avr64(avr, rB(ctx->opcode), false);                         \

> +        tcg_gen_andi_i64(t2, avr, 0xF);                                 \

> +        tcg_gen_add2_i64(avr, t2, t0, t1, t2, z);                       \

> +        set_avr64(rD(ctx->opcode), avr, false);                         \

>      } else {                                                            \

> -        tcg_gen_mulu2_i64(cpu_avrl[rD(ctx->opcode)], t2,                \

> -                          cpu_avrl[rA(ctx->opcode)], ten);              \

> +        get_avr64(avr, rA(ctx->opcode), false);                         \

> +        tcg_gen_mulu2_i64(avr, t2, avr, ten);                           \

> +        set_avr64(rD(ctx->opcode), avr, false);                         \

>      }                                                                   \

>                                                                          \

>      if (ret_carry) {                                                    \

> -        tcg_gen_mulu2_i64(t0, t1, cpu_avrh[rA(ctx->opcode)], ten);      \

> -        tcg_gen_add2_i64(t0, cpu_avrl[rD(ctx->opcode)], t0, t1, t2, z); \

> -        tcg_gen_movi_i64(cpu_avrh[rD(ctx->opcode)], 0);                 \

> +        get_avr64(avr, rA(ctx->opcode), true);                          \

> +        tcg_gen_mulu2_i64(t0, t1, avr, ten);                            \

> +        tcg_gen_add2_i64(t0, avr, t0, t1, t2, z);                       \

> +        set_avr64(rD(ctx->opcode), avr, false);                         \

> +        set_avr64(rD(ctx->opcode), z, true);                            \

>      } else {                                                            \

> -        tcg_gen_mul_i64(t0, cpu_avrh[rA(ctx->opcode)], ten);            \

> -        tcg_gen_add_i64(cpu_avrh[rD(ctx->opcode)], t0, t2);             \

> +        get_avr64(avr, rA(ctx->opcode), true);                          \

> +        tcg_gen_mul_i64(t0, avr, ten);                                  \

> +        tcg_gen_add_i64(avr, t0, t2);                                   \

> +        set_avr64(rD(ctx->opcode), avr, true);                          \

>      }                                                                   \

>                                                                          \

>      tcg_temp_free_i64(t0);                                              \

>      tcg_temp_free_i64(t1);                                              \

>      tcg_temp_free_i64(t2);                                              \

> +    tcg_temp_free_i64(avr);                                             \

>      tcg_temp_free_i64(ten);                                             \

>      tcg_temp_free_i64(z);                                               \

>  }                                                                       \

> @@ -232,12 +261,27 @@ GEN_VX_VMUL10(vmul10ecuq, 1, 1);

>  #define GEN_VX_LOGICAL(name, tcg_op, opc2, opc3)                        \

>  static void glue(gen_, name)(DisasContext *ctx)                                 \

>  {                                                                       \

> +    TCGv_i64 t0 = tcg_temp_new_i64();                                   \

> +    TCGv_i64 t1 = tcg_temp_new_i64();                                   \

> +    TCGv_i64 avr = tcg_temp_new_i64();                                  \

> +                                                                        \

>      if (unlikely(!ctx->altivec_enabled)) {                              \

>          gen_exception(ctx, POWERPC_EXCP_VPU);                           \

>          return;                                                         \

>      }                                                                   \

> -    tcg_op(cpu_avrh[rD(ctx->opcode)], cpu_avrh[rA(ctx->opcode)], cpu_avrh[rB(ctx->opcode)]); \

> -    tcg_op(cpu_avrl[rD(ctx->opcode)], cpu_avrl[rA(ctx->opcode)], cpu_avrl[rB(ctx->opcode)]); \

> +    get_avr64(t0, rA(ctx->opcode), true);                               \

> +    get_avr64(t1, rB(ctx->opcode), true);                               \

> +    tcg_op(avr, t0, t1);                                                \

> +    set_avr64(rD(ctx->opcode), avr, true);                              \

> +                                                                        \

> +    get_avr64(t0, rA(ctx->opcode), false);                              \

> +    get_avr64(t1, rB(ctx->opcode), false);                              \

> +    tcg_op(avr, t0, t1);                                                \

> +    set_avr64(rD(ctx->opcode), avr, false);                             \

> +                                                                        \

> +    tcg_temp_free_i64(t0);                                              \

> +    tcg_temp_free_i64(t1);                                              \

> +    tcg_temp_free_i64(avr);                                             \

>  }

>  

>  GEN_VX_LOGICAL(vand, tcg_gen_and_i64, 2, 16);

> @@ -406,6 +450,7 @@ GEN_VXFORM(vmrglw, 6, 6);

>  static void gen_vmrgew(DisasContext *ctx)

>  {

>      TCGv_i64 tmp;

> +    TCGv_i64 avr;

>      int VT, VA, VB;

>      if (unlikely(!ctx->altivec_enabled)) {

>          gen_exception(ctx, POWERPC_EXCP_VPU);

> @@ -415,15 +460,28 @@ static void gen_vmrgew(DisasContext *ctx)

>      VA = rA(ctx->opcode);

>      VB = rB(ctx->opcode);

>      tmp = tcg_temp_new_i64();

> -    tcg_gen_shri_i64(tmp, cpu_avrh[VB], 32);

> -    tcg_gen_deposit_i64(cpu_avrh[VT], cpu_avrh[VA], tmp, 0, 32);

> -    tcg_gen_shri_i64(tmp, cpu_avrl[VB], 32);

> -    tcg_gen_deposit_i64(cpu_avrl[VT], cpu_avrl[VA], tmp, 0, 32);

> +    avr = tcg_temp_new_i64();

> +

> +    get_avr64(avr, VB, true);

> +    tcg_gen_shri_i64(tmp, avr, 32);

> +    get_avr64(avr, VA, true);

> +    tcg_gen_deposit_i64(avr, avr, tmp, 0, 32);

> +    set_avr64(VT, avr, true);

> +

> +    get_avr64(avr, VB, false);

> +    tcg_gen_shri_i64(tmp, avr, 32);

> +    get_avr64(avr, VA, false);

> +    tcg_gen_deposit_i64(avr, avr, tmp, 0, 32);

> +    set_avr64(VT, avr, false);

> +

>      tcg_temp_free_i64(tmp);

> +    tcg_temp_free_i64(avr);

>  }

>  

>  static void gen_vmrgow(DisasContext *ctx)

>  {

> +    TCGv_i64 t0, t1;

> +    TCGv_i64 avr;

>      int VT, VA, VB;

>      if (unlikely(!ctx->altivec_enabled)) {

>          gen_exception(ctx, POWERPC_EXCP_VPU);

> @@ -432,9 +490,23 @@ static void gen_vmrgow(DisasContext *ctx)

>      VT = rD(ctx->opcode);

>      VA = rA(ctx->opcode);

>      VB = rB(ctx->opcode);

> +    t0 = tcg_temp_new_i64();

> +    t1 = tcg_temp_new_i64();

> +    avr = tcg_temp_new_i64();

>  

> -    tcg_gen_deposit_i64(cpu_avrh[VT], cpu_avrh[VB], cpu_avrh[VA], 32, 32);

> -    tcg_gen_deposit_i64(cpu_avrl[VT], cpu_avrl[VB], cpu_avrl[VA], 32, 32);

> +    get_avr64(t0, VB, true);

> +    get_avr64(t1, VA, true);

> +    tcg_gen_deposit_i64(avr, t0, t1, 32, 32);

> +    set_avr64(VT, avr, true);

> +

> +    get_avr64(t0, VB, false);

> +    get_avr64(t1, VA, false);

> +    tcg_gen_deposit_i64(avr, t0, t1, 32, 32);

> +    set_avr64(VT, avr, false);

> +

> +    tcg_temp_free_i64(t0);

> +    tcg_temp_free_i64(t1);

> +    tcg_temp_free_i64(avr);

>  }

>  

>  GEN_VXFORM(vmuloub, 4, 0);


-- 
David Gibson			| I'll have my music baroque, and my code
david AT gibson.dropbear.id.au	| minimalist, thank you.  NOT _the_ _other_
				| _way_ _around_!
http://www.ozlabs.org/~dgibson
diff mbox series

Patch

diff --git a/target/ppc/translate.c b/target/ppc/translate.c
index 1d4bf624a3..fa3e8dc114 100644
--- a/target/ppc/translate.c
+++ b/target/ppc/translate.c
@@ -6704,6 +6704,16 @@  static inline void set_fpr(int regno, TCGv_i64 src)
     tcg_gen_mov_i64(cpu_fpr[regno], src);
 }
 
+static inline void get_avr64(TCGv_i64 dst, int regno, bool high)
+{
+    tcg_gen_mov_i64(dst, (high ? cpu_avrh : cpu_avrl)[regno]);
+}
+
+static inline void set_avr64(int regno, TCGv_i64 src, bool high)
+{
+    tcg_gen_mov_i64((high ? cpu_avrh : cpu_avrl)[regno], src);
+}
+
 #include "translate/fp-impl.inc.c"
 
 #include "translate/vmx-impl.inc.c"
diff --git a/target/ppc/translate/vmx-impl.inc.c b/target/ppc/translate/vmx-impl.inc.c
index 3cb6fc2926..30046c6e31 100644
--- a/target/ppc/translate/vmx-impl.inc.c
+++ b/target/ppc/translate/vmx-impl.inc.c
@@ -18,52 +18,66 @@  static inline TCGv_ptr gen_avr_ptr(int reg)
 static void glue(gen_, name)(DisasContext *ctx)                                       \
 {                                                                             \
     TCGv EA;                                                                  \
+    TCGv_i64 avr;                                                             \
     if (unlikely(!ctx->altivec_enabled)) {                                    \
         gen_exception(ctx, POWERPC_EXCP_VPU);                                 \
         return;                                                               \
     }                                                                         \
     gen_set_access_type(ctx, ACCESS_INT);                                     \
+    avr = tcg_temp_new_i64();                                                 \
     EA = tcg_temp_new();                                                      \
     gen_addr_reg_index(ctx, EA);                                              \
     tcg_gen_andi_tl(EA, EA, ~0xf);                                            \
     /* We only need to swap high and low halves. gen_qemu_ld64_i64 does       \
        necessary 64-bit byteswap already. */                                  \
     if (ctx->le_mode) {                                                       \
-        gen_qemu_ld64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA);                \
+        gen_qemu_ld64_i64(ctx, avr, EA);                                      \
+        set_avr64(rD(ctx->opcode), avr, false);                               \
         tcg_gen_addi_tl(EA, EA, 8);                                           \
-        gen_qemu_ld64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA);                \
+        gen_qemu_ld64_i64(ctx, avr, EA);                                      \
+        set_avr64(rD(ctx->opcode), avr, true);                                \
     } else {                                                                  \
-        gen_qemu_ld64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA);                \
+        gen_qemu_ld64_i64(ctx, avr, EA);                                      \
+        set_avr64(rD(ctx->opcode), avr, true);                                \
         tcg_gen_addi_tl(EA, EA, 8);                                           \
-        gen_qemu_ld64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA);                \
+        gen_qemu_ld64_i64(ctx, avr, EA);                                      \
+        set_avr64(rD(ctx->opcode), avr, false);                               \
     }                                                                         \
     tcg_temp_free(EA);                                                        \
+    tcg_temp_free_i64(avr);                                                   \
 }
 
 #define GEN_VR_STX(name, opc2, opc3)                                          \
 static void gen_st##name(DisasContext *ctx)                                   \
 {                                                                             \
     TCGv EA;                                                                  \
+    TCGv_i64 avr;                                                             \
     if (unlikely(!ctx->altivec_enabled)) {                                    \
         gen_exception(ctx, POWERPC_EXCP_VPU);                                 \
         return;                                                               \
     }                                                                         \
     gen_set_access_type(ctx, ACCESS_INT);                                     \
+    avr = tcg_temp_new_i64();                                                 \
     EA = tcg_temp_new();                                                      \
     gen_addr_reg_index(ctx, EA);                                              \
     tcg_gen_andi_tl(EA, EA, ~0xf);                                            \
     /* We only need to swap high and low halves. gen_qemu_st64_i64 does       \
        necessary 64-bit byteswap already. */                                  \
     if (ctx->le_mode) {                                                       \
-        gen_qemu_st64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA);                \
+        get_avr64(avr, rD(ctx->opcode), false);                               \
+        gen_qemu_st64_i64(ctx, avr, EA);                                      \
         tcg_gen_addi_tl(EA, EA, 8);                                           \
-        gen_qemu_st64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA);                \
+        get_avr64(avr, rD(ctx->opcode), true);                                \
+        gen_qemu_st64_i64(ctx, avr, EA);                                      \
     } else {                                                                  \
-        gen_qemu_st64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA);                \
+        get_avr64(avr, rD(ctx->opcode), true);                                \
+        gen_qemu_st64_i64(ctx, avr, EA);                                      \
         tcg_gen_addi_tl(EA, EA, 8);                                           \
-        gen_qemu_st64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA);                \
+        get_avr64(avr, rD(ctx->opcode), false);                               \
+        gen_qemu_st64_i64(ctx, avr, EA);                                      \
     }                                                                         \
     tcg_temp_free(EA);                                                        \
+    tcg_temp_free_i64(avr);                                                   \
 }
 
 #define GEN_VR_LVE(name, opc2, opc3, size)                              \
@@ -159,15 +173,20 @@  static void gen_lvsr(DisasContext *ctx)
 static void gen_mfvscr(DisasContext *ctx)
 {
     TCGv_i32 t;
+    TCGv_i64 avr;
     if (unlikely(!ctx->altivec_enabled)) {
         gen_exception(ctx, POWERPC_EXCP_VPU);
         return;
     }
-    tcg_gen_movi_i64(cpu_avrh[rD(ctx->opcode)], 0);
+    avr = tcg_temp_new_i64();
+    tcg_gen_movi_i64(avr, 0);
+    set_avr64(rD(ctx->opcode), avr, true);
     t = tcg_temp_new_i32();
     tcg_gen_ld_i32(t, cpu_env, offsetof(CPUPPCState, vscr));
-    tcg_gen_extu_i32_i64(cpu_avrl[rD(ctx->opcode)], t);
+    tcg_gen_extu_i32_i64(avr, t);
+    set_avr64(rD(ctx->opcode), avr, false);
     tcg_temp_free_i32(t);
+    tcg_temp_free_i64(avr);
 }
 
 static void gen_mtvscr(DisasContext *ctx)
@@ -188,6 +207,7 @@  static void glue(gen_, name)(DisasContext *ctx)                         \
     TCGv_i64 t0 = tcg_temp_new_i64();                                   \
     TCGv_i64 t1 = tcg_temp_new_i64();                                   \
     TCGv_i64 t2 = tcg_temp_new_i64();                                   \
+    TCGv_i64 avr = tcg_temp_new_i64();                                  \
     TCGv_i64 ten, z;                                                    \
                                                                         \
     if (unlikely(!ctx->altivec_enabled)) {                              \
@@ -199,26 +219,35 @@  static void glue(gen_, name)(DisasContext *ctx)                         \
     z = tcg_const_i64(0);                                               \
                                                                         \
     if (add_cin) {                                                      \
-        tcg_gen_mulu2_i64(t0, t1, cpu_avrl[rA(ctx->opcode)], ten);      \
-        tcg_gen_andi_i64(t2, cpu_avrl[rB(ctx->opcode)], 0xF);           \
-        tcg_gen_add2_i64(cpu_avrl[rD(ctx->opcode)], t2, t0, t1, t2, z); \
+        get_avr64(avr, rA(ctx->opcode), false);                         \
+        tcg_gen_mulu2_i64(t0, t1, avr, ten);                            \
+        get_avr64(avr, rB(ctx->opcode), false);                         \
+        tcg_gen_andi_i64(t2, avr, 0xF);                                 \
+        tcg_gen_add2_i64(avr, t2, t0, t1, t2, z);                       \
+        set_avr64(rD(ctx->opcode), avr, false);                         \
     } else {                                                            \
-        tcg_gen_mulu2_i64(cpu_avrl[rD(ctx->opcode)], t2,                \
-                          cpu_avrl[rA(ctx->opcode)], ten);              \
+        get_avr64(avr, rA(ctx->opcode), false);                         \
+        tcg_gen_mulu2_i64(avr, t2, avr, ten);                           \
+        set_avr64(rD(ctx->opcode), avr, false);                         \
     }                                                                   \
                                                                         \
     if (ret_carry) {                                                    \
-        tcg_gen_mulu2_i64(t0, t1, cpu_avrh[rA(ctx->opcode)], ten);      \
-        tcg_gen_add2_i64(t0, cpu_avrl[rD(ctx->opcode)], t0, t1, t2, z); \
-        tcg_gen_movi_i64(cpu_avrh[rD(ctx->opcode)], 0);                 \
+        get_avr64(avr, rA(ctx->opcode), true);                          \
+        tcg_gen_mulu2_i64(t0, t1, avr, ten);                            \
+        tcg_gen_add2_i64(t0, avr, t0, t1, t2, z);                       \
+        set_avr64(rD(ctx->opcode), avr, false);                         \
+        set_avr64(rD(ctx->opcode), z, true);                            \
     } else {                                                            \
-        tcg_gen_mul_i64(t0, cpu_avrh[rA(ctx->opcode)], ten);            \
-        tcg_gen_add_i64(cpu_avrh[rD(ctx->opcode)], t0, t2);             \
+        get_avr64(avr, rA(ctx->opcode), true);                          \
+        tcg_gen_mul_i64(t0, avr, ten);                                  \
+        tcg_gen_add_i64(avr, t0, t2);                                   \
+        set_avr64(rD(ctx->opcode), avr, true);                          \
     }                                                                   \
                                                                         \
     tcg_temp_free_i64(t0);                                              \
     tcg_temp_free_i64(t1);                                              \
     tcg_temp_free_i64(t2);                                              \
+    tcg_temp_free_i64(avr);                                             \
     tcg_temp_free_i64(ten);                                             \
     tcg_temp_free_i64(z);                                               \
 }                                                                       \
@@ -232,12 +261,27 @@  GEN_VX_VMUL10(vmul10ecuq, 1, 1);
 #define GEN_VX_LOGICAL(name, tcg_op, opc2, opc3)                        \
 static void glue(gen_, name)(DisasContext *ctx)                                 \
 {                                                                       \
+    TCGv_i64 t0 = tcg_temp_new_i64();                                   \
+    TCGv_i64 t1 = tcg_temp_new_i64();                                   \
+    TCGv_i64 avr = tcg_temp_new_i64();                                  \
+                                                                        \
     if (unlikely(!ctx->altivec_enabled)) {                              \
         gen_exception(ctx, POWERPC_EXCP_VPU);                           \
         return;                                                         \
     }                                                                   \
-    tcg_op(cpu_avrh[rD(ctx->opcode)], cpu_avrh[rA(ctx->opcode)], cpu_avrh[rB(ctx->opcode)]); \
-    tcg_op(cpu_avrl[rD(ctx->opcode)], cpu_avrl[rA(ctx->opcode)], cpu_avrl[rB(ctx->opcode)]); \
+    get_avr64(t0, rA(ctx->opcode), true);                               \
+    get_avr64(t1, rB(ctx->opcode), true);                               \
+    tcg_op(avr, t0, t1);                                                \
+    set_avr64(rD(ctx->opcode), avr, true);                              \
+                                                                        \
+    get_avr64(t0, rA(ctx->opcode), false);                              \
+    get_avr64(t1, rB(ctx->opcode), false);                              \
+    tcg_op(avr, t0, t1);                                                \
+    set_avr64(rD(ctx->opcode), avr, false);                             \
+                                                                        \
+    tcg_temp_free_i64(t0);                                              \
+    tcg_temp_free_i64(t1);                                              \
+    tcg_temp_free_i64(avr);                                             \
 }
 
 GEN_VX_LOGICAL(vand, tcg_gen_and_i64, 2, 16);
@@ -406,6 +450,7 @@  GEN_VXFORM(vmrglw, 6, 6);
 static void gen_vmrgew(DisasContext *ctx)
 {
     TCGv_i64 tmp;
+    TCGv_i64 avr;
     int VT, VA, VB;
     if (unlikely(!ctx->altivec_enabled)) {
         gen_exception(ctx, POWERPC_EXCP_VPU);
@@ -415,15 +460,28 @@  static void gen_vmrgew(DisasContext *ctx)
     VA = rA(ctx->opcode);
     VB = rB(ctx->opcode);
     tmp = tcg_temp_new_i64();
-    tcg_gen_shri_i64(tmp, cpu_avrh[VB], 32);
-    tcg_gen_deposit_i64(cpu_avrh[VT], cpu_avrh[VA], tmp, 0, 32);
-    tcg_gen_shri_i64(tmp, cpu_avrl[VB], 32);
-    tcg_gen_deposit_i64(cpu_avrl[VT], cpu_avrl[VA], tmp, 0, 32);
+    avr = tcg_temp_new_i64();
+
+    get_avr64(avr, VB, true);
+    tcg_gen_shri_i64(tmp, avr, 32);
+    get_avr64(avr, VA, true);
+    tcg_gen_deposit_i64(avr, avr, tmp, 0, 32);
+    set_avr64(VT, avr, true);
+
+    get_avr64(avr, VB, false);
+    tcg_gen_shri_i64(tmp, avr, 32);
+    get_avr64(avr, VA, false);
+    tcg_gen_deposit_i64(avr, avr, tmp, 0, 32);
+    set_avr64(VT, avr, false);
+
     tcg_temp_free_i64(tmp);
+    tcg_temp_free_i64(avr);
 }
 
 static void gen_vmrgow(DisasContext *ctx)
 {
+    TCGv_i64 t0, t1;
+    TCGv_i64 avr;
     int VT, VA, VB;
     if (unlikely(!ctx->altivec_enabled)) {
         gen_exception(ctx, POWERPC_EXCP_VPU);
@@ -432,9 +490,23 @@  static void gen_vmrgow(DisasContext *ctx)
     VT = rD(ctx->opcode);
     VA = rA(ctx->opcode);
     VB = rB(ctx->opcode);
+    t0 = tcg_temp_new_i64();
+    t1 = tcg_temp_new_i64();
+    avr = tcg_temp_new_i64();
 
-    tcg_gen_deposit_i64(cpu_avrh[VT], cpu_avrh[VB], cpu_avrh[VA], 32, 32);
-    tcg_gen_deposit_i64(cpu_avrl[VT], cpu_avrl[VB], cpu_avrl[VA], 32, 32);
+    get_avr64(t0, VB, true);
+    get_avr64(t1, VA, true);
+    tcg_gen_deposit_i64(avr, t0, t1, 32, 32);
+    set_avr64(VT, avr, true);
+
+    get_avr64(t0, VB, false);
+    get_avr64(t1, VA, false);
+    tcg_gen_deposit_i64(avr, t0, t1, 32, 32);
+    set_avr64(VT, avr, false);
+
+    tcg_temp_free_i64(t0);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(avr);
 }
 
 GEN_VXFORM(vmuloub, 4, 0);