Message ID | 20170916023417.14599-5-richard.henderson@linaro.org |
---|---|
State | Superseded |
Headers | show |
Series | TCG vectorization and example conversion | expand |
Richard Henderson <richard.henderson@linaro.org> writes: > Signed-off-by: Richard Henderson <richard.henderson@linaro.org> Reviewed-by: Alex Bennée <alex.bennee@linaro.org> > --- > target/arm/translate-a64.c | 216 ++++++++++++++++++++++++++++++--------------- > 1 file changed, 143 insertions(+), 73 deletions(-) > > diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c > index a3984c9a0d..4759cc9829 100644 > --- a/target/arm/translate-a64.c > +++ b/target/arm/translate-a64.c > @@ -21,6 +21,7 @@ > #include "cpu.h" > #include "exec/exec-all.h" > #include "tcg-op.h" > +#include "tcg-op-gvec.h" > #include "qemu/log.h" > #include "arm_ldst.h" > #include "translate.h" > @@ -82,6 +83,7 @@ typedef void NeonGenTwoDoubleOPFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGv_ptr); > typedef void NeonGenOneOpFn(TCGv_i64, TCGv_i64); > typedef void CryptoTwoOpEnvFn(TCGv_ptr, TCGv_i32, TCGv_i32); > typedef void CryptoThreeOpEnvFn(TCGv_ptr, TCGv_i32, TCGv_i32, TCGv_i32); > +typedef void GVecGenTwoFn(uint32_t, uint32_t, uint32_t, uint32_t, uint32_t); > > /* initialize TCG globals. */ > void a64_translate_init(void) > @@ -537,6 +539,21 @@ static inline int vec_reg_offset(DisasContext *s, int regno, > return offs; > } > > +/* Return the offset info CPUARMState of the "whole" vector register Qn. */ > +static inline int vec_full_reg_offset(DisasContext *s, int regno) > +{ > + assert_fp_access_checked(s); > + return offsetof(CPUARMState, vfp.regs[regno * 2]); > +} > + > +/* Return the byte size of the "whole" vector register, VL / 8. */ > +static inline int vec_full_reg_size(DisasContext *s) > +{ > + /* FIXME SVE: We should put the composite ZCR_EL* value into tb->flags. > + In the meantime this is just the AdvSIMD length of 128. */ > + return 128 / 8; > +} > + > /* Return the offset into CPUARMState of a slice (from > * the least significant end) of FP register Qn (ie > * Dn, Sn, Hn or Bn). > @@ -9036,85 +9053,125 @@ static void disas_simd_three_reg_diff(DisasContext *s, uint32_t insn) > } > } > > +static void gen_bsl_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm) > +{ > + tcg_gen_xor_i64(rn, rn, rm); > + tcg_gen_and_i64(rn, rn, rd); > + tcg_gen_xor_i64(rd, rm, rn); > +} > + > +static void gen_bit_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm) > +{ > + tcg_gen_xor_i64(rn, rn, rd); > + tcg_gen_and_i64(rn, rn, rm); > + tcg_gen_xor_i64(rd, rd, rn); > +} > + > +static void gen_bif_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm) > +{ > + tcg_gen_xor_i64(rn, rn, rd); > + tcg_gen_andc_i64(rn, rn, rm); > + tcg_gen_xor_i64(rd, rd, rn); > +} > + > +static void gen_bsl_vec(TCGv_vec rd, TCGv_vec rn, TCGv_vec rm) > +{ > + tcg_gen_xor_vec(rn, rn, rm); > + tcg_gen_and_vec(rn, rn, rd); > + tcg_gen_xor_vec(rd, rm, rn); > +} > + > +static void gen_bit_vec(TCGv_vec rd, TCGv_vec rn, TCGv_vec rm) > +{ > + tcg_gen_xor_vec(rn, rn, rd); > + tcg_gen_and_vec(rn, rn, rm); > + tcg_gen_xor_vec(rd, rd, rn); > +} > + > +static void gen_bif_vec(TCGv_vec rd, TCGv_vec rn, TCGv_vec rm) > +{ > + tcg_gen_xor_vec(rn, rn, rd); > + tcg_gen_andc_vec(rn, rn, rm); > + tcg_gen_xor_vec(rd, rd, rn); > +} > + > /* Logic op (opcode == 3) subgroup of C3.6.16. */ > static void disas_simd_3same_logic(DisasContext *s, uint32_t insn) > { > + static const GVecGen3 bsl_op = { > + .fni8 = gen_bsl_i64, > + .fniv = gen_bsl_vec, > + .prefer_i64 = TCG_TARGET_REG_BITS == 64, > + .load_dest = true > + }; > + static const GVecGen3 bit_op = { > + .fni8 = gen_bit_i64, > + .fniv = gen_bit_vec, > + .prefer_i64 = TCG_TARGET_REG_BITS == 64, > + .load_dest = true > + }; > + static const GVecGen3 bif_op = { > + .fni8 = gen_bif_i64, > + .fniv = gen_bif_vec, > + .prefer_i64 = TCG_TARGET_REG_BITS == 64, > + .load_dest = true > + }; > + > int rd = extract32(insn, 0, 5); > int rn = extract32(insn, 5, 5); > int rm = extract32(insn, 16, 5); > int size = extract32(insn, 22, 2); > bool is_u = extract32(insn, 29, 1); > bool is_q = extract32(insn, 30, 1); > - TCGv_i64 tcg_op1, tcg_op2, tcg_res[2]; > - int pass; > + GVecGenTwoFn *gvec_fn; > + const GVecGen3 *gvec_op; > > if (!fp_access_check(s)) { > return; > } > > - tcg_op1 = tcg_temp_new_i64(); > - tcg_op2 = tcg_temp_new_i64(); > - tcg_res[0] = tcg_temp_new_i64(); > - tcg_res[1] = tcg_temp_new_i64(); > - > - for (pass = 0; pass < (is_q ? 2 : 1); pass++) { > - read_vec_element(s, tcg_op1, rn, pass, MO_64); > - read_vec_element(s, tcg_op2, rm, pass, MO_64); > - > - if (!is_u) { > - switch (size) { > - case 0: /* AND */ > - tcg_gen_and_i64(tcg_res[pass], tcg_op1, tcg_op2); > - break; > - case 1: /* BIC */ > - tcg_gen_andc_i64(tcg_res[pass], tcg_op1, tcg_op2); > - break; > - case 2: /* ORR */ > - tcg_gen_or_i64(tcg_res[pass], tcg_op1, tcg_op2); > - break; > - case 3: /* ORN */ > - tcg_gen_orc_i64(tcg_res[pass], tcg_op1, tcg_op2); > - break; > - } > - } else { > - if (size != 0) { > - /* B* ops need res loaded to operate on */ > - read_vec_element(s, tcg_res[pass], rd, pass, MO_64); > - } > - > - switch (size) { > - case 0: /* EOR */ > - tcg_gen_xor_i64(tcg_res[pass], tcg_op1, tcg_op2); > - break; > - case 1: /* BSL bitwise select */ > - tcg_gen_xor_i64(tcg_op1, tcg_op1, tcg_op2); > - tcg_gen_and_i64(tcg_op1, tcg_op1, tcg_res[pass]); > - tcg_gen_xor_i64(tcg_res[pass], tcg_op2, tcg_op1); > - break; > - case 2: /* BIT, bitwise insert if true */ > - tcg_gen_xor_i64(tcg_op1, tcg_op1, tcg_res[pass]); > - tcg_gen_and_i64(tcg_op1, tcg_op1, tcg_op2); > - tcg_gen_xor_i64(tcg_res[pass], tcg_res[pass], tcg_op1); > - break; > - case 3: /* BIF, bitwise insert if false */ > - tcg_gen_xor_i64(tcg_op1, tcg_op1, tcg_res[pass]); > - tcg_gen_andc_i64(tcg_op1, tcg_op1, tcg_op2); > - tcg_gen_xor_i64(tcg_res[pass], tcg_res[pass], tcg_op1); > - break; > - } > - } > - } > + switch (size + 4 * is_u) { > + case 0: /* AND */ > + gvec_fn = tcg_gen_gvec_and; > + goto do_fn; > + case 1: /* BIC */ > + gvec_fn = tcg_gen_gvec_andc; > + goto do_fn; > + case 2: /* ORR */ > + gvec_fn = tcg_gen_gvec_or; > + goto do_fn; > + case 3: /* ORN */ > + gvec_fn = tcg_gen_gvec_orc; > + goto do_fn; > + case 4: /* EOR */ > + gvec_fn = tcg_gen_gvec_xor; > + goto do_fn; > + do_fn: > + gvec_fn(vec_full_reg_offset(s, rd), > + vec_full_reg_offset(s, rn), > + vec_full_reg_offset(s, rm), > + is_q ? 16 : 8, vec_full_reg_size(s)); > + return; > + > + case 5: /* BSL bitwise select */ > + gvec_op = &bsl_op; > + goto do_op; > + case 6: /* BIT, bitwise insert if true */ > + gvec_op = &bit_op; > + goto do_op; > + case 7: /* BIF, bitwise insert if false */ > + gvec_op = &bif_op; > + goto do_op; > + do_op: > + tcg_gen_gvec_3(vec_full_reg_offset(s, rd), > + vec_full_reg_offset(s, rn), > + vec_full_reg_offset(s, rm), > + is_q ? 16 : 8, vec_full_reg_size(s), gvec_op); > + return; > > - write_vec_element(s, tcg_res[0], rd, 0, MO_64); > - if (!is_q) { > - tcg_gen_movi_i64(tcg_res[1], 0); > + default: > + g_assert_not_reached(); > } > - write_vec_element(s, tcg_res[1], rd, 1, MO_64); > - > - tcg_temp_free_i64(tcg_op1); > - tcg_temp_free_i64(tcg_op2); > - tcg_temp_free_i64(tcg_res[0]); > - tcg_temp_free_i64(tcg_res[1]); > } > > /* Helper functions for 32 bit comparisons */ > @@ -9375,6 +9432,7 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) > int rn = extract32(insn, 5, 5); > int rd = extract32(insn, 0, 5); > int pass; > + GVecGenTwoFn *gvec_op; > > switch (opcode) { > case 0x13: /* MUL, PMUL */ > @@ -9414,6 +9472,28 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) > return; > } > > + switch (opcode) { > + case 0x10: /* ADD, SUB */ > + { > + static GVecGenTwoFn * const fns[4][2] = { > + { tcg_gen_gvec_add8, tcg_gen_gvec_sub8 }, > + { tcg_gen_gvec_add16, tcg_gen_gvec_sub16 }, > + { tcg_gen_gvec_add32, tcg_gen_gvec_sub32 }, > + { tcg_gen_gvec_add64, tcg_gen_gvec_sub64 }, > + }; > + gvec_op = fns[size][u]; > + goto do_gvec; > + } > + break; > + > + do_gvec: > + gvec_op(vec_full_reg_offset(s, rd), > + vec_full_reg_offset(s, rn), > + vec_full_reg_offset(s, rm), > + is_q ? 16 : 8, vec_full_reg_size(s)); > + return; > + } > + > if (size == 3) { > assert(is_q); > for (pass = 0; pass < 2; pass++) { > @@ -9586,16 +9666,6 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) > genfn = fns[size][u]; > break; > } > - case 0x10: /* ADD, SUB */ > - { > - static NeonGenTwoOpFn * const fns[3][2] = { > - { gen_helper_neon_add_u8, gen_helper_neon_sub_u8 }, > - { gen_helper_neon_add_u16, gen_helper_neon_sub_u16 }, > - { tcg_gen_add_i32, tcg_gen_sub_i32 }, > - }; > - genfn = fns[size][u]; > - break; > - } > case 0x11: /* CMTST, CMEQ */ > { > static NeonGenTwoOpFn * const fns[3][2] = { -- Alex Bennée
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c index a3984c9a0d..4759cc9829 100644 --- a/target/arm/translate-a64.c +++ b/target/arm/translate-a64.c @@ -21,6 +21,7 @@ #include "cpu.h" #include "exec/exec-all.h" #include "tcg-op.h" +#include "tcg-op-gvec.h" #include "qemu/log.h" #include "arm_ldst.h" #include "translate.h" @@ -82,6 +83,7 @@ typedef void NeonGenTwoDoubleOPFn(TCGv_i64, TCGv_i64, TCGv_i64, TCGv_ptr); typedef void NeonGenOneOpFn(TCGv_i64, TCGv_i64); typedef void CryptoTwoOpEnvFn(TCGv_ptr, TCGv_i32, TCGv_i32); typedef void CryptoThreeOpEnvFn(TCGv_ptr, TCGv_i32, TCGv_i32, TCGv_i32); +typedef void GVecGenTwoFn(uint32_t, uint32_t, uint32_t, uint32_t, uint32_t); /* initialize TCG globals. */ void a64_translate_init(void) @@ -537,6 +539,21 @@ static inline int vec_reg_offset(DisasContext *s, int regno, return offs; } +/* Return the offset info CPUARMState of the "whole" vector register Qn. */ +static inline int vec_full_reg_offset(DisasContext *s, int regno) +{ + assert_fp_access_checked(s); + return offsetof(CPUARMState, vfp.regs[regno * 2]); +} + +/* Return the byte size of the "whole" vector register, VL / 8. */ +static inline int vec_full_reg_size(DisasContext *s) +{ + /* FIXME SVE: We should put the composite ZCR_EL* value into tb->flags. + In the meantime this is just the AdvSIMD length of 128. */ + return 128 / 8; +} + /* Return the offset into CPUARMState of a slice (from * the least significant end) of FP register Qn (ie * Dn, Sn, Hn or Bn). @@ -9036,85 +9053,125 @@ static void disas_simd_three_reg_diff(DisasContext *s, uint32_t insn) } } +static void gen_bsl_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm) +{ + tcg_gen_xor_i64(rn, rn, rm); + tcg_gen_and_i64(rn, rn, rd); + tcg_gen_xor_i64(rd, rm, rn); +} + +static void gen_bit_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm) +{ + tcg_gen_xor_i64(rn, rn, rd); + tcg_gen_and_i64(rn, rn, rm); + tcg_gen_xor_i64(rd, rd, rn); +} + +static void gen_bif_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm) +{ + tcg_gen_xor_i64(rn, rn, rd); + tcg_gen_andc_i64(rn, rn, rm); + tcg_gen_xor_i64(rd, rd, rn); +} + +static void gen_bsl_vec(TCGv_vec rd, TCGv_vec rn, TCGv_vec rm) +{ + tcg_gen_xor_vec(rn, rn, rm); + tcg_gen_and_vec(rn, rn, rd); + tcg_gen_xor_vec(rd, rm, rn); +} + +static void gen_bit_vec(TCGv_vec rd, TCGv_vec rn, TCGv_vec rm) +{ + tcg_gen_xor_vec(rn, rn, rd); + tcg_gen_and_vec(rn, rn, rm); + tcg_gen_xor_vec(rd, rd, rn); +} + +static void gen_bif_vec(TCGv_vec rd, TCGv_vec rn, TCGv_vec rm) +{ + tcg_gen_xor_vec(rn, rn, rd); + tcg_gen_andc_vec(rn, rn, rm); + tcg_gen_xor_vec(rd, rd, rn); +} + /* Logic op (opcode == 3) subgroup of C3.6.16. */ static void disas_simd_3same_logic(DisasContext *s, uint32_t insn) { + static const GVecGen3 bsl_op = { + .fni8 = gen_bsl_i64, + .fniv = gen_bsl_vec, + .prefer_i64 = TCG_TARGET_REG_BITS == 64, + .load_dest = true + }; + static const GVecGen3 bit_op = { + .fni8 = gen_bit_i64, + .fniv = gen_bit_vec, + .prefer_i64 = TCG_TARGET_REG_BITS == 64, + .load_dest = true + }; + static const GVecGen3 bif_op = { + .fni8 = gen_bif_i64, + .fniv = gen_bif_vec, + .prefer_i64 = TCG_TARGET_REG_BITS == 64, + .load_dest = true + }; + int rd = extract32(insn, 0, 5); int rn = extract32(insn, 5, 5); int rm = extract32(insn, 16, 5); int size = extract32(insn, 22, 2); bool is_u = extract32(insn, 29, 1); bool is_q = extract32(insn, 30, 1); - TCGv_i64 tcg_op1, tcg_op2, tcg_res[2]; - int pass; + GVecGenTwoFn *gvec_fn; + const GVecGen3 *gvec_op; if (!fp_access_check(s)) { return; } - tcg_op1 = tcg_temp_new_i64(); - tcg_op2 = tcg_temp_new_i64(); - tcg_res[0] = tcg_temp_new_i64(); - tcg_res[1] = tcg_temp_new_i64(); - - for (pass = 0; pass < (is_q ? 2 : 1); pass++) { - read_vec_element(s, tcg_op1, rn, pass, MO_64); - read_vec_element(s, tcg_op2, rm, pass, MO_64); - - if (!is_u) { - switch (size) { - case 0: /* AND */ - tcg_gen_and_i64(tcg_res[pass], tcg_op1, tcg_op2); - break; - case 1: /* BIC */ - tcg_gen_andc_i64(tcg_res[pass], tcg_op1, tcg_op2); - break; - case 2: /* ORR */ - tcg_gen_or_i64(tcg_res[pass], tcg_op1, tcg_op2); - break; - case 3: /* ORN */ - tcg_gen_orc_i64(tcg_res[pass], tcg_op1, tcg_op2); - break; - } - } else { - if (size != 0) { - /* B* ops need res loaded to operate on */ - read_vec_element(s, tcg_res[pass], rd, pass, MO_64); - } - - switch (size) { - case 0: /* EOR */ - tcg_gen_xor_i64(tcg_res[pass], tcg_op1, tcg_op2); - break; - case 1: /* BSL bitwise select */ - tcg_gen_xor_i64(tcg_op1, tcg_op1, tcg_op2); - tcg_gen_and_i64(tcg_op1, tcg_op1, tcg_res[pass]); - tcg_gen_xor_i64(tcg_res[pass], tcg_op2, tcg_op1); - break; - case 2: /* BIT, bitwise insert if true */ - tcg_gen_xor_i64(tcg_op1, tcg_op1, tcg_res[pass]); - tcg_gen_and_i64(tcg_op1, tcg_op1, tcg_op2); - tcg_gen_xor_i64(tcg_res[pass], tcg_res[pass], tcg_op1); - break; - case 3: /* BIF, bitwise insert if false */ - tcg_gen_xor_i64(tcg_op1, tcg_op1, tcg_res[pass]); - tcg_gen_andc_i64(tcg_op1, tcg_op1, tcg_op2); - tcg_gen_xor_i64(tcg_res[pass], tcg_res[pass], tcg_op1); - break; - } - } - } + switch (size + 4 * is_u) { + case 0: /* AND */ + gvec_fn = tcg_gen_gvec_and; + goto do_fn; + case 1: /* BIC */ + gvec_fn = tcg_gen_gvec_andc; + goto do_fn; + case 2: /* ORR */ + gvec_fn = tcg_gen_gvec_or; + goto do_fn; + case 3: /* ORN */ + gvec_fn = tcg_gen_gvec_orc; + goto do_fn; + case 4: /* EOR */ + gvec_fn = tcg_gen_gvec_xor; + goto do_fn; + do_fn: + gvec_fn(vec_full_reg_offset(s, rd), + vec_full_reg_offset(s, rn), + vec_full_reg_offset(s, rm), + is_q ? 16 : 8, vec_full_reg_size(s)); + return; + + case 5: /* BSL bitwise select */ + gvec_op = &bsl_op; + goto do_op; + case 6: /* BIT, bitwise insert if true */ + gvec_op = &bit_op; + goto do_op; + case 7: /* BIF, bitwise insert if false */ + gvec_op = &bif_op; + goto do_op; + do_op: + tcg_gen_gvec_3(vec_full_reg_offset(s, rd), + vec_full_reg_offset(s, rn), + vec_full_reg_offset(s, rm), + is_q ? 16 : 8, vec_full_reg_size(s), gvec_op); + return; - write_vec_element(s, tcg_res[0], rd, 0, MO_64); - if (!is_q) { - tcg_gen_movi_i64(tcg_res[1], 0); + default: + g_assert_not_reached(); } - write_vec_element(s, tcg_res[1], rd, 1, MO_64); - - tcg_temp_free_i64(tcg_op1); - tcg_temp_free_i64(tcg_op2); - tcg_temp_free_i64(tcg_res[0]); - tcg_temp_free_i64(tcg_res[1]); } /* Helper functions for 32 bit comparisons */ @@ -9375,6 +9432,7 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) int rn = extract32(insn, 5, 5); int rd = extract32(insn, 0, 5); int pass; + GVecGenTwoFn *gvec_op; switch (opcode) { case 0x13: /* MUL, PMUL */ @@ -9414,6 +9472,28 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) return; } + switch (opcode) { + case 0x10: /* ADD, SUB */ + { + static GVecGenTwoFn * const fns[4][2] = { + { tcg_gen_gvec_add8, tcg_gen_gvec_sub8 }, + { tcg_gen_gvec_add16, tcg_gen_gvec_sub16 }, + { tcg_gen_gvec_add32, tcg_gen_gvec_sub32 }, + { tcg_gen_gvec_add64, tcg_gen_gvec_sub64 }, + }; + gvec_op = fns[size][u]; + goto do_gvec; + } + break; + + do_gvec: + gvec_op(vec_full_reg_offset(s, rd), + vec_full_reg_offset(s, rn), + vec_full_reg_offset(s, rm), + is_q ? 16 : 8, vec_full_reg_size(s)); + return; + } + if (size == 3) { assert(is_q); for (pass = 0; pass < 2; pass++) { @@ -9586,16 +9666,6 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) genfn = fns[size][u]; break; } - case 0x10: /* ADD, SUB */ - { - static NeonGenTwoOpFn * const fns[3][2] = { - { gen_helper_neon_add_u8, gen_helper_neon_sub_u8 }, - { gen_helper_neon_add_u16, gen_helper_neon_sub_u16 }, - { tcg_gen_add_i32, tcg_gen_sub_i32 }, - }; - genfn = fns[size][u]; - break; - } case 0x11: /* CMTST, CMEQ */ { static NeonGenTwoOpFn * const fns[3][2] = {
Signed-off-by: Richard Henderson <richard.henderson@linaro.org> --- target/arm/translate-a64.c | 216 ++++++++++++++++++++++++++++++--------------- 1 file changed, 143 insertions(+), 73 deletions(-) -- 2.13.5