Message ID | 20180126045742.5487-18-richard.henderson@linaro.org |
---|---|
State | Superseded |
Headers | show |
Series | tcg: generic vector operations | expand |
Richard Henderson <richard.henderson@linaro.org> writes: > Reviewed-by: Peter Maydell <peter.maydell@linaro.org> > Signed-off-by: Richard Henderson <richard.henderson@linaro.org> Reviewed-by: Alex Bennée <alex.bennee@linaro.org> > --- > target/arm/translate-a64.c | 154 +++++++++++++++++++++++++++++++++++++-------- > 1 file changed, 129 insertions(+), 25 deletions(-) > > diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c > index c928c4787c..64a2c2df59 100644 > --- a/target/arm/translate-a64.c > +++ b/target/arm/translate-a64.c > @@ -9753,6 +9753,66 @@ static void disas_simd_3same_float(DisasContext *s, uint32_t insn) > } > } > > +static void gen_mla8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) > +{ > + gen_helper_neon_mul_u8(a, a, b); > + gen_helper_neon_add_u8(d, d, a); > +} > + > +static void gen_mla16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) > +{ > + gen_helper_neon_mul_u16(a, a, b); > + gen_helper_neon_add_u16(d, d, a); > +} > + > +static void gen_mla32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) > +{ > + tcg_gen_mul_i32(a, a, b); > + tcg_gen_add_i32(d, d, a); > +} > + > +static void gen_mla64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b) > +{ > + tcg_gen_mul_i64(a, a, b); > + tcg_gen_add_i64(d, d, a); > +} > + > +static void gen_mla_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b) > +{ > + tcg_gen_mul_vec(vece, a, a, b); > + tcg_gen_add_vec(vece, d, d, a); > +} > + > +static void gen_mls8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) > +{ > + gen_helper_neon_mul_u8(a, a, b); > + gen_helper_neon_sub_u8(d, d, a); > +} > + > +static void gen_mls16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) > +{ > + gen_helper_neon_mul_u16(a, a, b); > + gen_helper_neon_sub_u16(d, d, a); > +} > + > +static void gen_mls32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) > +{ > + tcg_gen_mul_i32(a, a, b); > + tcg_gen_sub_i32(d, d, a); > +} > + > +static void gen_mls64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b) > +{ > + tcg_gen_mul_i64(a, a, b); > + tcg_gen_sub_i64(d, d, a); > +} > + > +static void gen_mls_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b) > +{ > + tcg_gen_mul_vec(vece, a, a, b); > + tcg_gen_sub_vec(vece, d, d, a); > +} > + > /* Integer op subgroup of C3.6.16. */ > static void disas_simd_3same_int(DisasContext *s, uint32_t insn) > { > @@ -9771,6 +9831,52 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) > .prefer_i64 = TCG_TARGET_REG_BITS == 64, > .vece = MO_64 }, > }; > + static const GVecGen3 mla_op[4] = { > + { .fni4 = gen_mla8_i32, > + .fniv = gen_mla_vec, > + .opc = INDEX_op_mul_vec, > + .load_dest = true, > + .vece = MO_8 }, > + { .fni4 = gen_mla16_i32, > + .fniv = gen_mla_vec, > + .opc = INDEX_op_mul_vec, > + .load_dest = true, > + .vece = MO_16 }, > + { .fni4 = gen_mla32_i32, > + .fniv = gen_mla_vec, > + .opc = INDEX_op_mul_vec, > + .load_dest = true, > + .vece = MO_32 }, > + { .fni8 = gen_mla64_i64, > + .fniv = gen_mla_vec, > + .opc = INDEX_op_mul_vec, > + .prefer_i64 = TCG_TARGET_REG_BITS == 64, > + .load_dest = true, > + .vece = MO_64 }, > + }; > + static const GVecGen3 mls_op[4] = { > + { .fni4 = gen_mls8_i32, > + .fniv = gen_mls_vec, > + .opc = INDEX_op_mul_vec, > + .load_dest = true, > + .vece = MO_8 }, > + { .fni4 = gen_mls16_i32, > + .fniv = gen_mls_vec, > + .opc = INDEX_op_mul_vec, > + .load_dest = true, > + .vece = MO_16 }, > + { .fni4 = gen_mls32_i32, > + .fniv = gen_mls_vec, > + .opc = INDEX_op_mul_vec, > + .load_dest = true, > + .vece = MO_32 }, > + { .fni8 = gen_mls64_i64, > + .fniv = gen_mls_vec, > + .opc = INDEX_op_mul_vec, > + .prefer_i64 = TCG_TARGET_REG_BITS == 64, > + .load_dest = true, > + .vece = MO_64 }, > + }; > > int is_q = extract32(insn, 30, 1); > int u = extract32(insn, 29, 1); > @@ -9828,6 +9934,19 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) > gen_gvec_fn3(s, is_q, rd, rn, rm, tcg_gen_gvec_add, size); > } > return; > + case 0x13: /* MUL, PMUL */ > + if (!u) { /* MUL */ > + gen_gvec_fn3(s, is_q, rd, rn, rm, tcg_gen_gvec_mul, size); > + return; > + } > + break; > + case 0x12: /* MLA, MLS */ > + if (u) { > + gen_gvec_op3(s, is_q, rd, rn, rm, &mls_op[size]); > + } else { > + gen_gvec_op3(s, is_q, rd, rn, rm, &mla_op[size]); > + } > + return; > case 0x11: > if (!u) { /* CMTST */ > gen_gvec_op3(s, is_q, rd, rn, rm, &cmtst_op[size]); > @@ -10002,23 +10121,10 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) > break; > } > case 0x13: /* MUL, PMUL */ > - if (u) { > - /* PMUL */ > - assert(size == 0); > - genfn = gen_helper_neon_mul_p8; > - break; > - } > - /* fall through : MUL */ > - case 0x12: /* MLA, MLS */ > - { > - static NeonGenTwoOpFn * const fns[3] = { > - gen_helper_neon_mul_u8, > - gen_helper_neon_mul_u16, > - tcg_gen_mul_i32, > - }; > - genfn = fns[size]; > + assert(u); /* PMUL */ > + assert(size == 0); > + genfn = gen_helper_neon_mul_p8; > break; > - } > case 0x16: /* SQDMULH, SQRDMULH */ > { > static NeonGenTwoOpEnvFn * const fns[2][2] = { > @@ -10039,18 +10145,16 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) > genfn(tcg_res, tcg_op1, tcg_op2); > } > > - if (opcode == 0xf || opcode == 0x12) { > - /* SABA, UABA, MLA, MLS: accumulating ops */ > - static NeonGenTwoOpFn * const fns[3][2] = { > - { gen_helper_neon_add_u8, gen_helper_neon_sub_u8 }, > - { gen_helper_neon_add_u16, gen_helper_neon_sub_u16 }, > - { tcg_gen_add_i32, tcg_gen_sub_i32 }, > + if (opcode == 0xf) { > + /* SABA, UABA: accumulating ops */ > + static NeonGenTwoOpFn * const fns[3] = { > + gen_helper_neon_add_u8, > + gen_helper_neon_add_u16, > + tcg_gen_add_i32, > }; > - bool is_sub = (opcode == 0x12 && u); /* MLS */ > > - genfn = fns[size][is_sub]; > read_vec_element_i32(s, tcg_op1, rd, pass, MO_32); > - genfn(tcg_res, tcg_op1, tcg_res); > + fns[size](tcg_res, tcg_op1, tcg_res); > } > > write_vec_element_i32(s, tcg_res, rd, pass, MO_32); -- Alex Bennée
diff --git a/target/arm/translate-a64.c b/target/arm/translate-a64.c index c928c4787c..64a2c2df59 100644 --- a/target/arm/translate-a64.c +++ b/target/arm/translate-a64.c @@ -9753,6 +9753,66 @@ static void disas_simd_3same_float(DisasContext *s, uint32_t insn) } } +static void gen_mla8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) +{ + gen_helper_neon_mul_u8(a, a, b); + gen_helper_neon_add_u8(d, d, a); +} + +static void gen_mla16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) +{ + gen_helper_neon_mul_u16(a, a, b); + gen_helper_neon_add_u16(d, d, a); +} + +static void gen_mla32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) +{ + tcg_gen_mul_i32(a, a, b); + tcg_gen_add_i32(d, d, a); +} + +static void gen_mla64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b) +{ + tcg_gen_mul_i64(a, a, b); + tcg_gen_add_i64(d, d, a); +} + +static void gen_mla_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b) +{ + tcg_gen_mul_vec(vece, a, a, b); + tcg_gen_add_vec(vece, d, d, a); +} + +static void gen_mls8_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) +{ + gen_helper_neon_mul_u8(a, a, b); + gen_helper_neon_sub_u8(d, d, a); +} + +static void gen_mls16_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) +{ + gen_helper_neon_mul_u16(a, a, b); + gen_helper_neon_sub_u16(d, d, a); +} + +static void gen_mls32_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) +{ + tcg_gen_mul_i32(a, a, b); + tcg_gen_sub_i32(d, d, a); +} + +static void gen_mls64_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b) +{ + tcg_gen_mul_i64(a, a, b); + tcg_gen_sub_i64(d, d, a); +} + +static void gen_mls_vec(unsigned vece, TCGv_vec d, TCGv_vec a, TCGv_vec b) +{ + tcg_gen_mul_vec(vece, a, a, b); + tcg_gen_sub_vec(vece, d, d, a); +} + /* Integer op subgroup of C3.6.16. */ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) { @@ -9771,6 +9831,52 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) .prefer_i64 = TCG_TARGET_REG_BITS == 64, .vece = MO_64 }, }; + static const GVecGen3 mla_op[4] = { + { .fni4 = gen_mla8_i32, + .fniv = gen_mla_vec, + .opc = INDEX_op_mul_vec, + .load_dest = true, + .vece = MO_8 }, + { .fni4 = gen_mla16_i32, + .fniv = gen_mla_vec, + .opc = INDEX_op_mul_vec, + .load_dest = true, + .vece = MO_16 }, + { .fni4 = gen_mla32_i32, + .fniv = gen_mla_vec, + .opc = INDEX_op_mul_vec, + .load_dest = true, + .vece = MO_32 }, + { .fni8 = gen_mla64_i64, + .fniv = gen_mla_vec, + .opc = INDEX_op_mul_vec, + .prefer_i64 = TCG_TARGET_REG_BITS == 64, + .load_dest = true, + .vece = MO_64 }, + }; + static const GVecGen3 mls_op[4] = { + { .fni4 = gen_mls8_i32, + .fniv = gen_mls_vec, + .opc = INDEX_op_mul_vec, + .load_dest = true, + .vece = MO_8 }, + { .fni4 = gen_mls16_i32, + .fniv = gen_mls_vec, + .opc = INDEX_op_mul_vec, + .load_dest = true, + .vece = MO_16 }, + { .fni4 = gen_mls32_i32, + .fniv = gen_mls_vec, + .opc = INDEX_op_mul_vec, + .load_dest = true, + .vece = MO_32 }, + { .fni8 = gen_mls64_i64, + .fniv = gen_mls_vec, + .opc = INDEX_op_mul_vec, + .prefer_i64 = TCG_TARGET_REG_BITS == 64, + .load_dest = true, + .vece = MO_64 }, + }; int is_q = extract32(insn, 30, 1); int u = extract32(insn, 29, 1); @@ -9828,6 +9934,19 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) gen_gvec_fn3(s, is_q, rd, rn, rm, tcg_gen_gvec_add, size); } return; + case 0x13: /* MUL, PMUL */ + if (!u) { /* MUL */ + gen_gvec_fn3(s, is_q, rd, rn, rm, tcg_gen_gvec_mul, size); + return; + } + break; + case 0x12: /* MLA, MLS */ + if (u) { + gen_gvec_op3(s, is_q, rd, rn, rm, &mls_op[size]); + } else { + gen_gvec_op3(s, is_q, rd, rn, rm, &mla_op[size]); + } + return; case 0x11: if (!u) { /* CMTST */ gen_gvec_op3(s, is_q, rd, rn, rm, &cmtst_op[size]); @@ -10002,23 +10121,10 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) break; } case 0x13: /* MUL, PMUL */ - if (u) { - /* PMUL */ - assert(size == 0); - genfn = gen_helper_neon_mul_p8; - break; - } - /* fall through : MUL */ - case 0x12: /* MLA, MLS */ - { - static NeonGenTwoOpFn * const fns[3] = { - gen_helper_neon_mul_u8, - gen_helper_neon_mul_u16, - tcg_gen_mul_i32, - }; - genfn = fns[size]; + assert(u); /* PMUL */ + assert(size == 0); + genfn = gen_helper_neon_mul_p8; break; - } case 0x16: /* SQDMULH, SQRDMULH */ { static NeonGenTwoOpEnvFn * const fns[2][2] = { @@ -10039,18 +10145,16 @@ static void disas_simd_3same_int(DisasContext *s, uint32_t insn) genfn(tcg_res, tcg_op1, tcg_op2); } - if (opcode == 0xf || opcode == 0x12) { - /* SABA, UABA, MLA, MLS: accumulating ops */ - static NeonGenTwoOpFn * const fns[3][2] = { - { gen_helper_neon_add_u8, gen_helper_neon_sub_u8 }, - { gen_helper_neon_add_u16, gen_helper_neon_sub_u16 }, - { tcg_gen_add_i32, tcg_gen_sub_i32 }, + if (opcode == 0xf) { + /* SABA, UABA: accumulating ops */ + static NeonGenTwoOpFn * const fns[3] = { + gen_helper_neon_add_u8, + gen_helper_neon_add_u16, + tcg_gen_add_i32, }; - bool is_sub = (opcode == 0x12 && u); /* MLS */ - genfn = fns[size][is_sub]; read_vec_element_i32(s, tcg_op1, rd, pass, MO_32); - genfn(tcg_res, tcg_op1, tcg_res); + fns[size](tcg_res, tcg_op1, tcg_res); } write_vec_element_i32(s, tcg_res, rd, pass, MO_32);