target/arm: Use vector infrastructure for aa64 dup/movi

Reviewed-by: Alex Bennée <alex.bennee@linaro.org>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
This commit is contained in:
Richard Henderson 2017-11-13 19:31:31 +01:00
parent 377ef731a8
commit 861a1ded24

View file

@ -5890,10 +5890,7 @@ static void handle_simd_dupe(DisasContext *s, int is_q, int rd, int rn,
int imm5) int imm5)
{ {
int size = ctz32(imm5); int size = ctz32(imm5);
int esize = 8 << size; int index = imm5 >> (size + 1);
int elements = (is_q ? 128 : 64) / esize;
int index, i;
TCGv_i64 tmp;
if (size > 3 || (size == 3 && !is_q)) { if (size > 3 || (size == 3 && !is_q)) {
unallocated_encoding(s); unallocated_encoding(s);
@ -5904,20 +5901,9 @@ static void handle_simd_dupe(DisasContext *s, int is_q, int rd, int rn,
return; return;
} }
index = imm5 >> (size + 1); tcg_gen_gvec_dup_mem(size, vec_full_reg_offset(s, rd),
vec_reg_offset(s, rn, index, size),
tmp = tcg_temp_new_i64(); is_q ? 16 : 8, vec_full_reg_size(s));
read_vec_element(s, tmp, rn, index, size);
for (i = 0; i < elements; i++) {
write_vec_element(s, tmp, rd, i, size);
}
if (!is_q) {
clear_vec_high(s, rd);
}
tcg_temp_free_i64(tmp);
} }
/* DUP (element, scalar) /* DUP (element, scalar)
@ -5966,9 +5952,7 @@ static void handle_simd_dupg(DisasContext *s, int is_q, int rd, int rn,
int imm5) int imm5)
{ {
int size = ctz32(imm5); int size = ctz32(imm5);
int esize = 8 << size; uint32_t dofs, oprsz, maxsz;
int elements = (is_q ? 128 : 64)/esize;
int i = 0;
if (size > 3 || ((size == 3) && !is_q)) { if (size > 3 || ((size == 3) && !is_q)) {
unallocated_encoding(s); unallocated_encoding(s);
@ -5979,12 +5963,11 @@ static void handle_simd_dupg(DisasContext *s, int is_q, int rd, int rn,
return; return;
} }
for (i = 0; i < elements; i++) { dofs = vec_full_reg_offset(s, rd);
write_vec_element(s, cpu_reg(s, rn), rd, i, size); oprsz = is_q ? 16 : 8;
} maxsz = vec_full_reg_size(s);
if (!is_q) {
clear_vec_high(s, rd); tcg_gen_gvec_dup_i64(size, dofs, oprsz, maxsz, cpu_reg(s, rn));
}
} }
/* INS (Element) /* INS (Element)
@ -6175,7 +6158,6 @@ static void disas_simd_mod_imm(DisasContext *s, uint32_t insn)
bool is_neg = extract32(insn, 29, 1); bool is_neg = extract32(insn, 29, 1);
bool is_q = extract32(insn, 30, 1); bool is_q = extract32(insn, 30, 1);
uint64_t imm = 0; uint64_t imm = 0;
TCGv_i64 tcg_rd, tcg_imm;
int i; int i;
if (o2 != 0 || ((cmode == 0xf) && is_neg && !is_q)) { if (o2 != 0 || ((cmode == 0xf) && is_neg && !is_q)) {
@ -6257,16 +6239,21 @@ static void disas_simd_mod_imm(DisasContext *s, uint32_t insn)
imm = ~imm; imm = ~imm;
} }
tcg_imm = tcg_const_i64(imm); if (!((cmode & 0x9) == 0x1 || (cmode & 0xd) == 0x9)) {
tcg_rd = new_tmp_a64(s); /* MOVI or MVNI, with MVNI negation handled above. */
tcg_gen_gvec_dup64i(vec_full_reg_offset(s, rd), is_q ? 16 : 8,
vec_full_reg_size(s), imm);
} else {
TCGv_i64 tcg_imm = tcg_const_i64(imm);
TCGv_i64 tcg_rd = new_tmp_a64(s);
for (i = 0; i < 2; i++) { for (i = 0; i < 2; i++) {
int foffs = i ? fp_reg_hi_offset(s, rd) : fp_reg_offset(s, rd, MO_64); int foffs = vec_reg_offset(s, rd, i, MO_64);
if (i == 1 && !is_q) { if (i == 1 && !is_q) {
/* non-quad ops clear high half of vector */ /* non-quad ops clear high half of vector */
tcg_gen_movi_i64(tcg_rd, 0); tcg_gen_movi_i64(tcg_rd, 0);
} else if ((cmode & 0x9) == 0x1 || (cmode & 0xd) == 0x9) { } else {
tcg_gen_ld_i64(tcg_rd, cpu_env, foffs); tcg_gen_ld_i64(tcg_rd, cpu_env, foffs);
if (is_neg) { if (is_neg) {
/* AND (BIC) */ /* AND (BIC) */
@ -6275,14 +6262,12 @@ static void disas_simd_mod_imm(DisasContext *s, uint32_t insn)
/* ORR */ /* ORR */
tcg_gen_or_i64(tcg_rd, tcg_rd, tcg_imm); tcg_gen_or_i64(tcg_rd, tcg_rd, tcg_imm);
} }
} else {
/* MOVI */
tcg_gen_mov_i64(tcg_rd, tcg_imm);
} }
tcg_gen_st_i64(tcg_rd, cpu_env, foffs); tcg_gen_st_i64(tcg_rd, cpu_env, foffs);
} }
tcg_temp_free_i64(tcg_imm); tcg_temp_free_i64(tcg_imm);
}
} }
/* AdvSIMD scalar copy /* AdvSIMD scalar copy