Loading target-arm/translate.c +34 −33 Original line number Diff line number Diff line Loading @@ -3764,6 +3764,7 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) int load; int shift; int n; TCGv addr; TCGv tmp; TCGv tmp2; Loading @@ -3773,6 +3774,7 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) rn = (insn >> 16) & 0xf; rm = insn & 0xf; load = (insn & (1 << 21)) != 0; addr = new_tmp(); if ((insn & (1 << 23)) == 0) { /* Load store all elements. */ op = (insn >> 8) & 0xf; Loading @@ -3781,32 +3783,30 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) return 1; nregs = neon_ls_element_type[op].nregs; interleave = neon_ls_element_type[op].interleave; gen_movl_T1_reg(s, rn); tcg_gen_mov_i32(addr, cpu_R[rn]); stride = (1 << size) * interleave; for (reg = 0; reg < nregs; reg++) { if (interleave > 2 || (interleave == 2 && nregs == 2)) { gen_movl_T1_reg(s, rn); gen_op_addl_T1_im((1 << size) * reg); tcg_gen_addi_i32(addr, cpu_R[rn], (1 << size) * reg); } else if (interleave == 2 && nregs == 4 && reg == 2) { gen_movl_T1_reg(s, rn); gen_op_addl_T1_im(1 << size); tcg_gen_addi_i32(addr, cpu_R[rn], 1 << size); } for (pass = 0; pass < 2; pass++) { if (size == 2) { if (load) { tmp = gen_ld32(cpu_T[1], IS_USER(s)); tmp = gen_ld32(addr, IS_USER(s)); neon_store_reg(rd, pass, tmp); } else { tmp = neon_load_reg(rd, pass); gen_st32(tmp, cpu_T[1], IS_USER(s)); gen_st32(tmp, addr, IS_USER(s)); } gen_op_addl_T1_im(stride); tcg_gen_addi_i32(addr, addr, stride); } else if (size == 1) { if (load) { tmp = gen_ld16u(cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); tmp2 = gen_ld16u(cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); tmp = gen_ld16u(addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); tmp2 = gen_ld16u(addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); gen_bfi(tmp, tmp, tmp2, 16, 0xffff); dead_tmp(tmp2); neon_store_reg(rd, pass, tmp); Loading @@ -3814,17 +3814,17 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) tmp = neon_load_reg(rd, pass); tmp2 = new_tmp(); tcg_gen_shri_i32(tmp2, tmp, 16); gen_st16(tmp, cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); gen_st16(tmp2, cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); gen_st16(tmp, addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); gen_st16(tmp2, addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); } } else /* size == 0 */ { if (load) { TCGV_UNUSED(tmp2); for (n = 0; n < 4; n++) { tmp = gen_ld8u(cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); tmp = gen_ld8u(addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); if (n == 0) { tmp2 = tmp; } else { Loading @@ -3842,8 +3842,8 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) } else { tcg_gen_shri_i32(tmp, tmp2, n * 8); } gen_st8(tmp, cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); gen_st8(tmp, addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); } dead_tmp(tmp2); } Loading @@ -3861,26 +3861,26 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) size = (insn >> 6) & 3; nregs = ((insn >> 8) & 3) + 1; stride = (insn & (1 << 5)) ? 2 : 1; gen_movl_T1_reg(s, rn); tcg_gen_mov_i32(addr, cpu_R[rn]); for (reg = 0; reg < nregs; reg++) { switch (size) { case 0: tmp = gen_ld8u(cpu_T[1], IS_USER(s)); tmp = gen_ld8u(addr, IS_USER(s)); gen_neon_dup_u8(tmp, 0); break; case 1: tmp = gen_ld16u(cpu_T[1], IS_USER(s)); tmp = gen_ld16u(addr, IS_USER(s)); gen_neon_dup_low16(tmp); break; case 2: tmp = gen_ld32(cpu_T[0], IS_USER(s)); tmp = gen_ld32(addr, IS_USER(s)); break; case 3: return 1; default: /* Avoid compiler warnings. */ abort(); } gen_op_addl_T1_im(1 << size); tcg_gen_addi_i32(addr, addr, 1 << size); tmp2 = new_tmp(); tcg_gen_mov_i32(tmp2, tmp); neon_store_reg(rd, 0, tmp2); Loading Loading @@ -3908,18 +3908,18 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) abort(); } nregs = ((insn >> 8) & 3) + 1; gen_movl_T1_reg(s, rn); tcg_gen_mov_i32(addr, cpu_R[rn]); for (reg = 0; reg < nregs; reg++) { if (load) { switch (size) { case 0: tmp = gen_ld8u(cpu_T[1], IS_USER(s)); tmp = gen_ld8u(addr, IS_USER(s)); break; case 1: tmp = gen_ld16u(cpu_T[1], IS_USER(s)); tmp = gen_ld16u(addr, IS_USER(s)); break; case 2: tmp = gen_ld32(cpu_T[1], IS_USER(s)); tmp = gen_ld32(addr, IS_USER(s)); break; default: /* Avoid compiler warnings. */ abort(); Loading @@ -3936,22 +3936,23 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) tcg_gen_shri_i32(tmp, tmp, shift); switch (size) { case 0: gen_st8(tmp, cpu_T[1], IS_USER(s)); gen_st8(tmp, addr, IS_USER(s)); break; case 1: gen_st16(tmp, cpu_T[1], IS_USER(s)); gen_st16(tmp, addr, IS_USER(s)); break; case 2: gen_st32(tmp, cpu_T[1], IS_USER(s)); gen_st32(tmp, addr, IS_USER(s)); break; } } rd += stride; gen_op_addl_T1_im(1 << size); tcg_gen_addi_i32(addr, addr, 1 << size); } stride = nregs * (1 << size); } } dead_tmp(addr); if (rm != 15) { TCGv base; Loading Loading
target-arm/translate.c +34 −33 Original line number Diff line number Diff line Loading @@ -3764,6 +3764,7 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) int load; int shift; int n; TCGv addr; TCGv tmp; TCGv tmp2; Loading @@ -3773,6 +3774,7 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) rn = (insn >> 16) & 0xf; rm = insn & 0xf; load = (insn & (1 << 21)) != 0; addr = new_tmp(); if ((insn & (1 << 23)) == 0) { /* Load store all elements. */ op = (insn >> 8) & 0xf; Loading @@ -3781,32 +3783,30 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) return 1; nregs = neon_ls_element_type[op].nregs; interleave = neon_ls_element_type[op].interleave; gen_movl_T1_reg(s, rn); tcg_gen_mov_i32(addr, cpu_R[rn]); stride = (1 << size) * interleave; for (reg = 0; reg < nregs; reg++) { if (interleave > 2 || (interleave == 2 && nregs == 2)) { gen_movl_T1_reg(s, rn); gen_op_addl_T1_im((1 << size) * reg); tcg_gen_addi_i32(addr, cpu_R[rn], (1 << size) * reg); } else if (interleave == 2 && nregs == 4 && reg == 2) { gen_movl_T1_reg(s, rn); gen_op_addl_T1_im(1 << size); tcg_gen_addi_i32(addr, cpu_R[rn], 1 << size); } for (pass = 0; pass < 2; pass++) { if (size == 2) { if (load) { tmp = gen_ld32(cpu_T[1], IS_USER(s)); tmp = gen_ld32(addr, IS_USER(s)); neon_store_reg(rd, pass, tmp); } else { tmp = neon_load_reg(rd, pass); gen_st32(tmp, cpu_T[1], IS_USER(s)); gen_st32(tmp, addr, IS_USER(s)); } gen_op_addl_T1_im(stride); tcg_gen_addi_i32(addr, addr, stride); } else if (size == 1) { if (load) { tmp = gen_ld16u(cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); tmp2 = gen_ld16u(cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); tmp = gen_ld16u(addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); tmp2 = gen_ld16u(addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); gen_bfi(tmp, tmp, tmp2, 16, 0xffff); dead_tmp(tmp2); neon_store_reg(rd, pass, tmp); Loading @@ -3814,17 +3814,17 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) tmp = neon_load_reg(rd, pass); tmp2 = new_tmp(); tcg_gen_shri_i32(tmp2, tmp, 16); gen_st16(tmp, cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); gen_st16(tmp2, cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); gen_st16(tmp, addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); gen_st16(tmp2, addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); } } else /* size == 0 */ { if (load) { TCGV_UNUSED(tmp2); for (n = 0; n < 4; n++) { tmp = gen_ld8u(cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); tmp = gen_ld8u(addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); if (n == 0) { tmp2 = tmp; } else { Loading @@ -3842,8 +3842,8 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) } else { tcg_gen_shri_i32(tmp, tmp2, n * 8); } gen_st8(tmp, cpu_T[1], IS_USER(s)); gen_op_addl_T1_im(stride); gen_st8(tmp, addr, IS_USER(s)); tcg_gen_addi_i32(addr, addr, stride); } dead_tmp(tmp2); } Loading @@ -3861,26 +3861,26 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) size = (insn >> 6) & 3; nregs = ((insn >> 8) & 3) + 1; stride = (insn & (1 << 5)) ? 2 : 1; gen_movl_T1_reg(s, rn); tcg_gen_mov_i32(addr, cpu_R[rn]); for (reg = 0; reg < nregs; reg++) { switch (size) { case 0: tmp = gen_ld8u(cpu_T[1], IS_USER(s)); tmp = gen_ld8u(addr, IS_USER(s)); gen_neon_dup_u8(tmp, 0); break; case 1: tmp = gen_ld16u(cpu_T[1], IS_USER(s)); tmp = gen_ld16u(addr, IS_USER(s)); gen_neon_dup_low16(tmp); break; case 2: tmp = gen_ld32(cpu_T[0], IS_USER(s)); tmp = gen_ld32(addr, IS_USER(s)); break; case 3: return 1; default: /* Avoid compiler warnings. */ abort(); } gen_op_addl_T1_im(1 << size); tcg_gen_addi_i32(addr, addr, 1 << size); tmp2 = new_tmp(); tcg_gen_mov_i32(tmp2, tmp); neon_store_reg(rd, 0, tmp2); Loading Loading @@ -3908,18 +3908,18 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) abort(); } nregs = ((insn >> 8) & 3) + 1; gen_movl_T1_reg(s, rn); tcg_gen_mov_i32(addr, cpu_R[rn]); for (reg = 0; reg < nregs; reg++) { if (load) { switch (size) { case 0: tmp = gen_ld8u(cpu_T[1], IS_USER(s)); tmp = gen_ld8u(addr, IS_USER(s)); break; case 1: tmp = gen_ld16u(cpu_T[1], IS_USER(s)); tmp = gen_ld16u(addr, IS_USER(s)); break; case 2: tmp = gen_ld32(cpu_T[1], IS_USER(s)); tmp = gen_ld32(addr, IS_USER(s)); break; default: /* Avoid compiler warnings. */ abort(); Loading @@ -3936,22 +3936,23 @@ static int disas_neon_ls_insn(CPUState * env, DisasContext *s, uint32_t insn) tcg_gen_shri_i32(tmp, tmp, shift); switch (size) { case 0: gen_st8(tmp, cpu_T[1], IS_USER(s)); gen_st8(tmp, addr, IS_USER(s)); break; case 1: gen_st16(tmp, cpu_T[1], IS_USER(s)); gen_st16(tmp, addr, IS_USER(s)); break; case 2: gen_st32(tmp, cpu_T[1], IS_USER(s)); gen_st32(tmp, addr, IS_USER(s)); break; } } rd += stride; gen_op_addl_T1_im(1 << size); tcg_gen_addi_i32(addr, addr, 1 << size); } stride = nregs * (1 << size); } } dead_tmp(addr); if (rm != 15) { TCGv base; Loading