Loading target/arm/translate-vfp.inc.c +48 −0 Original line number Diff line number Diff line Loading @@ -2320,3 +2320,51 @@ static bool trans_VRINTX_dp(DisasContext *s, arg_VRINTX_dp *a) tcg_temp_free_i64(tmp); return true; } static bool trans_VCVT_sp(DisasContext *s, arg_VCVT_sp *a) { TCGv_i64 vd; TCGv_i32 vm; /* UNDEF accesses to D16-D31 if they don't exist. */ if (!dc_isar_feature(aa32_fp_d32, s) && (a->vd & 0x10)) { return false; } if (!vfp_access_check(s)) { return true; } vm = tcg_temp_new_i32(); vd = tcg_temp_new_i64(); neon_load_reg32(vm, a->vm); gen_helper_vfp_fcvtds(vd, vm, cpu_env); neon_store_reg64(vd, a->vd); tcg_temp_free_i32(vm); tcg_temp_free_i64(vd); return true; } static bool trans_VCVT_dp(DisasContext *s, arg_VCVT_dp *a) { TCGv_i64 vm; TCGv_i32 vd; /* UNDEF accesses to D16-D31 if they don't exist. */ if (!dc_isar_feature(aa32_fp_d32, s) && (a->vm & 0x10)) { return false; } if (!vfp_access_check(s)) { return true; } vd = tcg_temp_new_i32(); vm = tcg_temp_new_i64(); neon_load_reg64(vm, a->vm); gen_helper_vfp_fcvtsd(vd, vm, cpu_env); neon_store_reg32(vd, a->vd); tcg_temp_free_i32(vd); tcg_temp_free_i64(vm); return true; } target/arm/translate.c +1 −12 Original line number Diff line number Diff line Loading @@ -3050,7 +3050,7 @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn) return 1; case 15: switch (rn) { case 0 ... 14: case 0 ... 15: /* Already handled by decodetree */ return 1; default: Loading @@ -3063,10 +3063,6 @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn) if (op == 15) { /* rn is opcode, encoded as per VFP_SREG_N. */ switch (rn) { case 0x0f: /* vcvt double<->single */ rd_is_dp = !dp; break; case 0x10: /* vcvt.fxx.u32 */ case 0x11: /* vcvt.fxx.s32 */ rm_is_dp = false; Loading Loading @@ -3185,13 +3181,6 @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn) switch (op) { case 15: /* extension space */ switch (rn) { case 15: /* single<->double conversion */ if (dp) { gen_helper_vfp_fcvtsd(cpu_F0s, cpu_F0d, cpu_env); } else { gen_helper_vfp_fcvtds(cpu_F0d, cpu_F0s, cpu_env); } break; case 16: /* fuito */ gen_vfp_uito(dp, 0); break; Loading target/arm/vfp.decode +6 −0 Original line number Diff line number Diff line Loading @@ -208,3 +208,9 @@ VRINTX_sp ---- 1110 1.11 0111 .... 1010 01.0 .... \ vd=%vd_sp vm=%vm_sp VRINTX_dp ---- 1110 1.11 0111 .... 1011 01.0 .... \ vd=%vd_dp vm=%vm_dp # VCVT between single and double: Vm precision depends on size; Vd is its reverse VCVT_sp ---- 1110 1.11 0111 .... 1010 11.0 .... \ vd=%vd_dp vm=%vm_sp VCVT_dp ---- 1110 1.11 0111 .... 1011 11.0 .... \ vd=%vd_sp vm=%vm_dp Loading
target/arm/translate-vfp.inc.c +48 −0 Original line number Diff line number Diff line Loading @@ -2320,3 +2320,51 @@ static bool trans_VRINTX_dp(DisasContext *s, arg_VRINTX_dp *a) tcg_temp_free_i64(tmp); return true; } static bool trans_VCVT_sp(DisasContext *s, arg_VCVT_sp *a) { TCGv_i64 vd; TCGv_i32 vm; /* UNDEF accesses to D16-D31 if they don't exist. */ if (!dc_isar_feature(aa32_fp_d32, s) && (a->vd & 0x10)) { return false; } if (!vfp_access_check(s)) { return true; } vm = tcg_temp_new_i32(); vd = tcg_temp_new_i64(); neon_load_reg32(vm, a->vm); gen_helper_vfp_fcvtds(vd, vm, cpu_env); neon_store_reg64(vd, a->vd); tcg_temp_free_i32(vm); tcg_temp_free_i64(vd); return true; } static bool trans_VCVT_dp(DisasContext *s, arg_VCVT_dp *a) { TCGv_i64 vm; TCGv_i32 vd; /* UNDEF accesses to D16-D31 if they don't exist. */ if (!dc_isar_feature(aa32_fp_d32, s) && (a->vm & 0x10)) { return false; } if (!vfp_access_check(s)) { return true; } vd = tcg_temp_new_i32(); vm = tcg_temp_new_i64(); neon_load_reg64(vm, a->vm); gen_helper_vfp_fcvtsd(vd, vm, cpu_env); neon_store_reg32(vd, a->vd); tcg_temp_free_i32(vd); tcg_temp_free_i64(vm); return true; }
target/arm/translate.c +1 −12 Original line number Diff line number Diff line Loading @@ -3050,7 +3050,7 @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn) return 1; case 15: switch (rn) { case 0 ... 14: case 0 ... 15: /* Already handled by decodetree */ return 1; default: Loading @@ -3063,10 +3063,6 @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn) if (op == 15) { /* rn is opcode, encoded as per VFP_SREG_N. */ switch (rn) { case 0x0f: /* vcvt double<->single */ rd_is_dp = !dp; break; case 0x10: /* vcvt.fxx.u32 */ case 0x11: /* vcvt.fxx.s32 */ rm_is_dp = false; Loading Loading @@ -3185,13 +3181,6 @@ static int disas_vfp_insn(DisasContext *s, uint32_t insn) switch (op) { case 15: /* extension space */ switch (rn) { case 15: /* single<->double conversion */ if (dp) { gen_helper_vfp_fcvtsd(cpu_F0s, cpu_F0d, cpu_env); } else { gen_helper_vfp_fcvtds(cpu_F0d, cpu_F0s, cpu_env); } break; case 16: /* fuito */ gen_vfp_uito(dp, 0); break; Loading
target/arm/vfp.decode +6 −0 Original line number Diff line number Diff line Loading @@ -208,3 +208,9 @@ VRINTX_sp ---- 1110 1.11 0111 .... 1010 01.0 .... \ vd=%vd_sp vm=%vm_sp VRINTX_dp ---- 1110 1.11 0111 .... 1011 01.0 .... \ vd=%vd_dp vm=%vm_dp # VCVT between single and double: Vm precision depends on size; Vd is its reverse VCVT_sp ---- 1110 1.11 0111 .... 1010 11.0 .... \ vd=%vd_dp vm=%vm_sp VCVT_dp ---- 1110 1.11 0111 .... 1011 11.0 .... \ vd=%vd_sp vm=%vm_dp