Loading target-sparc/translate.c +63 −0 Original line number Diff line number Diff line Loading @@ -2036,6 +2036,8 @@ typedef enum { GET_ASI_DTWINX, GET_ASI_BLOCK, GET_ASI_SHORT, GET_ASI_BCOPY, GET_ASI_BFILL, } ASIType; typedef struct { Loading Loading @@ -2077,6 +2079,14 @@ static DisasASI get_asi(DisasContext *dc, int insn, TCGMemOp memop) mem_idx = MMU_PHYS_IDX; type = GET_ASI_DIRECT; break; case ASI_M_BCOPY: /* Block copy, sta access */ mem_idx = MMU_KERNEL_IDX; type = GET_ASI_BCOPY; break; case ASI_M_BFILL: /* Block fill, stda access */ mem_idx = MMU_KERNEL_IDX; type = GET_ASI_BFILL; break; } } else { gen_exception(dc, TT_PRIV_INSN); Loading Loading @@ -2294,6 +2304,38 @@ static void gen_st_asi(DisasContext *dc, TCGv src, TCGv addr, gen_address_mask(dc, addr); tcg_gen_qemu_st_tl(src, addr, da.mem_idx, da.memop); break; #if !defined(TARGET_SPARC64) && !defined(CONFIG_USER_ONLY) case GET_ASI_BCOPY: /* Copy 32 bytes from the address in SRC to ADDR. */ /* ??? The original qemu code suggests 4-byte alignment, dropping the low bits, but the only place I can see this used is in the Linux kernel with 32 byte alignment, which would make more sense as a cacheline-style operation. */ { TCGv saddr = tcg_temp_new(); TCGv daddr = tcg_temp_new(); TCGv four = tcg_const_tl(4); TCGv_i32 tmp = tcg_temp_new_i32(); int i; tcg_gen_andi_tl(saddr, src, -4); tcg_gen_andi_tl(daddr, addr, -4); for (i = 0; i < 32; i += 4) { /* Since the loads and stores are paired, allow the copy to happen in the host endianness. */ tcg_gen_qemu_ld_i32(tmp, saddr, da.mem_idx, MO_UL); tcg_gen_qemu_st_i32(tmp, daddr, da.mem_idx, MO_UL); tcg_gen_add_tl(saddr, saddr, four); tcg_gen_add_tl(daddr, daddr, four); } tcg_temp_free(saddr); tcg_temp_free(daddr); tcg_temp_free(four); tcg_temp_free_i32(tmp); } break; #endif default: { TCGv_i32 r_asi = tcg_const_i32(da.asi); Loading Loading @@ -2766,6 +2808,27 @@ static void gen_stda_asi(DisasContext *dc, TCGv hi, TCGv addr, gen_address_mask(dc, addr); tcg_gen_qemu_st_i64(t64, addr, da.mem_idx, da.memop); break; case GET_ASI_BFILL: /* Store 32 bytes of T64 to ADDR. */ /* ??? The original qemu code suggests 8-byte alignment, dropping the low bits, but the only place I can see this used is in the Linux kernel with 32 byte alignment, which would make more sense as a cacheline-style operation. */ { TCGv d_addr = tcg_temp_new(); TCGv eight = tcg_const_tl(8); int i; tcg_gen_andi_tl(d_addr, addr, -8); for (i = 0; i < 32; i += 8) { tcg_gen_qemu_st_i64(t64, d_addr, da.mem_idx, da.memop); tcg_gen_add_tl(d_addr, d_addr, eight); } tcg_temp_free(d_addr); tcg_temp_free(eight); } break; default: { TCGv_i32 r_asi = tcg_const_i32(da.asi); Loading Loading
target-sparc/translate.c +63 −0 Original line number Diff line number Diff line Loading @@ -2036,6 +2036,8 @@ typedef enum { GET_ASI_DTWINX, GET_ASI_BLOCK, GET_ASI_SHORT, GET_ASI_BCOPY, GET_ASI_BFILL, } ASIType; typedef struct { Loading Loading @@ -2077,6 +2079,14 @@ static DisasASI get_asi(DisasContext *dc, int insn, TCGMemOp memop) mem_idx = MMU_PHYS_IDX; type = GET_ASI_DIRECT; break; case ASI_M_BCOPY: /* Block copy, sta access */ mem_idx = MMU_KERNEL_IDX; type = GET_ASI_BCOPY; break; case ASI_M_BFILL: /* Block fill, stda access */ mem_idx = MMU_KERNEL_IDX; type = GET_ASI_BFILL; break; } } else { gen_exception(dc, TT_PRIV_INSN); Loading Loading @@ -2294,6 +2304,38 @@ static void gen_st_asi(DisasContext *dc, TCGv src, TCGv addr, gen_address_mask(dc, addr); tcg_gen_qemu_st_tl(src, addr, da.mem_idx, da.memop); break; #if !defined(TARGET_SPARC64) && !defined(CONFIG_USER_ONLY) case GET_ASI_BCOPY: /* Copy 32 bytes from the address in SRC to ADDR. */ /* ??? The original qemu code suggests 4-byte alignment, dropping the low bits, but the only place I can see this used is in the Linux kernel with 32 byte alignment, which would make more sense as a cacheline-style operation. */ { TCGv saddr = tcg_temp_new(); TCGv daddr = tcg_temp_new(); TCGv four = tcg_const_tl(4); TCGv_i32 tmp = tcg_temp_new_i32(); int i; tcg_gen_andi_tl(saddr, src, -4); tcg_gen_andi_tl(daddr, addr, -4); for (i = 0; i < 32; i += 4) { /* Since the loads and stores are paired, allow the copy to happen in the host endianness. */ tcg_gen_qemu_ld_i32(tmp, saddr, da.mem_idx, MO_UL); tcg_gen_qemu_st_i32(tmp, daddr, da.mem_idx, MO_UL); tcg_gen_add_tl(saddr, saddr, four); tcg_gen_add_tl(daddr, daddr, four); } tcg_temp_free(saddr); tcg_temp_free(daddr); tcg_temp_free(four); tcg_temp_free_i32(tmp); } break; #endif default: { TCGv_i32 r_asi = tcg_const_i32(da.asi); Loading Loading @@ -2766,6 +2808,27 @@ static void gen_stda_asi(DisasContext *dc, TCGv hi, TCGv addr, gen_address_mask(dc, addr); tcg_gen_qemu_st_i64(t64, addr, da.mem_idx, da.memop); break; case GET_ASI_BFILL: /* Store 32 bytes of T64 to ADDR. */ /* ??? The original qemu code suggests 8-byte alignment, dropping the low bits, but the only place I can see this used is in the Linux kernel with 32 byte alignment, which would make more sense as a cacheline-style operation. */ { TCGv d_addr = tcg_temp_new(); TCGv eight = tcg_const_tl(8); int i; tcg_gen_andi_tl(d_addr, addr, -8); for (i = 0; i < 32; i += 8) { tcg_gen_qemu_st_i64(t64, d_addr, da.mem_idx, da.memop); tcg_gen_add_tl(d_addr, d_addr, eight); } tcg_temp_free(d_addr); tcg_temp_free(eight); } break; default: { TCGv_i32 r_asi = tcg_const_i32(da.asi); Loading