Commit eabcd6fa authored by Richard Henderson's avatar Richard Henderson Committed by Peter Maydell
Browse files

target/arm: Use gvec for NEON_3R_LOGIC insns



Move expanders for VBSL, VBIT, and VBIF from translate-a64.c.

Signed-off-by: default avatarRichard Henderson <richard.henderson@linaro.org>
Message-id: 20181011205206.3552-9-richard.henderson@linaro.org
Reviewed-by: default avatarPeter Maydell <peter.maydell@linaro.org>
Signed-off-by: default avatarPeter Maydell <peter.maydell@linaro.org>
parent 246fa4ac
Loading
Loading
Loading
Loading
+0 −61
Original line number Diff line number Diff line
@@ -10398,70 +10398,9 @@ static void disas_simd_three_reg_diff(DisasContext *s, uint32_t insn)
    }
}

static void gen_bsl_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
{
    tcg_gen_xor_i64(rn, rn, rm);
    tcg_gen_and_i64(rn, rn, rd);
    tcg_gen_xor_i64(rd, rm, rn);
}

static void gen_bit_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
{
    tcg_gen_xor_i64(rn, rn, rd);
    tcg_gen_and_i64(rn, rn, rm);
    tcg_gen_xor_i64(rd, rd, rn);
}

static void gen_bif_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
{
    tcg_gen_xor_i64(rn, rn, rd);
    tcg_gen_andc_i64(rn, rn, rm);
    tcg_gen_xor_i64(rd, rd, rn);
}

static void gen_bsl_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
{
    tcg_gen_xor_vec(vece, rn, rn, rm);
    tcg_gen_and_vec(vece, rn, rn, rd);
    tcg_gen_xor_vec(vece, rd, rm, rn);
}

static void gen_bit_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
{
    tcg_gen_xor_vec(vece, rn, rn, rd);
    tcg_gen_and_vec(vece, rn, rn, rm);
    tcg_gen_xor_vec(vece, rd, rd, rn);
}

static void gen_bif_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
{
    tcg_gen_xor_vec(vece, rn, rn, rd);
    tcg_gen_andc_vec(vece, rn, rn, rm);
    tcg_gen_xor_vec(vece, rd, rd, rn);
}

/* Logic op (opcode == 3) subgroup of C3.6.16. */
static void disas_simd_3same_logic(DisasContext *s, uint32_t insn)
{
    static const GVecGen3 bsl_op = {
        .fni8 = gen_bsl_i64,
        .fniv = gen_bsl_vec,
        .prefer_i64 = TCG_TARGET_REG_BITS == 64,
        .load_dest = true
    };
    static const GVecGen3 bit_op = {
        .fni8 = gen_bit_i64,
        .fniv = gen_bit_vec,
        .prefer_i64 = TCG_TARGET_REG_BITS == 64,
        .load_dest = true
    };
    static const GVecGen3 bif_op = {
        .fni8 = gen_bif_i64,
        .fniv = gen_bif_vec,
        .prefer_i64 = TCG_TARGET_REG_BITS == 64,
        .load_dest = true
    };

    int rd = extract32(insn, 0, 5);
    int rn = extract32(insn, 5, 5);
    int rm = extract32(insn, 16, 5);
+118 −44
Original line number Diff line number Diff line
@@ -5262,14 +5262,6 @@ static int disas_neon_ls_insn(DisasContext *s, uint32_t insn)
    return 0;
}

/* Bitwise select.  dest = c ? t : f.  Clobbers T and F.  */
static void gen_neon_bsl(TCGv_i32 dest, TCGv_i32 t, TCGv_i32 f, TCGv_i32 c)
{
    tcg_gen_and_i32(t, t, c);
    tcg_gen_andc_i32(f, f, c);
    tcg_gen_or_i32(dest, t, f);
}

static inline void gen_neon_narrow(int size, TCGv_i32 dest, TCGv_i64 src)
{
    switch (size) {
@@ -5712,6 +5704,73 @@ static int do_v81_helper(DisasContext *s, gen_helper_gvec_3_ptr *fn,
    return 1;
}

/*
 * Expanders for VBitOps_VBIF, VBIT, VBSL.
 */
static void gen_bsl_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
{
    tcg_gen_xor_i64(rn, rn, rm);
    tcg_gen_and_i64(rn, rn, rd);
    tcg_gen_xor_i64(rd, rm, rn);
}

static void gen_bit_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
{
    tcg_gen_xor_i64(rn, rn, rd);
    tcg_gen_and_i64(rn, rn, rm);
    tcg_gen_xor_i64(rd, rd, rn);
}

static void gen_bif_i64(TCGv_i64 rd, TCGv_i64 rn, TCGv_i64 rm)
{
    tcg_gen_xor_i64(rn, rn, rd);
    tcg_gen_andc_i64(rn, rn, rm);
    tcg_gen_xor_i64(rd, rd, rn);
}

static void gen_bsl_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
{
    tcg_gen_xor_vec(vece, rn, rn, rm);
    tcg_gen_and_vec(vece, rn, rn, rd);
    tcg_gen_xor_vec(vece, rd, rm, rn);
}

static void gen_bit_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
{
    tcg_gen_xor_vec(vece, rn, rn, rd);
    tcg_gen_and_vec(vece, rn, rn, rm);
    tcg_gen_xor_vec(vece, rd, rd, rn);
}

static void gen_bif_vec(unsigned vece, TCGv_vec rd, TCGv_vec rn, TCGv_vec rm)
{
    tcg_gen_xor_vec(vece, rn, rn, rd);
    tcg_gen_andc_vec(vece, rn, rn, rm);
    tcg_gen_xor_vec(vece, rd, rd, rn);
}

const GVecGen3 bsl_op = {
    .fni8 = gen_bsl_i64,
    .fniv = gen_bsl_vec,
    .prefer_i64 = TCG_TARGET_REG_BITS == 64,
    .load_dest = true
};

const GVecGen3 bit_op = {
    .fni8 = gen_bit_i64,
    .fniv = gen_bit_vec,
    .prefer_i64 = TCG_TARGET_REG_BITS == 64,
    .load_dest = true
};

const GVecGen3 bif_op = {
    .fni8 = gen_bif_i64,
    .fniv = gen_bif_vec,
    .prefer_i64 = TCG_TARGET_REG_BITS == 64,
    .load_dest = true
};


/* Translate a NEON data processing instruction.  Return nonzero if the
   instruction is invalid.
   We process data in a mixture of 32-bit and 64-bit chunks.
@@ -5721,13 +5780,14 @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
{
    int op;
    int q;
    int rd, rn, rm;
    int rd, rn, rm, rd_ofs, rn_ofs, rm_ofs;
    int size;
    int shift;
    int pass;
    int count;
    int pairwise;
    int u;
    int vec_size;
    uint32_t imm, mask;
    TCGv_i32 tmp, tmp2, tmp3, tmp4, tmp5;
    TCGv_ptr ptr1, ptr2, ptr3;
@@ -5751,6 +5811,11 @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
    VFP_DREG_N(rn, insn);
    VFP_DREG_M(rm, insn);
    size = (insn >> 20) & 3;
    vec_size = q ? 16 : 8;
    rd_ofs = neon_reg_offset(rd, 0);
    rn_ofs = neon_reg_offset(rn, 0);
    rm_ofs = neon_reg_offset(rm, 0);

    if ((insn & (1 << 23)) == 0) {
        /* Three register same length.  */
        op = ((insn >> 7) & 0x1e) | ((insn >> 4) & 1);
@@ -5841,8 +5906,51 @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
                                     q, rd, rn, rm);
            }
            return 1;

        case NEON_3R_LOGIC: /* Logic ops.  */
            switch ((u << 2) | size) {
            case 0: /* VAND */
                tcg_gen_gvec_and(0, rd_ofs, rn_ofs, rm_ofs,
                                 vec_size, vec_size);
                break;
            case 1: /* VBIC */
                tcg_gen_gvec_andc(0, rd_ofs, rn_ofs, rm_ofs,
                                  vec_size, vec_size);
                break;
            case 2:
                if (rn == rm) {
                    /* VMOV */
                    tcg_gen_gvec_mov(0, rd_ofs, rn_ofs, vec_size, vec_size);
                } else {
                    /* VORR */
                    tcg_gen_gvec_or(0, rd_ofs, rn_ofs, rm_ofs,
                                    vec_size, vec_size);
                }
        if (size == 3 && op != NEON_3R_LOGIC) {
                break;
            case 3: /* VORN */
                tcg_gen_gvec_orc(0, rd_ofs, rn_ofs, rm_ofs,
                                 vec_size, vec_size);
                break;
            case 4: /* VEOR */
                tcg_gen_gvec_xor(0, rd_ofs, rn_ofs, rm_ofs,
                                 vec_size, vec_size);
                break;
            case 5: /* VBSL */
                tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs,
                               vec_size, vec_size, &bsl_op);
                break;
            case 6: /* VBIT */
                tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs,
                               vec_size, vec_size, &bit_op);
                break;
            case 7: /* VBIF */
                tcg_gen_gvec_3(rd_ofs, rn_ofs, rm_ofs,
                               vec_size, vec_size, &bif_op);
                break;
            }
            return 0;
        }
        if (size == 3) {
            /* 64-bit element instructions. */
            for (pass = 0; pass < (q ? 2 : 1); pass++) {
                neon_load_reg64(cpu_V0, rn + pass);
@@ -6000,40 +6108,6 @@ static int disas_neon_data_insn(DisasContext *s, uint32_t insn)
        case NEON_3R_VRHADD:
            GEN_NEON_INTEGER_OP(rhadd);
            break;
        case NEON_3R_LOGIC: /* Logic ops.  */
            switch ((u << 2) | size) {
            case 0: /* VAND */
                tcg_gen_and_i32(tmp, tmp, tmp2);
                break;
            case 1: /* BIC */
                tcg_gen_andc_i32(tmp, tmp, tmp2);
                break;
            case 2: /* VORR */
                tcg_gen_or_i32(tmp, tmp, tmp2);
                break;
            case 3: /* VORN */
                tcg_gen_orc_i32(tmp, tmp, tmp2);
                break;
            case 4: /* VEOR */
                tcg_gen_xor_i32(tmp, tmp, tmp2);
                break;
            case 5: /* VBSL */
                tmp3 = neon_load_reg(rd, pass);
                gen_neon_bsl(tmp, tmp, tmp2, tmp3);
                tcg_temp_free_i32(tmp3);
                break;
            case 6: /* VBIT */
                tmp3 = neon_load_reg(rd, pass);
                gen_neon_bsl(tmp, tmp, tmp3, tmp2);
                tcg_temp_free_i32(tmp3);
                break;
            case 7: /* VBIF */
                tmp3 = neon_load_reg(rd, pass);
                gen_neon_bsl(tmp, tmp3, tmp, tmp2);
                tcg_temp_free_i32(tmp3);
                break;
            }
            break;
        case NEON_3R_VHSUB:
            GEN_NEON_INTEGER_OP(hsub);
            break;
+6 −0
Original line number Diff line number Diff line
@@ -191,6 +191,12 @@ static inline TCGv_i32 get_ahp_flag(void)
    return ret;
}


/* Vector operations shared between ARM and AArch64.  */
extern const GVecGen3 bsl_op;
extern const GVecGen3 bit_op;
extern const GVecGen3 bif_op;

/*
 * Forward to the isar_feature_* tests given a DisasContext pointer.
 */