Skip to content

Commit

Permalink
target/ppc: implement vrlqnm
Browse files Browse the repository at this point in the history
Signed-off-by: Matheus Ferst <matheus.ferst@eldorado.org.br>
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
Message-Id: <20220225210936.1749575-27-matheus.ferst@eldorado.org.br>
Signed-off-by: Cédric Le Goater <clg@kaod.org>
  • Loading branch information
mferst authored and legoater committed Mar 2, 2022
1 parent aa0f34e commit 4e27266
Show file tree
Hide file tree
Showing 2 changed files with 77 additions and 5 deletions.
1 change: 1 addition & 0 deletions target/ppc/insn32.decode
Expand Up @@ -498,6 +498,7 @@ VRLDMI 000100 ..... ..... ..... 00011000101 @VX

VRLWNM 000100 ..... ..... ..... 00110000101 @VX
VRLDNM 000100 ..... ..... ..... 00111000101 @VX
VRLQNM 000100 ..... ..... ..... 00101000101 @VX

## Vector Integer Arithmetic Instructions

Expand Down
81 changes: 76 additions & 5 deletions target/ppc/translate/vmx-impl.c.inc
Expand Up @@ -1055,28 +1055,83 @@ TRANS_FLAGS2(ISA310, VSLQ, do_vector_shift_quad, false, false);
TRANS_FLAGS2(ISA310, VSRQ, do_vector_shift_quad, true, false);
TRANS_FLAGS2(ISA310, VSRAQ, do_vector_shift_quad, true, true);

static bool trans_VRLQ(DisasContext *ctx, arg_VX *a)
static void do_vrlq_mask(TCGv_i64 mh, TCGv_i64 ml, TCGv_i64 b, TCGv_i64 e)
{
TCGv_i64 ah, al, n, t0, t1, zero = tcg_constant_i64(0);
TCGv_i64 th, tl, t0, t1, zero = tcg_constant_i64(0),
ones = tcg_constant_i64(-1);

th = tcg_temp_new_i64();
tl = tcg_temp_new_i64();
t0 = tcg_temp_new_i64();
t1 = tcg_temp_new_i64();

/* m = ~0 >> b */
tcg_gen_andi_i64(t0, b, 64);
tcg_gen_movcond_i64(TCG_COND_NE, t1, t0, zero, zero, ones);
tcg_gen_andi_i64(t0, b, 0x3F);
tcg_gen_shr_i64(mh, t1, t0);
tcg_gen_shr_i64(ml, ones, t0);
tcg_gen_xori_i64(t0, t0, 63);
tcg_gen_shl_i64(t1, t1, t0);
tcg_gen_shli_i64(t1, t1, 1);
tcg_gen_or_i64(ml, t1, ml);

/* t = ~0 >> e */
tcg_gen_andi_i64(t0, e, 64);
tcg_gen_movcond_i64(TCG_COND_NE, t1, t0, zero, zero, ones);
tcg_gen_andi_i64(t0, e, 0x3F);
tcg_gen_shr_i64(th, t1, t0);
tcg_gen_shr_i64(tl, ones, t0);
tcg_gen_xori_i64(t0, t0, 63);
tcg_gen_shl_i64(t1, t1, t0);
tcg_gen_shli_i64(t1, t1, 1);
tcg_gen_or_i64(tl, t1, tl);

/* t = t >> 1 */
tcg_gen_shli_i64(t0, th, 63);
tcg_gen_shri_i64(tl, tl, 1);
tcg_gen_shri_i64(th, th, 1);
tcg_gen_or_i64(tl, t0, tl);

/* m = m ^ t */
tcg_gen_xor_i64(mh, mh, th);
tcg_gen_xor_i64(ml, ml, tl);

/* Negate the mask if begin > end */
tcg_gen_movcond_i64(TCG_COND_GT, t0, b, e, ones, zero);

tcg_gen_xor_i64(mh, mh, t0);
tcg_gen_xor_i64(ml, ml, t0);

tcg_temp_free_i64(th);
tcg_temp_free_i64(tl);
tcg_temp_free_i64(t0);
tcg_temp_free_i64(t1);
}

static bool do_vector_rotl_quad(DisasContext *ctx, arg_VX *a, bool mask)
{
TCGv_i64 ah, al, vrb, n, t0, t1, zero = tcg_constant_i64(0);

REQUIRE_VECTOR(ctx);
REQUIRE_INSNS_FLAGS2(ctx, ISA310);

ah = tcg_temp_new_i64();
al = tcg_temp_new_i64();
vrb = tcg_temp_new_i64();
n = tcg_temp_new_i64();
t0 = tcg_temp_new_i64();
t1 = tcg_temp_new_i64();

get_avr64(ah, a->vra, true);
get_avr64(al, a->vra, false);
get_avr64(n, a->vrb, true);
get_avr64(vrb, a->vrb, true);

tcg_gen_mov_i64(t0, ah);
tcg_gen_andi_i64(t1, n, 64);
tcg_gen_andi_i64(t1, vrb, 64);
tcg_gen_movcond_i64(TCG_COND_NE, ah, t1, zero, al, ah);
tcg_gen_movcond_i64(TCG_COND_NE, al, t1, zero, t0, al);
tcg_gen_andi_i64(n, n, 0x3F);
tcg_gen_andi_i64(n, vrb, 0x3F);

tcg_gen_shl_i64(t0, ah, n);
tcg_gen_shl_i64(t1, al, n);
Expand All @@ -1091,18 +1146,34 @@ static bool trans_VRLQ(DisasContext *ctx, arg_VX *a)
tcg_gen_shri_i64(ah, ah, 1);
tcg_gen_or_i64(t1, ah, t1);

if (mask) {
tcg_gen_shri_i64(n, vrb, 8);
tcg_gen_shri_i64(vrb, vrb, 16);
tcg_gen_andi_i64(n, n, 0x7f);
tcg_gen_andi_i64(vrb, vrb, 0x7f);

do_vrlq_mask(ah, al, vrb, n);

tcg_gen_and_i64(t0, t0, ah);
tcg_gen_and_i64(t1, t1, al);
}

set_avr64(a->vrt, t0, true);
set_avr64(a->vrt, t1, false);

tcg_temp_free_i64(ah);
tcg_temp_free_i64(al);
tcg_temp_free_i64(vrb);
tcg_temp_free_i64(n);
tcg_temp_free_i64(t0);
tcg_temp_free_i64(t1);

return true;
}

TRANS(VRLQ, do_vector_rotl_quad, false)
TRANS(VRLQNM, do_vector_rotl_quad, true)

#define GEN_VXFORM_SAT(NAME, VECE, NORM, SAT, OPC2, OPC3) \
static void glue(glue(gen_, NAME), _vec)(unsigned vece, TCGv_vec t, \
TCGv_vec sat, TCGv_vec a, \
Expand Down

0 comments on commit 4e27266

Please sign in to comment.