mirror of
https://github.com/yuzu-emu/unicorn.git
synced 2024-12-23 00:35:34 +00:00
tcg: Increase tcg_out_dupi_vec immediate to int64_t
While we don't store more than tcg_target_long in TCGTemp, we shouldn't be limited to that for code generation. We will be able to use this for INDEX_op_dup2_vec with 2 constants. Also pass along the minimal vece that may be said to apply to the constant. This allows some simplification in the various backends. Backports 4e18617555955503628a004ed97e1fc2fa7818b9
This commit is contained in:
parent
c5c19529c5
commit
520ec7ca76
|
@ -849,14 +849,14 @@ static void tcg_out_logicali(TCGContext *s, AArch64Insn insn, TCGType ext,
|
|||
tcg_out_insn_3404(s, insn, ext, rd, rn, ext, r, c);
|
||||
}
|
||||
|
||||
static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
|
||||
TCGReg rd, tcg_target_long v64)
|
||||
static void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
|
||||
TCGReg rd, int64_t v64)
|
||||
{
|
||||
bool q = type == TCG_TYPE_V128;
|
||||
int cmode, imm8, i;
|
||||
|
||||
/* Test all bytes equal first. */
|
||||
if (v64 == dup_const(MO_8, v64)) {
|
||||
if (vece == MO_8) {
|
||||
imm8 = (uint8_t)v64;
|
||||
tcg_out_insn(s, 3606, MOVI, q, rd, 0, 0xe, imm8);
|
||||
return;
|
||||
|
@ -883,7 +883,7 @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
|
|||
* cannot find an expansion there's no point checking a larger
|
||||
* width because we already know by replication it cannot match.
|
||||
*/
|
||||
if (v64 == dup_const(MO_16, v64)) {
|
||||
if (vece == MO_16) {
|
||||
uint16_t v16 = v64;
|
||||
|
||||
if (is_shimm16(v16, &cmode, &imm8)) {
|
||||
|
@ -902,7 +902,7 @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
|
|||
tcg_out_insn(s, 3606, MOVI, q, rd, 0, 0x8, v16 & 0xff);
|
||||
tcg_out_insn(s, 3606, ORR, q, rd, 0, 0xa, v16 >> 8);
|
||||
return;
|
||||
} else if (v64 == dup_const(MO_32, v64)) {
|
||||
} else if (vece == MO_32) {
|
||||
uint32_t v32 = v64;
|
||||
uint32_t n32 = ~v32;
|
||||
|
||||
|
@ -2429,7 +2429,7 @@ static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc,
|
|||
tcg_out_insn_3617(s, insn, is_q, vece, a0, a1);
|
||||
break;
|
||||
}
|
||||
tcg_out_dupi_vec(s, type, TCG_VEC_TMP, 0);
|
||||
tcg_out_dupi_vec(s, type, MO_8, TCG_VEC_TMP, 0);
|
||||
a2 = TCG_VEC_TMP;
|
||||
}
|
||||
insn = cmp_insn[cond];
|
||||
|
|
|
@ -948,8 +948,8 @@ static bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece,
|
|||
return true;
|
||||
}
|
||||
|
||||
static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
|
||||
TCGReg ret, tcg_target_long arg)
|
||||
static void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
|
||||
TCGReg ret, int64_t arg)
|
||||
{
|
||||
int vex_l = (type == TCG_TYPE_V256 ? P_VEXL : 0);
|
||||
|
||||
|
@ -962,7 +962,14 @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
|
|||
return;
|
||||
}
|
||||
|
||||
if (TCG_TARGET_REG_BITS == 64) {
|
||||
if (TCG_TARGET_REG_BITS == 32 && vece < MO_64) {
|
||||
if (have_avx2) {
|
||||
tcg_out_vex_modrm_pool(s, OPC_VPBROADCASTD + vex_l, ret);
|
||||
} else {
|
||||
tcg_out_vex_modrm_pool(s, OPC_VBROADCASTSS, ret);
|
||||
}
|
||||
new_pool_label(s, arg, R_386_32, s->code_ptr - 4, 0);
|
||||
} else {
|
||||
if (type == TCG_TYPE_V64) {
|
||||
tcg_out_vex_modrm_pool(s, OPC_MOVQ_VqWq, ret);
|
||||
} else if (have_avx2) {
|
||||
|
@ -970,14 +977,11 @@ static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
|
|||
} else {
|
||||
tcg_out_vex_modrm_pool(s, OPC_MOVDDUP, ret);
|
||||
}
|
||||
if (TCG_TARGET_REG_BITS == 64) {
|
||||
new_pool_label(s, arg, R_386_PC32, s->code_ptr - 4, -4);
|
||||
} else {
|
||||
if (have_avx2) {
|
||||
tcg_out_vex_modrm_pool(s, OPC_VPBROADCASTD + vex_l, ret);
|
||||
} else {
|
||||
tcg_out_vex_modrm_pool(s, OPC_VBROADCASTSS, ret);
|
||||
new_pool_l2(s, R_386_32, s->code_ptr - 4, 0, arg, arg >> 32);
|
||||
}
|
||||
new_pool_label(s, arg, R_386_32, s->code_ptr - 4, 0);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -104,8 +104,8 @@ static bool tcg_out_dup_vec(TCGContext *s, TCGType type, unsigned vece,
|
|||
TCGReg dst, TCGReg src);
|
||||
static bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece,
|
||||
TCGReg dst, TCGReg base, intptr_t offset);
|
||||
static void tcg_out_dupi_vec(TCGContext *s, TCGType type,
|
||||
TCGReg dst, tcg_target_long arg);
|
||||
static void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
|
||||
TCGReg dst, int64_t arg);
|
||||
static void tcg_out_vec_op(TCGContext *s, TCGOpcode opc, unsigned vecl,
|
||||
unsigned vece, const TCGArg *args,
|
||||
const int *const_args);
|
||||
|
@ -120,8 +120,8 @@ static inline bool tcg_out_dupm_vec(TCGContext *s, TCGType type, unsigned vece,
|
|||
{
|
||||
g_assert_not_reached();
|
||||
}
|
||||
static inline void tcg_out_dupi_vec(TCGContext *s, TCGType type,
|
||||
TCGReg dst, tcg_target_long arg)
|
||||
static inline void tcg_out_dupi_vec(TCGContext *s, TCGType type, unsigned vece,
|
||||
TCGReg dst, int64_t arg)
|
||||
{
|
||||
g_assert_not_reached();
|
||||
}
|
||||
|
@ -2741,7 +2741,28 @@ static void temp_load(TCGContext *s, TCGTemp *ts, TCGRegSet desired_regs,
|
|||
if (ts->type <= TCG_TYPE_I64) {
|
||||
tcg_out_movi(s, ts->type, reg, ts->val);
|
||||
} else {
|
||||
tcg_out_dupi_vec(s, ts->type, reg, ts->val);
|
||||
uint64_t val = ts->val;
|
||||
MemOp vece = MO_64;
|
||||
|
||||
/*
|
||||
* Find the minimal vector element that matches the constant.
|
||||
* The targets will, in general, have to do this search anyway,
|
||||
* do this generically.
|
||||
*/
|
||||
if (TCG_TARGET_REG_BITS == 32) {
|
||||
val = dup_const(MO_32, val);
|
||||
vece = MO_32;
|
||||
}
|
||||
if (val == dup_const(MO_8, val)) {
|
||||
vece = MO_8;
|
||||
} else if (val == dup_const(MO_16, val)) {
|
||||
vece = MO_16;
|
||||
} else if (TCG_TARGET_REG_BITS == 64 &&
|
||||
val == dup_const(MO_32, val)) {
|
||||
vece = MO_32;
|
||||
}
|
||||
|
||||
tcg_out_dupi_vec(s, ts->type, vece, reg, ts->val);
|
||||
}
|
||||
ts->mem_coherent = 0;
|
||||
break;
|
||||
|
|
Loading…
Reference in a new issue