mirror of
https://github.com/yuzu-emu/unicorn.git
synced 2024-12-23 18:15:30 +00:00
b3e89e9996
The x86 vector instruction set is extremely irregular. With newer editions, Intel has filled in some of the blanks. However, we don't get many 64-bit operations until SSE4.2, introduced in 2009. The subsequent edition was for AVX1, introduced in 2011, which added three-operand addressing, and adjusts how all instructions should be encoded. Given the relatively narrow 2 year window between possible to support and desirable to support, and to vastly simplify code maintainence, I am only planning to support AVX1 and later cpus. Backports commit 770c2fc7bb70804ae9869995fd02dadd6d7656ac from qemu
265 lines
8.1 KiB
C
265 lines
8.1 KiB
C
/*
|
|
* Tiny Code Generator for QEMU
|
|
*
|
|
* Copyright (c) 2008 Fabrice Bellard
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
* of this software and associated documentation files (the "Software"), to deal
|
|
* in the Software without restriction, including without limitation the rights
|
|
* to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
* copies of the Software, and to permit persons to whom the Software is
|
|
* furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice shall be included in
|
|
* all copies or substantial portions of the Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
|
|
* THE SOFTWARE.
|
|
*/
|
|
|
|
#ifndef I386_TCG_TARGET_H
|
|
#define I386_TCG_TARGET_H
|
|
|
|
#define TCG_TARGET_INSN_UNIT_SIZE 1
|
|
#define TCG_TARGET_TLB_DISPLACEMENT_BITS 31
|
|
|
|
#ifdef __x86_64__
|
|
# define TCG_TARGET_REG_BITS 64
|
|
# define TCG_TARGET_NB_REGS 32
|
|
#else
|
|
# define TCG_TARGET_REG_BITS 32
|
|
# define TCG_TARGET_NB_REGS 24
|
|
#endif
|
|
|
|
typedef enum {
|
|
TCG_REG_EAX = 0,
|
|
TCG_REG_ECX,
|
|
TCG_REG_EDX,
|
|
TCG_REG_EBX,
|
|
TCG_REG_ESP,
|
|
TCG_REG_EBP,
|
|
TCG_REG_ESI,
|
|
TCG_REG_EDI,
|
|
|
|
/* 64-bit registers; always define the symbols to avoid
|
|
too much if-deffing. */
|
|
TCG_REG_R8,
|
|
TCG_REG_R9,
|
|
TCG_REG_R10,
|
|
TCG_REG_R11,
|
|
TCG_REG_R12,
|
|
TCG_REG_R13,
|
|
TCG_REG_R14,
|
|
TCG_REG_R15,
|
|
|
|
TCG_REG_XMM0,
|
|
TCG_REG_XMM1,
|
|
TCG_REG_XMM2,
|
|
TCG_REG_XMM3,
|
|
TCG_REG_XMM4,
|
|
TCG_REG_XMM5,
|
|
TCG_REG_XMM6,
|
|
TCG_REG_XMM7,
|
|
|
|
/* 64-bit registers; likewise always define. */
|
|
TCG_REG_XMM8,
|
|
TCG_REG_XMM9,
|
|
TCG_REG_XMM10,
|
|
TCG_REG_XMM11,
|
|
TCG_REG_XMM12,
|
|
TCG_REG_XMM13,
|
|
TCG_REG_XMM14,
|
|
TCG_REG_XMM15,
|
|
|
|
TCG_REG_RAX = TCG_REG_EAX,
|
|
TCG_REG_RCX = TCG_REG_ECX,
|
|
TCG_REG_RDX = TCG_REG_EDX,
|
|
TCG_REG_RBX = TCG_REG_EBX,
|
|
TCG_REG_RSP = TCG_REG_ESP,
|
|
TCG_REG_RBP = TCG_REG_EBP,
|
|
TCG_REG_RSI = TCG_REG_ESI,
|
|
TCG_REG_RDI = TCG_REG_EDI,
|
|
} TCGReg;
|
|
|
|
/* used for function call generation */
|
|
#define TCG_REG_CALL_STACK TCG_REG_ESP
|
|
#define TCG_TARGET_STACK_ALIGN 16
|
|
#if defined(_WIN64) || (defined(__CYGWIN__) && defined(__x86_64__))
|
|
#define TCG_TARGET_CALL_STACK_OFFSET 32
|
|
#else
|
|
#define TCG_TARGET_CALL_STACK_OFFSET 0
|
|
#endif
|
|
|
|
extern bool have_bmi1;
|
|
extern bool have_popcnt;
|
|
extern bool have_avx1;
|
|
extern bool have_avx2;
|
|
|
|
// UNICORN FIXME:
|
|
// Taken from cpuid.h in mainline qemu.
|
|
// The cpuid mechanism should just be
|
|
// backported instead
|
|
|
|
/* Leaf 1, %ecx */
|
|
#ifndef bit_SSE4_1
|
|
#define bit_SSE4_1 (1 << 19)
|
|
#endif
|
|
#ifndef bit_MOVBE
|
|
#define bit_MOVBE (1 << 22)
|
|
#endif
|
|
#ifndef bit_OSXSAVE
|
|
#define bit_OSXSAVE (1 << 27)
|
|
#endif
|
|
#ifndef bit_AVX
|
|
#define bit_AVX (1 << 28)
|
|
#endif
|
|
/* Leaf 7, %ebx */
|
|
#ifndef bit_BMI
|
|
#define bit_BMI (1 << 3)
|
|
#endif
|
|
#ifndef bit_AVX2
|
|
#define bit_AVX2 (1 << 5)
|
|
#endif
|
|
#ifndef bit_BMI2
|
|
#define bit_BMI2 (1 << 8)
|
|
#endif
|
|
/* Leaf 0x80000001, %ecx */
|
|
#ifndef bit_LZCNT
|
|
#define bit_LZCNT (1 << 5)
|
|
#endif
|
|
|
|
/* optional instructions */
|
|
#define TCG_TARGET_HAS_div2_i32 1
|
|
#define TCG_TARGET_HAS_rot_i32 1
|
|
#define TCG_TARGET_HAS_ext8s_i32 1
|
|
#define TCG_TARGET_HAS_ext16s_i32 1
|
|
#define TCG_TARGET_HAS_ext8u_i32 1
|
|
#define TCG_TARGET_HAS_ext16u_i32 1
|
|
#define TCG_TARGET_HAS_bswap16_i32 1
|
|
#define TCG_TARGET_HAS_bswap32_i32 1
|
|
#define TCG_TARGET_HAS_neg_i32 1
|
|
#define TCG_TARGET_HAS_not_i32 1
|
|
#define TCG_TARGET_HAS_andc_i32 have_bmi1
|
|
#define TCG_TARGET_HAS_orc_i32 0
|
|
#define TCG_TARGET_HAS_eqv_i32 0
|
|
#define TCG_TARGET_HAS_nand_i32 0
|
|
#define TCG_TARGET_HAS_nor_i32 0
|
|
#define TCG_TARGET_HAS_clz_i32 1
|
|
#define TCG_TARGET_HAS_ctz_i32 1
|
|
#define TCG_TARGET_HAS_ctpop_i32 have_popcnt
|
|
#define TCG_TARGET_HAS_deposit_i32 1
|
|
#define TCG_TARGET_HAS_extract_i32 1
|
|
#define TCG_TARGET_HAS_sextract_i32 1
|
|
#define TCG_TARGET_HAS_movcond_i32 1
|
|
#define TCG_TARGET_HAS_add2_i32 1
|
|
#define TCG_TARGET_HAS_sub2_i32 1
|
|
#define TCG_TARGET_HAS_mulu2_i32 1
|
|
#define TCG_TARGET_HAS_muls2_i32 1
|
|
#define TCG_TARGET_HAS_muluh_i32 0
|
|
#define TCG_TARGET_HAS_mulsh_i32 0
|
|
#define TCG_TARGET_HAS_goto_ptr 1
|
|
#define TCG_TARGET_HAS_direct_jump 1
|
|
|
|
#if TCG_TARGET_REG_BITS == 64
|
|
#define TCG_TARGET_HAS_extrl_i64_i32 0
|
|
#define TCG_TARGET_HAS_extrh_i64_i32 0
|
|
#define TCG_TARGET_HAS_div2_i64 1
|
|
#define TCG_TARGET_HAS_rot_i64 1
|
|
#define TCG_TARGET_HAS_ext8s_i64 1
|
|
#define TCG_TARGET_HAS_ext16s_i64 1
|
|
#define TCG_TARGET_HAS_ext32s_i64 1
|
|
#define TCG_TARGET_HAS_ext8u_i64 1
|
|
#define TCG_TARGET_HAS_ext16u_i64 1
|
|
#define TCG_TARGET_HAS_ext32u_i64 1
|
|
#define TCG_TARGET_HAS_bswap16_i64 1
|
|
#define TCG_TARGET_HAS_bswap32_i64 1
|
|
#define TCG_TARGET_HAS_bswap64_i64 1
|
|
#define TCG_TARGET_HAS_neg_i64 1
|
|
#define TCG_TARGET_HAS_not_i64 1
|
|
#define TCG_TARGET_HAS_andc_i64 have_bmi1
|
|
#define TCG_TARGET_HAS_orc_i64 0
|
|
#define TCG_TARGET_HAS_eqv_i64 0
|
|
#define TCG_TARGET_HAS_nand_i64 0
|
|
#define TCG_TARGET_HAS_nor_i64 0
|
|
#define TCG_TARGET_HAS_clz_i64 1
|
|
#define TCG_TARGET_HAS_ctz_i64 1
|
|
#define TCG_TARGET_HAS_ctpop_i64 have_popcnt
|
|
#define TCG_TARGET_HAS_deposit_i64 1
|
|
#define TCG_TARGET_HAS_extract_i64 1
|
|
#define TCG_TARGET_HAS_sextract_i64 0
|
|
#define TCG_TARGET_HAS_movcond_i64 1
|
|
#define TCG_TARGET_HAS_add2_i64 1
|
|
#define TCG_TARGET_HAS_sub2_i64 1
|
|
#define TCG_TARGET_HAS_mulu2_i64 1
|
|
#define TCG_TARGET_HAS_muls2_i64 1
|
|
#define TCG_TARGET_HAS_muluh_i64 0
|
|
#define TCG_TARGET_HAS_mulsh_i64 0
|
|
#endif
|
|
|
|
/* We do not support older SSE systems, only beginning with AVX1. */
|
|
#define TCG_TARGET_HAS_v64 have_avx1
|
|
#define TCG_TARGET_HAS_v128 have_avx1
|
|
#define TCG_TARGET_HAS_v256 have_avx2
|
|
|
|
#define TCG_TARGET_HAS_andc_vec 1
|
|
#define TCG_TARGET_HAS_orc_vec 0
|
|
#define TCG_TARGET_HAS_not_vec 0
|
|
#define TCG_TARGET_HAS_neg_vec 0
|
|
#define TCG_TARGET_HAS_shi_vec 1
|
|
#define TCG_TARGET_HAS_shs_vec 0
|
|
#define TCG_TARGET_HAS_shv_vec 0
|
|
#define TCG_TARGET_HAS_cmp_vec 1
|
|
#define TCG_TARGET_HAS_mul_vec 1
|
|
|
|
#define TCG_TARGET_deposit_i32_valid(ofs, len) \
|
|
(((ofs) == 0 && (len) == 8) || ((ofs) == 8 && (len) == 8) || \
|
|
((ofs) == 0 && (len) == 16))
|
|
#define TCG_TARGET_deposit_i64_valid TCG_TARGET_deposit_i32_valid
|
|
|
|
/* Check for the possibility of high-byte extraction and, for 64-bit,
|
|
zero-extending 32-bit right-shift. */
|
|
#define TCG_TARGET_extract_i32_valid(ofs, len) ((ofs) == 8 && (len) == 8)
|
|
#define TCG_TARGET_extract_i64_valid(ofs, len) \
|
|
(((ofs) == 8 && (len) == 8) || ((ofs) + (len)) == 32)
|
|
|
|
#if TCG_TARGET_REG_BITS == 64
|
|
# define TCG_AREG0 TCG_REG_R14
|
|
#else
|
|
# define TCG_AREG0 TCG_REG_EBP
|
|
#endif
|
|
|
|
static inline void flush_icache_range(uintptr_t start, uintptr_t stop)
|
|
{
|
|
}
|
|
|
|
static inline void tb_target_set_jmp_target(uintptr_t tc_ptr,
|
|
uintptr_t jmp_addr, uintptr_t addr)
|
|
{
|
|
/* patch the branch destination */
|
|
atomic_set((int32_t *)jmp_addr, addr - (jmp_addr + 4));
|
|
/* no need to flush icache explicitly */
|
|
}
|
|
|
|
/* This defines the natural memory order supported by this
|
|
* architecture before guarantees made by various barrier
|
|
* instructions.
|
|
*
|
|
* The x86 has a pretty strong memory ordering which only really
|
|
* allows for some stores to be re-ordered after loads.
|
|
*/
|
|
#include "tcg-mo.h"
|
|
|
|
#define TCG_TARGET_DEFAULT_MO (TCG_MO_ALL & ~TCG_MO_ST_LD)
|
|
|
|
#ifdef CONFIG_SOFTMMU
|
|
#define TCG_TARGET_NEED_LDST_LABELS
|
|
#endif
|
|
#define TCG_TARGET_NEED_POOL_LABELS
|
|
|
|
#endif
|