mirror of
https://github.com/c3lang/c3c.git
synced 2026-02-27 20:11:17 +00:00
959 lines
36 KiB
C
959 lines
36 KiB
C
// Copyright (c) 2022-2023 Christoffer Lerno. All rights reserved.
|
|
// Use of this source code is governed by a LGPLv3.0
|
|
// a copy of which can be found in the LICENSE file.
|
|
#include "compiler_internal.h"
|
|
#include "compiler/asm/x86.h"
|
|
#include "compiler/asm/aarch64.h"
|
|
#include "compiler/asm/riscv.h"
|
|
|
|
#define ASM_PTR_HASH(name__) (uint32_t)(((uintptr_t)name__ * 31) ^ ((uintptr_t)name__ >> 15))
|
|
|
|
const Clobbers NO_CLOBBER = { .mask[0] = 0 };
|
|
|
|
|
|
INLINE AsmInstruction *insert_instruction_named(PlatformTarget *target, const char *name)
|
|
{
|
|
TokenType token_type = TOKEN_IDENT;
|
|
unsigned len = (unsigned)strlen(name);
|
|
const char *interned = symtab_add(name, len, fnv1a(name, len), &token_type);
|
|
uint32_t hash = ASM_PTR_HASH(interned);
|
|
uint32_t slot = hash & ASM_INSTRUCTION_MASK;
|
|
while (1)
|
|
{
|
|
AsmInstruction *instr = &target->instructions[slot];
|
|
if (!instr->name)
|
|
{
|
|
instr->name = interned;
|
|
return instr;
|
|
}
|
|
slot = (slot + 1) & ASM_INSTRUCTION_MASK;
|
|
}
|
|
}
|
|
|
|
INLINE AsmArgBits parse_bits(const char **desc)
|
|
{
|
|
if (str_start_with(*desc, "80"))
|
|
{
|
|
*desc += 2;
|
|
return ARG_BITS_80;
|
|
}
|
|
if (str_start_with(*desc, "8"))
|
|
{
|
|
*desc += 1;
|
|
return ARG_BITS_8;
|
|
}
|
|
if (str_start_with(*desc, "16"))
|
|
{
|
|
*desc += 2;
|
|
return ARG_BITS_16;
|
|
}
|
|
if (str_start_with(*desc, "20"))
|
|
{
|
|
*desc += 2;
|
|
return ARG_BITS_20;
|
|
}
|
|
if (str_start_with(*desc, "32"))
|
|
{
|
|
*desc += 2;
|
|
return ARG_BITS_32;
|
|
}
|
|
if (str_start_with(*desc, "64"))
|
|
{
|
|
*desc += 2;
|
|
return ARG_BITS_64;
|
|
}
|
|
if (str_start_with(*desc, "128"))
|
|
{
|
|
*desc += 3;
|
|
return ARG_BITS_128;
|
|
}
|
|
if (memcmp("12", *desc, 2) == 0)
|
|
{
|
|
*desc += 2;
|
|
return ARG_BITS_12;
|
|
}
|
|
if (str_start_with(*desc, "256"))
|
|
{
|
|
*desc += 3;
|
|
return ARG_BITS_256;
|
|
}
|
|
if (str_start_with(*desc, "512"))
|
|
{
|
|
*desc += 3;
|
|
return ARG_BITS_512;
|
|
}
|
|
if (str_start_with(*desc, "5"))
|
|
{
|
|
*desc += 1;
|
|
return ARG_BITS_5;
|
|
}
|
|
error_exit("Invalid bits: %s.", *desc);
|
|
}
|
|
|
|
INLINE AsmArgType decode_arg_type(const char **desc)
|
|
{
|
|
AsmArgType arg_type = { .is_readwrite = false };
|
|
if (**desc == 'w')
|
|
{
|
|
arg_type.is_write = true;
|
|
assert((*desc)[1] == ':');
|
|
*desc += 2;
|
|
}
|
|
else if (**desc == 'r' && (*desc)[1] == 'w')
|
|
{
|
|
arg_type.is_readwrite = true;
|
|
arg_type.is_write = true;
|
|
assert((*desc)[2] == ':');
|
|
*desc += 3;
|
|
}
|
|
char c;
|
|
while ((c = ((*desc)++)[0]) != 0 && c != ',')
|
|
{
|
|
switch (c)
|
|
{
|
|
case 'r':
|
|
arg_type.ireg_bits |= parse_bits(desc);
|
|
goto NEXT;
|
|
case 'm':
|
|
if (memcmp("mem", &(*desc)[-1], 3) == 0)
|
|
{
|
|
arg_type.is_address = true;
|
|
*desc += 2;
|
|
goto NEXT;
|
|
}
|
|
error_exit("Unexpected string %s", &(*desc)[-1]);
|
|
case 'v':
|
|
arg_type.vec_bits |= parse_bits(desc);
|
|
goto NEXT;
|
|
case 'i':
|
|
if (memcmp("mm", *desc, 2) == 0)
|
|
{
|
|
*desc += 2;
|
|
c = **desc;
|
|
if (c == 'i')
|
|
{
|
|
(*desc)++;
|
|
arg_type.imm_arg_ibits |= parse_bits(desc);
|
|
goto NEXT;
|
|
}
|
|
if (c == 'u')
|
|
{
|
|
(*desc)++;
|
|
arg_type.imm_arg_ubits |= parse_bits(desc);
|
|
goto NEXT;
|
|
}
|
|
AsmArgBits bits = parse_bits(desc);
|
|
arg_type.imm_arg_ibits |= bits;
|
|
arg_type.imm_arg_ubits |= bits;
|
|
goto NEXT;
|
|
}
|
|
default:
|
|
error_exit("Unexpected string '%s'.", &(*desc)[-1]);
|
|
}
|
|
NEXT:
|
|
switch (**desc)
|
|
{
|
|
case '/':
|
|
(*desc)++;
|
|
continue;
|
|
case ',':
|
|
(*desc)++;
|
|
while (**desc == ' ') (*desc)++;
|
|
FALLTHROUGH;
|
|
case 0:
|
|
return arg_type;
|
|
default:
|
|
error_exit("Expected '/' or end: '%s'.", desc);
|
|
}
|
|
}
|
|
return arg_type;
|
|
}
|
|
static inline void reg_instr_clob(PlatformTarget *target, const char *name, Clobbers mask, const char *args)
|
|
{
|
|
AsmInstruction *instr = insert_instruction_named(target, name);
|
|
instr->mask = mask;
|
|
unsigned param_count = 0;
|
|
while (args && args[0] != 0)
|
|
{
|
|
ASSERT(param_count <= MAX_ASM_INSTRUCTION_PARAMS);
|
|
instr->param[param_count++] = decode_arg_type(&args);
|
|
}
|
|
instr->param_count = param_count;
|
|
}
|
|
|
|
static inline void reg_instr(PlatformTarget *target, const char *name, const char *args)
|
|
{
|
|
AsmInstruction *instr = insert_instruction_named(target, name);
|
|
instr->mask = NO_CLOBBER;
|
|
int param_count = 0;
|
|
while (args && args[0] != 0)
|
|
{
|
|
ASSERT(param_count <= MAX_ASM_INSTRUCTION_PARAMS);
|
|
instr->param[param_count++] = decode_arg_type(&args);
|
|
}
|
|
instr->param_count = param_count;
|
|
}
|
|
|
|
INLINE void reg_register(PlatformTarget *target, const char *name, AsmRegisterType reg_type, AsmArgBits bits, unsigned clobber_id)
|
|
{
|
|
TokenType token_type = TOKEN_CT_IDENT;
|
|
unsigned len = (unsigned)strlen(name);
|
|
const char *interned = symtab_add(name, len, fnv1a(name, len), &token_type);
|
|
uint32_t hash = ASM_PTR_HASH(interned);
|
|
uint32_t slot = hash & ASM_REGISTER_MASK;
|
|
while (1)
|
|
{
|
|
AsmRegister *reg = &target->registers[slot];
|
|
if (!reg->name)
|
|
{
|
|
*reg = (AsmRegister) { .name = interned, .type = reg_type, .bits = bits, .clobber_index = clobber_id };
|
|
target->register_count++;
|
|
return;
|
|
}
|
|
slot = (slot + 1) & ASM_REGISTER_MASK;
|
|
}
|
|
}
|
|
|
|
INLINE void reg_register_list(PlatformTarget *target, const char **names, unsigned count, AsmRegisterType param, unsigned bitsize, unsigned first_clobber)
|
|
{
|
|
for (unsigned i = 0; i < count; i++) reg_register(target, names[i], param, bitsize, i + first_clobber);
|
|
}
|
|
|
|
AsmInstruction *asm_instr_by_name(const char *name)
|
|
{
|
|
uint32_t hash = ASM_PTR_HASH(name);
|
|
uint32_t slot = hash & ASM_INSTRUCTION_MASK;
|
|
while (1)
|
|
{
|
|
AsmInstruction *inst = &compiler.platform.instructions[slot];
|
|
if (inst->name == name) return inst;
|
|
if (inst->name == NULL) return NULL;
|
|
slot = (slot + 1) & ASM_INSTRUCTION_MASK;
|
|
}
|
|
}
|
|
|
|
AsmRegister *asm_reg_by_name(PlatformTarget *target, const char *name)
|
|
{
|
|
uint32_t hash = ASM_PTR_HASH(name);
|
|
uint32_t slot = hash & ASM_REGISTER_MASK;
|
|
while (1)
|
|
{
|
|
AsmRegister *reg = &target->registers[slot];
|
|
const char *reg_name = reg->name;
|
|
if (reg_name == name) return reg;
|
|
if (!reg_name) return NULL;
|
|
slot = (slot + 1) & ASM_REGISTER_MASK;
|
|
}
|
|
}
|
|
|
|
|
|
static void init_asm_aarch64(PlatformTarget *target)
|
|
{
|
|
Clobbers cc_flag_mask = clobbers_make(AARCH64_CC, -1);
|
|
target->clobber_name_list = Aarch64ClobberNames;
|
|
target->extra_clobbers = NULL;
|
|
reg_instr(target, "abs", "w:r32/r64, r32/r64");
|
|
reg_instr(target, "adc", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr_clob(target, "adcs", cc_flag_mask, "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "add", "w:r32/r64, r32/r64, immu12/r32/r64");
|
|
reg_instr_clob(target, "adds", cc_flag_mask, "w:r32/r64, r32/r64, immu12/r32/r64");
|
|
reg_instr(target, "adr", "w:r64, imm32");
|
|
reg_instr(target, "adrp", "w:r64, imm64");
|
|
reg_instr(target, "and", "w:r32/r64, r32/r64, immu32/r32/r64");
|
|
reg_instr_clob(target, "ands", cc_flag_mask, "w:r32/r64, r32/r64, immu32/r32/r64");
|
|
reg_instr(target, "asr", "w:r32/r64, r32/r64, immu12/r32/r64");
|
|
reg_instr(target, "b", "imm64");
|
|
reg_instr(target, "beq", "imm64");
|
|
reg_instr(target, "bne", "imm64");
|
|
reg_instr(target, "bcs", "imm64");
|
|
reg_instr(target, "bcc", "imm64");
|
|
reg_instr(target, "bge", "imm64");
|
|
reg_instr(target, "blt", "imm64");
|
|
reg_instr(target, "bgt", "imm64");
|
|
reg_instr(target, "ble", "imm64");
|
|
reg_instr(target, "bfi", "w:r32/r64, r32/r64, immu12, immu12");
|
|
reg_instr(target, "bfm", "w:r32/r64, r32/r64, immu12, immu12");
|
|
reg_instr(target, "bfxil", "w:r32/r64, r32/r64, immu12, immu12");
|
|
reg_instr(target, "bic", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr_clob(target, "bics", cc_flag_mask, "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "bl", "imm64");
|
|
reg_instr(target, "blr", "r64");
|
|
reg_instr(target, "br", "r64");
|
|
reg_instr(target, "brk", "immu16");
|
|
reg_instr(target, "cbnz", "r32/r64, imm64");
|
|
reg_instr(target, "cbz", "r32/r64, imm64");
|
|
reg_instr_clob(target, "cfinv", cc_flag_mask, NULL);
|
|
reg_instr(target, "cls", "w:r32/r64, r32/r64");
|
|
reg_instr(target, "clz", "w:r32/r64, r32/r64");
|
|
reg_instr_clob(target, "cmn", cc_flag_mask, "r32/r64, r32/r64/immu12");
|
|
reg_instr_clob(target, "cmp", cc_flag_mask, "r32/r64, r32/r64/immu12");
|
|
reg_instr(target, "cnt", "r32/r64, r32/r64");
|
|
reg_instr(target, "ctz", "r32/r64, r32/r64");
|
|
reg_instr(target, "eon", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "eor", "w:r32/r64, r32/r64, immu16/r32/r64");
|
|
reg_instr(target, "extr", "w:r32/r64, r32/r64, r32/r64, immu8");
|
|
reg_instr(target, "hlt", "immu16");
|
|
reg_instr(target, "ldar", "w:r32/r64, mem");
|
|
reg_instr(target, "ldarb", "w:r32, mem");
|
|
reg_instr(target, "ldarh", "w:r32, mem");
|
|
reg_instr(target, "ldaxp", "w:r32/r64, w:r32/r64, mem");
|
|
reg_instr(target, "ldaxr", "w:r32/r64, mem");
|
|
reg_instr(target, "ldaxrb", "w:r32, mem");
|
|
reg_instr(target, "ldaxrh", "w:r32, mem");
|
|
reg_instr(target, "ldr", "w:r32/r64, mem");
|
|
reg_instr(target, "ldrb", "w:r32, mem");
|
|
reg_instr(target, "ldrsb", "w:r32/r64, mem");
|
|
reg_instr(target, "ldrh", "w:r32, mem");
|
|
reg_instr(target, "ldrsh", "w:r32/r64, mem");
|
|
reg_instr(target, "ldrsw", "w:r64, mem");
|
|
reg_instr(target, "ldp", "w:r32/r64, w:r32/r64, mem");
|
|
reg_instr(target, "ldpsw", "w:r64, w:r64, mem");
|
|
reg_instr(target, "ldxp", "w:r32/r64, w:r32/r64, mem");
|
|
reg_instr(target, "ldxr", "w:r32/r64, mem");
|
|
reg_instr(target, "ldxrb", "w:r32, mem");
|
|
reg_instr(target, "ldxrh", "w:r32, mem");
|
|
reg_instr(target, "lsl", "w:r32/r64, r32/r64, immu8/r32/r64");
|
|
reg_instr(target, "lslv", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "lsr", "w:r32/r64, r32/r64, immu8/r32/r64");
|
|
reg_instr(target, "lsrv", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "madd", "w:r32/r64, r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "mneg", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "mov", "w:r32/r64, r32/r64/imm64");
|
|
reg_instr(target, "movk", "w:r32/r64, immu16");
|
|
reg_instr(target, "movn", "w:r32/r64, immu16");
|
|
reg_instr(target, "movz", "w:r32/r64, immu16");
|
|
reg_instr(target, "msub", "w:r32/r64, r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "mul", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "mvn", "w:r32/r64, r32/r64");
|
|
reg_instr(target, "neg", "w:r32/r64, r32/r64");
|
|
reg_instr_clob(target, "negs", cc_flag_mask, "w:r32/r64, r32/r64");
|
|
reg_instr(target, "ngc", "w:r32/r64, r32/r64");
|
|
reg_instr_clob(target, "ngcs", cc_flag_mask, "w:r32/r64, r32/r64");
|
|
reg_instr(target, "nop", NULL);
|
|
reg_instr(target, "orn", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "orr", "w:r32/r64, r32/r64, r32/r64/imm64");
|
|
reg_instr(target, "rbit", "w:r32/r64, r32/r64");
|
|
reg_instr(target, "ret", NULL);
|
|
reg_instr(target, "rev", "w:r32/r64, r32/r64");
|
|
reg_instr(target, "rev16", "w:r32/r64, r32/r64");
|
|
reg_instr(target, "rev32", "w:r64, r64");
|
|
reg_instr(target, "ror", "w:r32/r64, r32/r64, r32/r64/immu8");
|
|
reg_instr(target, "rorv", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "sbc", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr_clob(target, "sbcs", cc_flag_mask, "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "sbfiz", "w:r32/r64, r32/r64, immu8, immu8");
|
|
reg_instr(target, "sbfm", "w:r32/r64, r32/r64, immu8, immu8");
|
|
reg_instr(target, "sbfx", "w:r32/r64, r32/r64, immu8, immu8");
|
|
reg_instr(target, "sdiv", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "smaddl", "w:r64, r32, r32, r64");
|
|
reg_instr(target, "smax", "w:r32/r64, r32/r64, r32/r64/imm8");
|
|
reg_instr(target, "smin", "w:r32/r64, r32/r64, r32/r64/imm8");
|
|
reg_instr(target, "smnegl", "w:r64, r32, r32");
|
|
reg_instr(target, "smsubl", "w:r64, r32, r32, r64");
|
|
reg_instr(target, "smulh", "w:r64, r64, r64");
|
|
reg_instr(target, "smull", "w:r64, r32, r32");
|
|
reg_instr(target, "stlr", "r32/r64, w:mem");
|
|
reg_instr(target, "stlrb", "r32, w:mem");
|
|
reg_instr(target, "stlrh", "r32, w:mem");
|
|
reg_instr(target, "stlxp", "r32, r32/r64, r32/r64, w:mem");
|
|
reg_instr(target, "stlxr", "r32, r32/r64, w:mem");
|
|
reg_instr(target, "stlxrb", "r32, r32, w:mem");
|
|
reg_instr(target, "stlxrh", "r32, r32, w:mem");
|
|
reg_instr(target, "stp", "r32/r64, r32/r64, w:mem");
|
|
reg_instr(target, "str", "r32/r64, w:mem");
|
|
reg_instr(target, "strb", "r32, w:mem");
|
|
reg_instr(target, "strh", "r32, w:mem");
|
|
reg_instr(target, "stxp", "r32, r32/r64, r32/r64, w:mem");
|
|
reg_instr(target, "stxr", "r32, r32/r64, w:mem");
|
|
reg_instr(target, "stxrb", "r32, r32, w:mem");
|
|
reg_instr(target, "stxrh", "r32, r32, w:mem");
|
|
reg_instr(target, "sub", "w:r32/r64, r32/r64, immu12/r32/r64");
|
|
reg_instr_clob(target, "subs", cc_flag_mask, "w:r32/r64, r32/r64, immu12/r32/r64");
|
|
reg_instr(target, "svc", "immu16");
|
|
reg_instr(target, "sxtb", "w:r32/r64, r32");
|
|
reg_instr(target, "sxth", "w:r32/r64, r32");
|
|
reg_instr(target, "sxtw", "w:r64, r32");
|
|
reg_instr_clob(target, "tst", cc_flag_mask, "r32/r64, r32/r64/imm64");
|
|
reg_instr(target, "ubfiz", "w:r32/r64, r32/r64, immu8, immu8");
|
|
reg_instr(target, "ubfm", "w:r32/r64, r32/r64, immu8, immu8");
|
|
reg_instr(target, "ubfx", "w:r32/r64, r32/r64, immu8, immu8");
|
|
reg_instr(target, "udiv", "w:r32/r64, r32/r64, r32/r64");
|
|
reg_instr(target, "umaddl", "w:r64, r32, r32, r64");
|
|
reg_instr(target, "umax", "w:r32/r64, r32/r64, r32/r64/imm8");
|
|
reg_instr(target, "umin", "w:r32/r64, r32/r64, r32/r64/imm8");
|
|
reg_instr(target, "umnegl", "w:r64, r32, r32");
|
|
reg_instr(target, "umsubl", "w:r64, r32, r32, r64");
|
|
reg_instr(target, "umulh", "w:r64, r64, r64");
|
|
reg_instr(target, "umull", "w:r64, r32, r32");
|
|
reg_instr(target, "uxtb", "w:r32, r32");
|
|
reg_instr(target, "uxth", "w:r32, r32");
|
|
|
|
reg_register_list(target, aarch64_quad_regs, 32, ASM_REG_INT, ARG_BITS_64, AARCH64_R0);
|
|
reg_register_list(target, aarch64_long_regs, 32, ASM_REG_INT, ARG_BITS_32, AARCH64_R0);
|
|
reg_register_list(target, aarch64_f128_regs, 32, ASM_REG_FLOAT, ARG_BITS_128, AARCH64_Q0);
|
|
reg_register_list(target, aarch64_double_regs, 32, ASM_REG_FLOAT, ARG_BITS_64, AARCH64_Q0);
|
|
reg_register_list(target, aarch64_float_regs, 32, ASM_REG_FLOAT, ARG_BITS_32, AARCH64_Q0);
|
|
reg_register_list(target, aarch64_f16_regs, 32, ASM_REG_FLOAT, ARG_BITS_16, AARCH64_Q0);
|
|
reg_register_list(target, aarch64_f8_regs, 32, ASM_REG_FLOAT, ARG_BITS_8, AARCH64_Q0);
|
|
reg_register_list(target, aarch64_v8b_regs, 32, ASM_REG_IVEC, ARG_BITS_64, AARCH64_FIRST_RV_CLOBBER);
|
|
reg_register_list(target, aarch64_v16b_regs, 32, ASM_REG_IVEC, ARG_BITS_128, AARCH64_FIRST_RV_CLOBBER);
|
|
reg_register_list(target, aarch64_v4h_regs, 32, ASM_REG_IVEC, ARG_BITS_64, AARCH64_FIRST_RV_CLOBBER);
|
|
reg_register_list(target, aarch64_v8h_regs, 32, ASM_REG_IVEC, ARG_BITS_128, AARCH64_FIRST_RV_CLOBBER);
|
|
reg_register_list(target, aarch64_v2s_regs, 32, ASM_REG_IVEC, ARG_BITS_64, AARCH64_FIRST_RV_CLOBBER);
|
|
reg_register_list(target, aarch64_v4s_regs, 32, ASM_REG_IVEC, ARG_BITS_128, AARCH64_FIRST_RV_CLOBBER);
|
|
reg_register_list(target, aarch64_v1d_regs, 32, ASM_REG_IVEC, ARG_BITS_64, AARCH64_FIRST_RV_CLOBBER);
|
|
reg_register_list(target, aarch64_v2d_regs, 32, ASM_REG_IVEC, ARG_BITS_128, AARCH64_FIRST_RV_CLOBBER);
|
|
reg_register(target, "$sp", ASM_REG_INT, ARG_BITS_64, AARCH64_R31);
|
|
}
|
|
|
|
static void init_asm_wasm(PlatformTarget *target)
|
|
{
|
|
error_exit("WASM asm not complete.");
|
|
}
|
|
|
|
static void init_asm_arm(PlatformTarget *target)
|
|
{
|
|
error_exit("ARM asm not complete.");
|
|
}
|
|
|
|
static void init_asm_riscv(PlatformTarget *target)
|
|
{
|
|
target->clobber_name_list = RISCVClobberNames;
|
|
target->extra_clobbers = NULL;
|
|
unsigned int bits = 0;
|
|
switch(target->arch) {
|
|
case ARCH_TYPE_RISCV64:
|
|
// math
|
|
reg_instr(target, "add", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "sub", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "addi", "w:r64/mem, r64/mem, immi12");
|
|
reg_instr(target, "neg", "w:r64/mem, r64/mem");
|
|
// bit
|
|
reg_instr(target, "and", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "or", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "xor", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "not", "w:r64/mem, r64/mem");
|
|
reg_instr(target, "andi", "w:r64/mem, r64/mem, immi12");
|
|
reg_instr(target, "ori", "w:r64/mem, r64/mem, immi12");
|
|
reg_instr(target, "xori", "w:r64/mem, r64/mem, immi12");
|
|
// shift
|
|
reg_instr(target, "sll", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "srl", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "sra", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "slli", "w:r64/mem, r64/mem, immu5");
|
|
reg_instr(target, "srli", "w:r64/mem, r64/mem, immu5");
|
|
reg_instr(target, "srai", "w:r64/mem, r64/mem, immu5");
|
|
// load
|
|
reg_instr(target, "li", "w:r64/mem, immi64");
|
|
reg_instr(target, "lui", "w:r64, immu20");
|
|
reg_instr(target, "auipc", "w:r64, immu20");
|
|
reg_instr(target, "mv", "w:r64/mem, r64/mem");
|
|
reg_instr(target, "ld", "w:r64/mem, mem");
|
|
reg_instr(target, "lw", "w:r64/mem, mem");
|
|
reg_instr(target, "lh", "w:r64/mem, mem");
|
|
reg_instr(target, "lhu", "w:r64/mem, mem");
|
|
reg_instr(target, "lb", "w:r64/mem, mem");
|
|
reg_instr(target, "lbu", "w:r64/mem, mem");
|
|
// store
|
|
reg_instr(target, "sd", "r64/mem, w:mem");
|
|
reg_instr(target, "sw", "r64/mem, w:mem");
|
|
reg_instr(target, "sh", "r64/mem, w:mem");
|
|
reg_instr(target, "sb", "r64/mem, w:mem");
|
|
// Misc
|
|
reg_instr(target, "nop", NULL);
|
|
reg_instr(target, "ebreak", NULL);
|
|
reg_instr(target, "ecall", NULL);
|
|
reg_instr(target, "eret", NULL);
|
|
// Jump
|
|
reg_instr(target, "j", "immi20");
|
|
reg_instr(target, "jal", "w:r64/mem, immi20");
|
|
reg_instr(target, "jalr", "w:r64/mem, r64/mem, immi20");
|
|
reg_instr(target, "ret", NULL);
|
|
// Set
|
|
reg_instr(target, "slt", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "slti", "w:r64/mem, r64/mem, immi12");
|
|
reg_instr(target, "sltu", "w:r64/mem, r64/mem, r64/mem");
|
|
reg_instr(target, "sltiu", "w:r64/mem, r64/mem, immu12");
|
|
reg_instr(target, "seqz", "w:r64/mem, r64/mem");
|
|
reg_instr(target, "snez", "w:r64/mem, r64/mem");
|
|
reg_instr(target, "sltz", "w:r64/mem, r64/mem");
|
|
reg_instr(target, "sgtz", "w:r64/mem, r64/mem");
|
|
// CSR
|
|
reg_instr(target, "csrw", "w:r64, r64/mem");
|
|
reg_instr(target, "csrr", "w:r64/mem, r64");
|
|
reg_instr(target, "csrrw", "w:r64/mem, rw:r64, r64/mem");
|
|
reg_instr(target, "csrrs", "w:r64/mem, rw:r64, r64/mem");
|
|
reg_instr(target, "csrrc", "w:r64/mem, rw:r64, r64/mem");
|
|
reg_instr(target, "csrrwi", "w:r64/mem, rw:r64, immu5");
|
|
reg_instr(target, "csrrsi", "w:r64/mem, rw:r64, immu5");
|
|
reg_instr(target, "csrrci", "w:r64/mem, rw:r64, immu5");
|
|
// Interrupt
|
|
reg_instr(target, "wfi", NULL);
|
|
reg_instr(target, "mret", NULL);
|
|
|
|
bits = ARG_BITS_64;
|
|
break;
|
|
case ARCH_TYPE_RISCV32:
|
|
// math
|
|
reg_instr(target, "add", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "sub", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "addi", "w:r32/mem, r32/mem, immi12");
|
|
reg_instr(target, "neg", "w:r32/mem, r32/mem");
|
|
// bit
|
|
reg_instr(target, "and", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "or", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "xor", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "not", "w:r32/mem, r32/mem");
|
|
reg_instr(target, "andi", "w:r32/mem, r32/mem, immi12");
|
|
reg_instr(target, "ori", "w:r32/mem, r32/mem, immi12");
|
|
reg_instr(target, "xori", "w:r32/mem, r32/mem, immi12");
|
|
// shift
|
|
reg_instr(target, "sll", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "srl", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "sra", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "slli", "w:r32/mem, r32/mem, immu5");
|
|
reg_instr(target, "srli", "w:r32/mem, r32/mem, immu5");
|
|
reg_instr(target, "srai", "w:r32/mem, r32/mem, immu5");
|
|
// load
|
|
reg_instr(target, "li", "w:r32/mem, immi32");
|
|
reg_instr(target, "lui", "w:r32, immu20");
|
|
reg_instr(target, "auipc", "w:r32, immu20");
|
|
reg_instr(target, "mv", "w:r32/mem, r32/mem");
|
|
reg_instr(target, "lw", "w:r32/mem, mem");
|
|
reg_instr(target, "lh", "w:r32/mem, mem");
|
|
reg_instr(target, "lhu", "w:r32/mem, mem");
|
|
reg_instr(target, "lb", "w:r32/mem, mem");
|
|
reg_instr(target, "lbu", "w:r32/mem, mem");
|
|
// store
|
|
reg_instr(target, "sw", "r32/mem, w:mem");
|
|
reg_instr(target, "sh", "r32/mem, w:mem");
|
|
reg_instr(target, "sb", "r32/mem, w:mem");
|
|
// Misc
|
|
reg_instr(target, "nop", NULL);
|
|
reg_instr(target, "ebreak", NULL);
|
|
reg_instr(target, "ecall", NULL);
|
|
reg_instr(target, "eret", NULL);
|
|
// Jump
|
|
reg_instr(target, "j", "immi20");
|
|
reg_instr(target, "jal", "w:r32/mem, immi20");
|
|
reg_instr(target, "jalr", "w:r32/mem, r32/mem, immi20");
|
|
reg_instr(target, "ret", NULL);
|
|
// Set
|
|
reg_instr(target, "slt", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "slti", "w:r32/mem, r32/mem, immi12");
|
|
reg_instr(target, "sltu", "w:r32/mem, r32/mem, r32/mem");
|
|
reg_instr(target, "sltiu", "w:r32/mem, r32/mem, immu12");
|
|
reg_instr(target, "seqz", "w:r32/mem, r32/mem");
|
|
reg_instr(target, "snez", "w:r32/mem, r32/mem");
|
|
reg_instr(target, "sltz", "w:r32/mem, r32/mem");
|
|
reg_instr(target, "sgtz", "w:r32/mem, r32/mem");
|
|
// CSR
|
|
reg_instr(target, "csrw", "w:r32, r32/mem");
|
|
reg_instr(target, "csrr", "w:r32/mem, r32");
|
|
reg_instr(target, "csrrw", "w:r32/mem, rw:r32, r32/mem");
|
|
reg_instr(target, "csrrs", "w:r32/mem, rw:r32, r32/mem");
|
|
reg_instr(target, "csrrc", "w:r32/mem, rw:r32, r32/mem");
|
|
reg_instr(target, "csrrwi", "w:r32/mem, rw:r32, immu5");
|
|
reg_instr(target, "csrrsi", "w:r32/mem, rw:r32, immu5");
|
|
reg_instr(target, "csrrci", "w:r32/mem, rw:r32, immu5");
|
|
// Interrupt
|
|
reg_instr(target, "wfi", NULL);
|
|
reg_instr(target, "mret", NULL);
|
|
|
|
bits = ARG_BITS_32;
|
|
break;
|
|
default:
|
|
UNREACHABLE
|
|
}
|
|
reg_register_list(target, riscv_gp_integer_regs, 32, ASM_REG_INT, bits, RISCV_X0);
|
|
reg_register_list(target, riscv_arg_integer_regs, 8, ASM_REG_INT, bits, RISCV_X10);
|
|
reg_register_list(target, riscv_temp_integer_regs, 3, ASM_REG_INT, bits, RISCV_X5);
|
|
reg_register_list(target, &riscv_temp_integer_regs[3], 4, ASM_REG_INT, bits, RISCV_X28);
|
|
reg_register_list(target, riscv_save_integer_regs, 2, ASM_REG_INT, bits, RISCV_X8);
|
|
reg_register_list(target, &riscv_save_integer_regs[2], 10, ASM_REG_INT, bits, RISCV_X18);
|
|
reg_register_list(target, riscv_machine_integer_regs, 3, ASM_REG_INT, bits, RISCV_MIE);
|
|
reg_register(target, "$ra", ASM_REG_INT, bits, RISCV_X1);
|
|
reg_register(target, "$sp", ASM_REG_INT, bits, RISCV_X2);
|
|
reg_register(target, "$gp", ASM_REG_INT, bits, RISCV_X3);
|
|
reg_register(target, "$tp", ASM_REG_INT, bits, RISCV_X4);
|
|
reg_register(target, "$zero", ASM_REG_INT, bits, RISCV_X0);
|
|
}
|
|
|
|
static void init_asm_ppc(PlatformTarget *target)
|
|
{
|
|
error_exit("PPC asm not complete.");
|
|
}
|
|
|
|
static void init_asm_x86(PlatformTarget* target)
|
|
{
|
|
Clobbers rax_mask = clobbers_make(X86_RAX, -1);
|
|
Clobbers cc_flag_mask = clobbers_make(X86_CC, -1);
|
|
Clobbers rax_cc_mask = clobbers_make_from(cc_flag_mask, X86_RAX, -1);
|
|
Clobbers rax_rdx_cc_mask = clobbers_make_from(cc_flag_mask, X86_RAX, X86_RDX, -1);
|
|
bool is_x64 = target->arch == ARCH_TYPE_X86_64;
|
|
if (!is_x64)
|
|
{
|
|
reg_instr_clob(target, "aaa", rax_mask, 0);
|
|
reg_instr_clob(target, "into", cc_flag_mask, NULL);
|
|
reg_instr(target, "pushl", "r32/mem/imm32");
|
|
reg_instr(target, "popl", "w:r32/mem/imm32");
|
|
}
|
|
if (is_x64)
|
|
{
|
|
reg_instr_clob(target, "syscall", clobbers_make_from(cc_flag_mask, X86_RAX, X86_R11, X86_RCX, -1), NULL);
|
|
reg_instr(target, "pushq", "r64/mem");
|
|
reg_instr(target, "popq", "w:r64/mem");
|
|
}
|
|
reg_instr_clob(target, "adcb", cc_flag_mask, "rw:r8/mem, r8/mem/imm8");
|
|
reg_instr_clob(target, "adcw", cc_flag_mask, "rw:r16/mem, r16/mem/imm16/immi8");
|
|
reg_instr_clob(target, "adcl", cc_flag_mask, "rw:r32/mem, r32/mem/imm32/immi8");
|
|
reg_instr_clob(target, "adcq", cc_flag_mask, "rw:r64/mem, r64/mem/immi32/immi8");
|
|
|
|
reg_instr_clob(target, "adcxl", cc_flag_mask, "r32, rw:r32/mem");
|
|
reg_instr_clob(target, "adcxq", cc_flag_mask, "r64, rw:r64/mem");
|
|
|
|
reg_instr_clob(target, "addb", cc_flag_mask, "rw:r8/mem, r8/mem/imm8");
|
|
reg_instr_clob(target, "addw", cc_flag_mask, "rw:r16/mem, r16/mem/imm16/immi8");
|
|
reg_instr_clob(target, "addl", cc_flag_mask, "rw:r32/mem, r32/mem/imm32/immi8");
|
|
reg_instr_clob(target, "addq", cc_flag_mask, "rw:r64/mem, r64/mem/immi32/immi8");
|
|
|
|
reg_instr(target, "addpd", "rw:v128, v128/mem");
|
|
reg_instr(target, "addps", "rw:v128, v128/mem");
|
|
reg_instr(target, "addsd", "rw:v128, v128/mem");
|
|
reg_instr(target, "addss", "rw:v128, v128/mem");
|
|
reg_instr(target, "vaddpd", "w:v128/v256/v512, v128/v256/v512, v128/v256/v512/mem");
|
|
reg_instr(target, "vaddps", "w:v128/v256/v512, v128/v256/v512, v128/v256/v512/mem");
|
|
reg_instr(target, "vaddsd", "w:v128, v128, v128/mem");
|
|
reg_instr(target, "vaddss", "w:v128, v128, v128/mem");
|
|
|
|
reg_instr_clob(target, "cbtw", rax_mask, NULL);
|
|
reg_instr_clob(target, "cwtl", rax_mask, NULL);
|
|
reg_instr_clob(target, "cltq", rax_mask, NULL);
|
|
reg_instr_clob(target, "clc", rax_mask, NULL);
|
|
reg_instr_clob(target, "cld", rax_mask, NULL);
|
|
reg_instr(target, "clflush", "mem");
|
|
reg_instr(target, "movb", "w:r8/mem, r8/mem/imm8");
|
|
reg_instr(target, "movsbw", "w:r16/mem, r8/mem");
|
|
reg_instr(target, "movzbw", "w:r16/mem, r8/mem");
|
|
reg_instr(target, "movsbl", "w:r32/mem, r8/mem");
|
|
reg_instr(target, "movzbl", "w:r32/mem, r8/mem");
|
|
reg_instr(target, "movsbq", "w:r64/mem, r8/mem");
|
|
reg_instr(target, "movzbq", "w:r64/mem, r8/mem");
|
|
reg_instr(target, "movw", "w:r16/mem, r16/mem/imm16"); // Add seg
|
|
reg_instr(target, "movswl", "w:r32/mem, r16/mem");
|
|
reg_instr(target, "movzwl", "w:r32/mem, r16/mem");
|
|
reg_instr(target, "movswq", "w:r64/mem, r16/mem");
|
|
reg_instr(target, "movzwq", "w:r64/mem, r16/mem");
|
|
reg_instr(target, "movl", "w:r32/mem, r32/mem/imm32");
|
|
reg_instr(target, "movslq", "w:r64/mem, r32/mem");
|
|
reg_instr(target, "movzlq", "w:r64/mem, r32/mem");
|
|
reg_instr(target, "movq", "w:r64/mem, r64/mem/immi32/imm64"); // Seg
|
|
//reg_instr("mov", "wAARG_SEG | AARG_W, AARG_R16 | AARG_R64, 0); // Seg
|
|
/*
|
|
reg_instr("movabsb", AARG_R8 | AARG_IMM64, AARG_R8 | AARG_IMM64, 0); // Missing segment
|
|
reg_instr("movawsw", AARG_R16 | AARG_IMM64, AARG_R16 | AARG_IMM64, 0); // Missing segment
|
|
reg_instr("movalsl", AARG_R32 | AARG_IMM64, AARG_R32 | AARG_IMM64, 0); // Missing segment
|
|
reg_instr("movaqsq", AARG_R64 | AARG_IMM64, AARG_R64 | AARG_IMM64, 0); // Missing segment*/
|
|
reg_instr(target, "nop", NULL);
|
|
reg_instr(target, "nopw", NULL);
|
|
reg_instr(target, "nopl", NULL);
|
|
reg_instr_clob(target, "orb", cc_flag_mask, "rw:r8/mem, r8/mem/imm8");
|
|
reg_instr_clob(target, "orw", cc_flag_mask, "rw:r16/mem, r16/mem/imm16");
|
|
reg_instr_clob(target, "orl", cc_flag_mask, "rw:r32/mem, r32/mem/imm32");
|
|
reg_instr_clob(target, "orq", cc_flag_mask, "rw:r64/mem, r64/mem/immi32/imm64");
|
|
reg_instr_clob(target, "negb", cc_flag_mask, "rw:r8/mem");
|
|
reg_instr_clob(target, "negw", cc_flag_mask, "rw:r16/mem");
|
|
reg_instr_clob(target, "negl", cc_flag_mask, "rw:r32/mem");
|
|
reg_instr_clob(target, "negq", cc_flag_mask, "rw:r64/mem");
|
|
reg_instr(target, "notb", "rw:r8/mem");
|
|
reg_instr(target, "notw", "rw:r16/mem");
|
|
reg_instr(target, "notl", "rw:r32/mem");
|
|
reg_instr(target, "notq", "rw:r64/mem");
|
|
reg_instr_clob(target, "xorb", cc_flag_mask, "rw:r8/mem, r8/mem/imm8");
|
|
reg_instr_clob(target, "xorw", cc_flag_mask, "rw:r16/mem, r16/mem/imm16");
|
|
reg_instr_clob(target, "xorl", cc_flag_mask, "rw:r32/mem, r32/mem/imm32");
|
|
reg_instr_clob(target, "xorq", cc_flag_mask, "rw:r64/mem, r64/mem/immi32/imm64");
|
|
reg_instr_clob(target, "mulb", rax_cc_mask, "r8/mem");
|
|
reg_instr_clob(target, "mulw", rax_rdx_cc_mask, "r16/mem");
|
|
reg_instr_clob(target, "mull", rax_rdx_cc_mask, "r32/mem");
|
|
reg_instr_clob(target, "mulq", rax_rdx_cc_mask, "rw:r64/mem");
|
|
reg_instr_clob(target, "subb", rax_cc_mask, "rw:r8/mem, r8/mem/imm8");
|
|
reg_instr_clob(target, "subw", rax_cc_mask, "rw:r16/mem, r16/mem/imm16");
|
|
reg_instr_clob(target, "subl", rax_cc_mask, "rw:r32/mem, r32/mem/imm32");
|
|
reg_instr_clob(target, "subq", rax_cc_mask, "rw:r64/mem, r64/mem/immi32/imm64");
|
|
reg_instr_clob(target, "cpuid", clobbers_make_from(cc_flag_mask, X86_RAX, X86_RBX, X86_RCX, X86_RDX, -1), NULL);
|
|
reg_instr(target, "hlt", NULL);
|
|
reg_instr(target, "in", "w:r8/r16/r32, r16/imm8"); // Actually ensure reg_al_ax and dx
|
|
reg_instr_clob(target, "incb", cc_flag_mask, "rw:r8/mem");
|
|
reg_instr_clob(target, "incw", cc_flag_mask, "rw:r16/mem");
|
|
reg_instr_clob(target, "incl", cc_flag_mask, "rw:r32/mem");
|
|
reg_instr_clob(target, "incq", cc_flag_mask, "rw:r64/mem");
|
|
reg_instr(target, "insb", NULL);
|
|
reg_instr(target, "insw", NULL);
|
|
reg_instr(target, "insl", NULL);
|
|
reg_instr_clob(target, "int", cc_flag_mask, "imm8");
|
|
reg_instr_clob(target, "int3", cc_flag_mask, NULL);
|
|
reg_instr_clob(target, "int1", cc_flag_mask, NULL);
|
|
reg_instr(target, "invd", NULL);
|
|
reg_instr(target, "invpcid", "r32/r64, mem");
|
|
reg_instr(target, "invlpg", "w:mem");
|
|
reg_instr(target, "invlpga", "r32, r64"); // c, a check this one!
|
|
reg_instr(target, "iret", NULL);
|
|
reg_instr(target, "iretl", NULL);
|
|
reg_instr(target, "iretw", NULL);
|
|
reg_instr(target, "iretq", NULL);
|
|
reg_instr_clob(target, "rdtsc", clobbers_make_from(rax_mask, X86_RDX, -1), NULL);
|
|
reg_instr_clob(target, "rdtscp", clobbers_make_from(rax_mask, X86_RDX, X86_RCX, -1), NULL);
|
|
reg_instr(target, "ret", NULL);
|
|
reg_instr(target, "push", "imm8");
|
|
reg_instr(target, "pushw", "r16/mem/imm16");
|
|
reg_instr(target, "popw", "w:r16/mem");
|
|
|
|
target->clobber_name_list = X86ClobberNames;
|
|
target->extra_clobbers = "~{flags},~{dirflag},~{fspr}";
|
|
if (target->arch == ARCH_TYPE_X86)
|
|
{
|
|
reg_register_list(target, x86_long_regs, 8, ASM_REG_INT, ARG_BITS_32, X86_RAX);
|
|
reg_register_list(target, x86_word_regs, 8, ASM_REG_INT, ARG_BITS_16, X86_RAX);
|
|
reg_register_list(target, x86_low_byte_regs, 8, ASM_REG_INT, ARG_BITS_8, X86_RAX);
|
|
reg_register_list(target, x86_float_regs, 8, ASM_REG_FLOAT, ARG_BITS_80, X86_ST0);
|
|
reg_register_list(target, x86_xmm_regs, 8, ASM_REF_FVEC, ARG_BITS_128, X86_MM0);
|
|
}
|
|
else
|
|
{
|
|
reg_register_list(target, x64_quad_regs, 16, ASM_REG_INT, ARG_BITS_64, X86_RAX);
|
|
reg_register_list(target, x86_long_regs, 16, ASM_REG_INT, ARG_BITS_32, X86_RAX);
|
|
reg_register_list(target, x86_word_regs, 16, ASM_REG_INT, ARG_BITS_16, X86_RAX);
|
|
reg_register_list(target, x86_low_byte_regs, 16, ASM_REG_INT, ARG_BITS_8, X86_RAX);
|
|
reg_register_list(target, x86_high_byte_regs, 4, ASM_REG_INT, ARG_BITS_8, X86_RAX);
|
|
reg_register_list(target, x86_xmm_regs, 16, ASM_REF_FVEC, ARG_BITS_128, X86_XMM0);
|
|
reg_register_list(target, x86_ymm_regs, 16, ASM_REF_FVEC, ARG_BITS_256, X86_XMM0);
|
|
reg_register_list(target, x86_zmm_regs, 16, ASM_REF_FVEC, ARG_BITS_512, X86_XMM0);
|
|
}
|
|
}
|
|
|
|
bool asm_is_supported(ArchType arch)
|
|
{
|
|
switch (arch)
|
|
{
|
|
case ARCH_TYPE_X86:
|
|
case ARCH_TYPE_X86_64:
|
|
case ARCH_TYPE_RISCV32:
|
|
case ARCH_TYPE_RISCV64:
|
|
case ARCH_TYPE_AARCH64:
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static void scratch_append_bit(const char *name, const char *string, bool *has_print_reg)
|
|
{
|
|
if (*has_print_reg)
|
|
{
|
|
scratch_buffer_append("/");
|
|
}
|
|
else
|
|
{
|
|
*has_print_reg = true;
|
|
}
|
|
scratch_buffer_append(name);
|
|
scratch_buffer_append(string);
|
|
}
|
|
|
|
static void scratch_append_bits(const char *name, AsmArgBits bits)
|
|
{
|
|
bool has_print = false;
|
|
if (bits & ARG_BITS_512) scratch_append_bit(name, "512", &has_print);
|
|
if (bits & ARG_BITS_256) scratch_append_bit(name, "256", &has_print);
|
|
if (bits & ARG_BITS_128) scratch_append_bit(name, "128", &has_print);
|
|
if (bits & ARG_BITS_80) scratch_append_bit(name, "80", &has_print);
|
|
if (bits & ARG_BITS_64) scratch_append_bit(name, "64", &has_print);
|
|
if (bits & ARG_BITS_32) scratch_append_bit(name, "32", &has_print);
|
|
if (bits & ARG_BITS_20) scratch_append_bit(name, "20", &has_print);
|
|
if (bits & ARG_BITS_16) scratch_append_bit(name, "16", &has_print);
|
|
if (bits & ARG_BITS_12) scratch_append_bit(name, "12", &has_print);
|
|
if (bits & ARG_BITS_8) scratch_append_bit(name, "8", &has_print);
|
|
if (bits & ARG_BITS_5) scratch_append_bit(name, "5", &has_print);
|
|
}
|
|
|
|
static void arm_arg_to_scratch(AsmArgType type)
|
|
{
|
|
scratch_buffer_clear();
|
|
if (type.is_write && !type.is_readwrite) scratch_buffer_append("w");
|
|
if (type.is_readwrite) scratch_buffer_append("rw");
|
|
if (scratch_buffer.len) scratch_buffer_append(":");
|
|
bool has_print = false;
|
|
if (type.imm_arg_ubits)
|
|
{
|
|
scratch_append_bits("immu", type.imm_arg_ubits);
|
|
has_print = true;
|
|
}
|
|
if (type.imm_arg_ibits)
|
|
{
|
|
if (has_print) scratch_buffer_append("/");
|
|
scratch_append_bits("imm", type.imm_arg_ibits);
|
|
has_print = true;
|
|
}
|
|
if (type.ireg_bits)
|
|
{
|
|
if (has_print) scratch_buffer_append("/");
|
|
scratch_append_bits("r", type.ireg_bits);
|
|
has_print = true;
|
|
}
|
|
if (type.float_bits)
|
|
{
|
|
if (has_print) scratch_buffer_append("/");
|
|
scratch_append_bits("f", type.float_bits);
|
|
has_print = true;
|
|
}
|
|
if (type.vec_bits)
|
|
{
|
|
if (has_print) scratch_buffer_append("/");
|
|
scratch_append_bits("v", type.vec_bits);
|
|
has_print = true;
|
|
}
|
|
if (type.is_address)
|
|
{
|
|
if (has_print) scratch_buffer_append("/");
|
|
scratch_buffer_append("mem");
|
|
}
|
|
}
|
|
|
|
static int comp(const void *a_ptr, const void *b_ptr)
|
|
{
|
|
const AsmInstruction *a = a_ptr;
|
|
const AsmInstruction *b = b_ptr;
|
|
if (a->name == b->name) return 0;
|
|
if (!a->name) return -1;
|
|
if (!b->name) return 1;
|
|
return strcmp(a->name, b->name);
|
|
}
|
|
|
|
static void print_arch_asm(PlatformTarget *target)
|
|
{
|
|
AsmInstruction instructions[ASM_INSTRUCTION_MAX];
|
|
memcpy(instructions, target->instructions, sizeof(target->instructions));
|
|
qsort(instructions, ASM_INSTRUCTION_MAX, sizeof(instructions[0]), comp);
|
|
printf("+------------+--------------------------------+-----------------------------------------------------------------------+\n");
|
|
printf("| instr | clobbers | args |\n");
|
|
printf("+------------+--------------------------------+-----------------------------------------------------------------------+\n");
|
|
for (int i = 0; i < ASM_INSTRUCTION_MAX; i++)
|
|
{
|
|
AsmInstruction *instruction = &instructions[i];
|
|
if (!instruction->name) continue;
|
|
printf("| %-10s | ", instruction->name);
|
|
scratch_buffer_clear();
|
|
Clobbers clobbers = instruction->mask;
|
|
switch (target->arch)
|
|
{
|
|
case ARCH_TYPE_X86_64:
|
|
case ARCH_TYPE_X86:
|
|
{
|
|
for (X86Clobbers cl = 0; cl <= X86_TMM7; cl++)
|
|
{
|
|
if (clobbers.mask[cl / 64] & (1ULL << (cl % 64)))
|
|
{
|
|
scratch_buffer_append(X86ClobberNames[cl]);
|
|
scratch_buffer_append(", ");
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
case ARCH_TYPE_AARCH64:
|
|
case ARCH_TYPE_AARCH64_32:
|
|
case ARCH_TYPE_AARCH64_BE:
|
|
{
|
|
for (Aarch64Clobbers cl = 0; cl <= AARCH64_V31; cl++)
|
|
{
|
|
if (clobbers.mask[cl / 64] & (1ULL << (cl % 64)))
|
|
{
|
|
scratch_buffer_append(Aarch64ClobberNames[cl]);
|
|
scratch_buffer_append(", ");
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
UNREACHABLE
|
|
}
|
|
if (scratch_buffer.len) scratch_buffer.len -= 2;
|
|
printf("%-30s | ", scratch_buffer_to_string());
|
|
int len = 0;
|
|
for (unsigned j = 0; j < instruction->param_count; j++)
|
|
{
|
|
if (j != 0)
|
|
{
|
|
len += 2;
|
|
printf(", ");
|
|
}
|
|
arm_arg_to_scratch(instruction->param[j]);
|
|
printf("%s", scratch_buffer_to_string());
|
|
len += scratch_buffer.len;
|
|
}
|
|
for (int j = 0; j < 70 - len; j++) printf(" ");
|
|
printf("|\n");
|
|
}
|
|
printf("+------------+--------------------------------+-----------------------------------------------------------------------+\n");
|
|
}
|
|
void print_asm(PlatformTarget *target)
|
|
{
|
|
init_asm(target);
|
|
switch (target->arch)
|
|
{
|
|
case ARCH_TYPE_X86_64:
|
|
case ARCH_TYPE_X86:
|
|
case ARCH_TYPE_AARCH64:
|
|
case ARCH_TYPE_AARCH64_32:
|
|
case ARCH_TYPE_AARCH64_BE:
|
|
print_arch_asm(target);
|
|
return;
|
|
default:
|
|
printf("Printing ASM for target is not yet supported.");
|
|
return;
|
|
}
|
|
}
|
|
void init_asm(PlatformTarget *target)
|
|
{
|
|
if (target->asm_initialized) return;
|
|
target->asm_initialized = true;
|
|
switch (target->arch)
|
|
{
|
|
case ARCH_TYPE_X86_64:
|
|
case ARCH_TYPE_X86:
|
|
init_asm_x86(target);
|
|
return;
|
|
case ARCH_TYPE_AARCH64:
|
|
case ARCH_TYPE_AARCH64_BE:
|
|
init_asm_aarch64(target);
|
|
return;
|
|
case ARCH_TYPE_ARM:
|
|
case ARCH_TYPE_ARMB:
|
|
case ARCH_TYPE_THUMB:
|
|
case ARCH_TYPE_THUMBEB:
|
|
init_asm_arm(target);
|
|
return;
|
|
case ARCH_TYPE_WASM64:
|
|
case ARCH_TYPE_WASM32:
|
|
init_asm_wasm(target);
|
|
return;
|
|
case ARCH_TYPE_XTENSA:
|
|
error_exit("Xtensa asm support not yet available.");
|
|
case ARCH_TYPE_UNKNOWN:
|
|
error_exit("Unknown arch does not support asm.");
|
|
UNREACHABLE
|
|
case ARCH_TYPE_PPC:
|
|
case ARCH_TYPE_PPC64:
|
|
case ARCH_TYPE_PPC64LE:
|
|
init_asm_ppc(target);
|
|
return;
|
|
case ARCH_TYPE_RISCV32:
|
|
case ARCH_TYPE_RISCV64:
|
|
init_asm_riscv(target);
|
|
return;
|
|
case ARCH_UNSUPPORTED:
|
|
error_exit("Arch is unsupported and does not support inline asm.");
|
|
}
|
|
UNREACHABLE
|
|
}
|