#define IR_X64 1 #include "ir.h" #include "ir_private.h" #include "ir_x86.h" #ifdef _WIN32 # define IR_SET_ALIGNED(alignment, decl) __declspec(align(alignment)) decl #elif defined(HAVE_ATTRIBUTE_ALIGNED) # define IR_SET_ALIGNED(alignment, decl) decl __attribute__ ((__aligned__ (alignment))) #else # define IR_SET_ALIGNED(alignment, decl) decl #endif #include "dynasm/dasm_proto.h" #include "dynasm/dasm_x86.h" #define IR_IS_SIGNED_32BIT(val) ((((intptr_t)(val)) <= 0x7fffffff) && (((intptr_t)(val)) >= (-2147483647 - 1))) #define IR_IS_UNSIGNED_32BIT(val) ((((uintptr_t)(val)) <= 0xffffffff) && (((uintptr_t)(val)) >= 0)) #define IR_IS_32BIT(type, val) (IR_IS_TYPE_SIGNED(type) ? IR_IS_SIGNED_32BIT((val).i64) : IR_IS_UNSIGNED_32BIT((val).u64)) |.if X64 |.arch x64 |.else |.arch x86 |.endif |.actionlist dasm_actions |.globals ir_lb |.section code, cold_code, rodata, jmp_table |.macro ASM_REG_OP, op, type, reg || switch (ir_type_size[type]) { || case 1: | op Rb(reg) || break; || case 2: | op Rw(reg) || break; || case 4: | op Rd(reg) || break; || case 8: | op Rq(reg) || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_MEM_OP, op, type, mem || switch (ir_type_size[type]) { || case 1: | op byte mem || break; || case 2: | op word mem || break; || case 4: | op dword mem || break; || case 8: | op qword mem || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_MREF_OP, op, type, ref || do { || int32_t offset = ir_ref_spill_slot(ctx, ref); || ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | ASM_MEM_OP op, type, [Ra(fp)+offset] || } while (0); |.endmacro |.macro ASM_REG_REG_OP, op, type, dst, src || switch (ir_type_size[type]) { || case 1: | op Rb(dst), Rb(src) || break; || case 2: | op Rw(dst), Rw(src) || break; || case 4: | op Rd(dst), Rd(src) || break; || case 8: | op Rq(dst), Rq(src) || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_REG_IMM_OP, op, type, dst, src || switch (ir_type_size[type]) { || case 1: | op Rb(dst), src || break; || case 2: | op Rw(dst), src || break; || case 4: | op Rd(dst), src || break; || case 8: | op Rq(dst), src || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_MEM_REG_OP, op, type, dst, src || switch (ir_type_size[type]) { || case 1: | op byte dst, Rb(src) || break; || case 2: | op word dst, Rw(src) || break; || case 4: | op dword dst, Rd(src) || break; || case 8: | op qword dst, Rq(src) || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_MEM_IMM_OP, op, type, dst, src || switch (ir_type_size[type]) { || case 1: | op byte dst, src || break; || case 2: | op word dst, src || break; || case 4: | op dword dst, src || break; || case 8: | op qword dst, src || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_REG_MEM_OP, op, type, dst, src || switch (ir_type_size[type]) { || case 1: | op Rb(dst), byte src || break; || case 2: | op Rw(dst), word src || break; || case 4: | op Rd(dst), dword src || break; || case 8: | op Rq(dst), qword src || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_REG_MREF_OP, _op, type, dst, src || if (IR_IS_CONST_REF(src)) { || ir_insn *_insn = &ctx->ir_base[src]; || IR_ASSERT(IR_IS_SIGNED_32BIT(_insn->val.i64)); | ASM_REG_IMM_OP _op, type, dst, _insn->val.i32 || } else { || int32_t offset = ir_ref_spill_slot(ctx, src); || ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | ASM_REG_MEM_OP _op, type, dst, [Ra(fp)+offset] || } |.endmacro |.macro ASM_MREF_REG_OP, op, type, dst, src || do { || int32_t offset = ir_ref_spill_slot(ctx, dst); || ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | ASM_MEM_REG_OP op, type, [Ra(fp)+offset], src || } while (0); |.endmacro |.macro ASM_MREF_IMM_OP, op, type, dst, src || do { || int32_t offset = ir_ref_spill_slot(ctx, dst); || ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | ASM_MEM_IMM_OP op, type, [Ra(fp)+offset], src || } while (0); |.endmacro |.macro ASM_REG_REG_IMUL, type, dst, src || switch (ir_type_size[type]) { || case 2: | imul Rw(dst), Rw(src) || break; || case 4: | imul Rd(dst), Rd(src) || break; || case 8: | imul Rq(dst), Rq(src) || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_REG_IMM_IMUL, type, dst, src || switch (ir_type_size[type]) { || case 2: | imul Rw(dst), src || break; || case 4: | imul Rd(dst), src || break; || case 8: | imul Rq(dst), src || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_REG_MEM_IMUL, type, dst, src || switch (ir_type_size[type]) { || case 2: | imul Rw(dst), word src || break; || case 4: | imul Rd(dst), dword src || break; || case 8: | imul Rq(dst), qword src || break; || default: || IR_ASSERT(0); || } |.endmacro |.macro ASM_REG_MREF_IMUL, type, dst, src || if (IR_IS_CONST_REF(src)) { || ir_insn *_insn = &ctx->ir_base[src]; || IR_ASSERT(IR_IS_SIGNED_32BIT(_insn->val.i64)); | ASM_REG_IMM_IMUL type, dst, _insn->val.i32 || } else { || int32_t offset = ir_ref_spill_slot(ctx, src); || ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | ASM_REG_MEM_IMUL type, dst, [Ra(fp)+offset] || } |.endmacro |.macro ASM_SSE2_REG_REG_OP, fop, dop, type, dst, src || if (type == IR_FLOAT) { | fop xmm(dst-IR_REG_XMM0), xmm(src-IR_REG_XMM0) || } else if (type == IR_DOUBLE) { | dop xmm(dst-IR_REG_XMM0), xmm(src-IR_REG_XMM0) || } else { || IR_ASSERT(0); || } |.endmacro |.macro ASM_SSE2_MEM_REG_OP, fop, dop, type, dst, src || if (type == IR_FLOAT) { | fop dword dst, xmm(src-IR_REG_XMM0) || } else if (type == IR_DOUBLE) { | dop qword dst, xmm(src-IR_REG_XMM0) || } else { || IR_ASSERT(0); || } |.endmacro |.macro ASM_SSE2_REG_MEM_OP, fop, dop, type, dst, src || if (type == IR_FLOAT) { | fop xmm(dst-IR_REG_XMM0), dword src || } else if (type == IR_DOUBLE) { | dop xmm(dst-IR_REG_XMM0), qword src || } else { || IR_ASSERT(0); || } |.endmacro |.macro ASM_SSE2_REG_MREF_OP, fop, dop, type, dst, src || if (IR_IS_CONST_REF(src)) { || ir_insn *_insn = &ctx->ir_base[src]; || int label = ctx->cfg_blocks_count - src; | ASM_SSE2_REG_MEM_OP fop, dop, type, dst, [=>label] || _insn->emit_const = 1; || } else { || int32_t offset = ir_ref_spill_slot(ctx, src); || ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | ASM_SSE2_REG_MEM_OP fop, dop, type, dst, [Ra(fp)+offset] || } |.endmacro |.macro ASM_AVX_REG_REG_REG_OP, fop, dop, type, dst, op1, op2 || if (type == IR_FLOAT) { | fop xmm(dst-IR_REG_XMM0), xmm(op1-IR_REG_XMM0), xmm(op2-IR_REG_XMM0) || } else if (type == IR_DOUBLE) { | dop xmm(dst-IR_REG_XMM0), xmm(op1-IR_REG_XMM0), xmm(op2-IR_REG_XMM0) || } else { || IR_ASSERT(0); || } |.endmacro |.macro ASM_AVX_REG_REG_MEM_OP, fop, dop, type, dst, op1, op2 || if (type == IR_FLOAT) { | fop xmm(dst-IR_REG_XMM0), xmm(op1-IR_REG_XMM0), dword op2 || } else if (type == IR_DOUBLE) { | dop xmm(dst-IR_REG_XMM0), xmm(op1-IR_REG_XMM0), qword op2 || } else { || IR_ASSERT(0); || } |.endmacro |.macro ASM_AVX_REG_REG_MREF_OP, fop, dop, type, dst, op1, op2 || if (IR_IS_CONST_REF(op2)) { || ir_insn *_insn = &ctx->ir_base[op2]; || int label = ctx->cfg_blocks_count - op2; | ASM_AVX_REG_REG_MEM_OP fop, dop, type, dst, op1, [=>label] || _insn->emit_const = 1; || } else { || int32_t offset = ir_ref_spill_slot(ctx, op2); || ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | ASM_AVX_REG_REG_MEM_OP fop, dop, type, dst, op1, [Ra(fp)+offset] || } |.endmacro |.macro ASM_FP_REG_REG_OP, fop, dop, avx_fop, avx_dop, type, dst, src || if (type == IR_FLOAT) { || if (ctx->flags & IR_AVX) { | avx_fop xmm(dst-IR_REG_XMM0), xmm(src-IR_REG_XMM0) || } else { | fop xmm(dst-IR_REG_XMM0), xmm(src-IR_REG_XMM0) || } || } else if (type == IR_DOUBLE) { || if (ctx->flags & IR_AVX) { | avx_dop xmm(dst-IR_REG_XMM0), xmm(src-IR_REG_XMM0) || } else { | dop xmm(dst-IR_REG_XMM0), xmm(src-IR_REG_XMM0) || } || } else { || IR_ASSERT(0); || } |.endmacro |.macro ASM_FP_MEM_REG_OP, fop, dop, avx_fop, avx_dop, type, dst, src || if (type == IR_FLOAT) { || if (ctx->flags & IR_AVX) { | avx_fop dword dst, xmm(src-IR_REG_XMM0) || } else { | fop dword dst, xmm(src-IR_REG_XMM0) || } || } else if (type == IR_DOUBLE) { || if (ctx->flags & IR_AVX) { | avx_dop qword dst, xmm(src-IR_REG_XMM0) || } else { | dop qword dst, xmm(src-IR_REG_XMM0) || } || } else { || IR_ASSERT(0); || } |.endmacro |.macro ASM_FP_REG_MEM_OP, fop, dop, avx_fop, avx_dop, type, dst, src || if (type == IR_FLOAT) { || if (ctx->flags & IR_AVX) { | avx_fop xmm(dst-IR_REG_XMM0), dword src || } else { | fop xmm(dst-IR_REG_XMM0), dword src || } || } else if (type == IR_DOUBLE) { || if (ctx->flags & IR_AVX) { | avx_dop xmm(dst-IR_REG_XMM0), qword src || } else { | dop xmm(dst-IR_REG_XMM0), qword src || } || } else { || IR_ASSERT(0); || } |.endmacro |.macro ASM_FP_REG_MREF_OP, fop, dop, avx_fop, avx_dop, type, dst, src || if (IR_IS_CONST_REF(src)) { || ir_insn *_insn = &ctx->ir_base[src]; || int label = ctx->cfg_blocks_count - src; | ASM_FP_REG_MEM_OP fop, dop, avx_fop, avx_dop, type, dst, [=>label] || _insn->emit_const = 1; || } else { || int32_t offset = ir_ref_spill_slot(ctx, src); || ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | ASM_FP_REG_MEM_OP fop, dop, avx_fop, avx_dop, type, dst, [Ra(fp)+offset] || } |.endmacro typedef struct _ir_backend_data { int32_t stack_frame_size; int32_t unused_slot_4; int32_t unused_slot_2; int32_t unused_slot_1; int32_t stack_frame_alignment; ir_regset used_preserved_regs; uint32_t dessa_from_block; dasm_State *dasm_state; int rodata_label, jmp_table_label; bool double_neg_const; bool float_neg_const; bool double_abs_const; bool float_abs_const; } ir_backend_data; #define IR_GP_REG_NAME(code, name64, name32, name16, name8, name8h) \ #name64, #define IR_GP_REG_NAME32(code, name64, name32, name16, name8, name8h) \ #name32, #define IR_GP_REG_NAME16(code, name64, name32, name16, name8, name8h) \ #name16, #define IR_GP_REG_NAME8(code, name64, name32, name16, name8, name8h) \ #name8, #define IR_FP_REG_NAME(code, name) \ #name, static const char *_ir_reg_name[IR_REG_NUM] = { IR_GP_REGS(IR_GP_REG_NAME) IR_FP_REGS(IR_FP_REG_NAME) }; static const char *_ir_reg_name32[IR_REG_NUM] = { IR_GP_REGS(IR_GP_REG_NAME32) }; static const char *_ir_reg_name16[IR_REG_NUM] = { IR_GP_REGS(IR_GP_REG_NAME16) }; static const char *_ir_reg_name8[IR_REG_NUM] = { IR_GP_REGS(IR_GP_REG_NAME8) }; /* Calling Convention */ #ifdef _WIN64 static const int8_t _ir_int_reg_params[IR_REG_INT_ARGS] = { IR_REG_INT_ARG1, IR_REG_INT_ARG2, IR_REG_INT_ARG3, IR_REG_INT_ARG4, }; static const int8_t _ir_fp_reg_params[IR_REG_FP_ARGS] = { IR_REG_FP_ARG1, IR_REG_FP_ARG2, IR_REG_FP_ARG3, IR_REG_FP_ARG4, }; #elif defined(__x86_64__) static const int8_t _ir_int_reg_params[IR_REG_INT_ARGS] = { IR_REG_INT_ARG1, IR_REG_INT_ARG2, IR_REG_INT_ARG3, IR_REG_INT_ARG4, IR_REG_INT_ARG5, IR_REG_INT_ARG6, }; static const int8_t _ir_fp_reg_params[IR_REG_FP_ARGS] = { IR_REG_FP_ARG1, IR_REG_FP_ARG2, IR_REG_FP_ARG3, IR_REG_FP_ARG4, IR_REG_FP_ARG5, IR_REG_FP_ARG6, IR_REG_FP_ARG7, IR_REG_FP_ARG8, }; #else static const int8_t *_ir_int_reg_params = NULL; static const int8_t *_ir_fp_reg_params = NULL; #endif const char *ir_reg_name(int8_t reg, ir_type type) { IR_ASSERT(reg >= 0 && reg < IR_REG_NUM); if (IR_IS_TYPE_FP(type) || ir_type_size[type] == 8) { return _ir_reg_name[reg]; } else if (ir_type_size[type] == 4) { return _ir_reg_name32[reg]; } else if (ir_type_size[type] == 2) { return _ir_reg_name16[reg]; } else if (ir_type_size[type] == 1) { return _ir_reg_name8[reg]; } IR_ASSERT(0); return NULL; } typedef enum _ir_rule { IR_SKIP = IR_LAST_OP, IR_SKIP_REG, IR_CMP_INT, // res=reg({%rax,%rbx,%rcx,%rdx}), op1=reg(GP), op2=reg(GP)|mem|imm // res=reg({%rax,%rbx,%rcx,%rdx}), op1=mem, op2=reg(GP)|imm IR_CMP_FP, // res=reg({%rax,%rbx,%rcx,%rdx}), op1=reg(FP), op2=reg(FP)|mem IR_MUL_INT, // res=%rax op1=%rax, op2=reg(GP)|mem, change=%rdx IR_DIV_INT, // res=%rax op1=%rax, op2=reg(GP)|mem, change=%rdx IR_MOD_INT, // res=%rdx op1=%rax, op2=reg(GP)|mem, change=%rax IR_LEA_OB, // res=reg(GP), op1=reg(GP) IR_LEA_SI, // res=reg(GP), op1=reg(GP) IR_LEA_SIB, // res=reg(GP), op1=reg(GP) IR_LEA_IB, // res=reg(GP), op1=reg(GP), op2=reg(GP) IR_LEA_SI_O, IR_LEA_SIB_O, IR_LEA_IB_O, IR_LEA_I_OB, IR_LEA_OB_I, IR_LEA_OB_SI, IR_LEA_SI_OB, IR_LEA_B_SI, IR_LEA_SI_B, IR_INC, IR_DEC, IR_MUL_2, IR_MUL_PWR2, IR_DIV_PWR2, IR_MOD_PWR2, IR_BOOL_NOT_INT, IR_ABS_INT, IR_OP_INT, IR_OP_FP, IR_BINOP_INT, // res=reg(GP), op1.reg(GP, hint=res), op2=any IR_BINOP_SSE2, // res=reg(FP), op1.reg(FP, hint=res), op2=reg(FP)|mem IR_BINOP_AVX, // res=reg(FP), op1.reg(FP), op2=reg(FP)|mem IR_SHIFT, // res=reg(GP), op1.reg(GP, hint=res), op2=%rcx IR_SHIFT_CONST, // res=reg(GP), op1.reg(GP, hint=res), imm IR_COPY_INT, // res=reg(GP), op1=reg(GP)|mem|imm // res=mem, op1=reg(GP)|imm IR_COPY_FP, // res=reg(FP), op1=reg(FP)|mem // res=mem, op2=reg(GP) IR_CMP_AND_BRANCH_INT, // op1=reg(GP), op2=reg(GP)|mem|imm // op1=mem, op2=reg(GP)|imm IR_CMP_AND_BRANCH_FP, // op1=reg(FP), op2=reg(FP)|mem IR_VLOAD_INT, IR_VLOAD_FP, IR_VSTORE_INT, IR_VSTORE_FP, IR_LOAD_INT, IR_LOAD_FP, IR_STORE_INT, IR_STORE_FP, IR_IF_INT, // op1=reg(GP)|mem IR_RETURN_VOID, // IR_RETURN_INT, // op1=reg(GP, hint=%rax)|mem|imm IR_RETURN_FP, // op1=reg(FP, hint=%xmm0)|mem } ir_rule; /* instruction selection */ bool ir_needs_vreg(ir_ctx *ctx, ir_ref ref) { IR_ASSERT(ctx->rules); return ctx->rules[ref] != IR_SKIP; } ir_regset ir_get_scratch_regset(ir_ctx *ctx, ir_ref ref, ir_live_pos *start, ir_live_pos *end) { ir_ref rule; rule = ctx->rules[ref]; if (rule == IR_SHIFT) { *start = IR_LOAD_SUB_REF; *end = IR_DEF_SUB_REF; return IR_REGSET(IR_REG_RCX); } else if (rule == IR_MUL_INT) { *start = IR_USE_SUB_REF; *end = IR_DEF_SUB_REF; return IR_REGSET(IR_REG_RDX); /* %rax - used as input and result */ } else if (rule == IR_DIV_INT) { *start = IR_LOAD_SUB_REF; *end = IR_DEF_SUB_REF; return IR_REGSET(IR_REG_RDX); /* %rax - used as input and result */ } else if (rule == IR_MOD_INT) { *start = IR_LOAD_SUB_REF; *end = IR_DEF_SUB_REF; return IR_REGSET(IR_REG_RAX) | IR_REGSET(IR_REG_RDX); /* %rdx - used as result */ } else if (rule == IR_CALL) { *start = IR_USE_SUB_REF; *end = IR_DEF_SUB_REF; return IR_REGSET_SCRATCH; } return IR_REGSET_EMPTY; } static ir_reg ir_get_param_reg(ir_ctx *ctx, ir_ref ref) { ir_use_list *use_list = &ctx->use_lists[1]; int i; ir_ref use, *p; ir_insn *insn; int int_param = 0; int fp_param = 0; int int_reg_params_count = IR_REG_INT_ARGS; int fp_reg_params_count = IR_REG_FP_ARGS; const int8_t *int_reg_params = _ir_int_reg_params; const int8_t *fp_reg_params = _ir_fp_reg_params; for (i = 0, p = &ctx->use_edges[use_list->refs]; i < use_list->count; i++, p++) { use = *p; insn = &ctx->ir_base[use]; if (insn->op == IR_PARAM) { if (IR_IS_TYPE_INT(insn->type)) { if (use == ref) { if (int_param < int_reg_params_count) { return int_reg_params[int_param]; } else { return IR_REG_NONE; } } int_param++; } else if (IR_IS_TYPE_FP(insn->type)) { if (use == ref) { if (fp_param < fp_reg_params_count) { return fp_reg_params[fp_param]; } else { return IR_REG_NONE; } } fp_param++; } else { IR_ASSERT(0); } } } return IR_REG_NONE; } static ir_reg ir_get_arg_reg(ir_ctx *ctx, ir_insn *insn, int op_num) { int j, n; ir_type type; int int_param = 0; int fp_param = 0; int int_reg_params_count = IR_REG_INT_ARGS; int fp_reg_params_count = IR_REG_FP_ARGS; const int8_t *int_reg_params = _ir_int_reg_params; const int8_t *fp_reg_params = _ir_fp_reg_params; n = ir_input_edges_count(ctx, insn); for (j = 3; j <= n; j++) { type = ctx->ir_base[insn->ops[j]].type; if (IR_IS_TYPE_INT(type)) { if (j == op_num) { if (int_param < int_reg_params_count) { return int_reg_params[int_param]; } else { return IR_REG_NONE; } } int_param++; } else if (IR_IS_TYPE_FP(type)) { if (j == op_num) { if (fp_param < fp_reg_params_count) { return fp_reg_params[fp_param]; } else { return IR_REG_NONE; } } fp_param++; } else { IR_ASSERT(0); } } return IR_REG_NONE; } ir_reg ir_uses_fixed_reg(ir_ctx *ctx, ir_ref ref, int op_num) { ir_ref rule; rule = ctx->rules[ref]; if (rule == IR_SHIFT) { if (op_num == 2) { return IR_REG_RCX; } } else if (rule == IR_MUL_INT || rule == IR_DIV_INT) { if (op_num == 0 || op_num == 1) { return IR_REG_RAX; } } else if (rule == IR_MOD_INT) { if (op_num == 0) { return IR_REG_RDX; } else if (op_num == 1) { return IR_REG_RAX; } } else if (rule == IR_RETURN_INT) { if (op_num == 2) { return IR_REG_RAX; } } else if (rule == IR_RETURN_FP) { if (op_num == 2) { return IR_REG_XMM0; } } else if (rule == IR_SKIP_REG) { if (ctx->ir_base[ref].op == IR_PARAM && op_num == 0) { return ir_get_param_reg(ctx, ref); } } else if (rule == IR_CALL || (rule == IR_TAILCALL && op_num > 0)) { ir_insn *insn = &ctx->ir_base[ref]; if (op_num == 0) { if (IR_IS_TYPE_INT(insn->type)) { return IR_REG_RAX; } else { return IR_REG_XMM0; } } else { return ir_get_arg_reg(ctx, insn, op_num); } } return IR_REG_NONE; } static bool ir_may_need_tmp_reg_for_int_args_swap(ir_ctx *ctx, ir_ref ref) { ir_insn *insn = &ctx->ir_base[ref]; int j, n; ir_type type; int int_param = 0; int int_reg_params_count = IR_REG_INT_ARGS; n = ir_input_edges_count(ctx, insn); for (j = 3; j <= n; j++) { type = ctx->ir_base[insn->ops[j]].type; if (IR_IS_TYPE_INT(type)) { if (int_param < int_reg_params_count) { if (int_param > 0) { return 1; } } else { return 0; } int_param++; } } return 0; } bool ir_result_reuses_op1_reg(ir_ctx *ctx, ir_ref ref) { ir_ref rule; rule = ctx->rules[ref]; switch (rule) { case IR_BINOP_INT: case IR_BINOP_SSE2: case IR_SHIFT: case IR_SHIFT_CONST: case IR_MUL_INT: case IR_DIV_INT: case IR_COPY_INT: case IR_COPY_FP: case IR_INC: case IR_DEC: case IR_MUL_2: case IR_MUL_PWR2: case IR_DIV_PWR2: case IR_MOD_PWR2: case IR_OP_INT: case IR_OP_FP: return 1; } return 0; } uint8_t ir_get_use_flags(ir_ctx *ctx, ir_ref ref, int op_num) { ir_ref rule; ir_insn *insn; rule = ctx->rules[ref]; switch (rule) { case IR_BINOP_INT: case IR_BINOP_SSE2: case IR_BINOP_AVX: case IR_MUL_INT: case IR_DIV_INT: case IR_MOD_INT: case IR_IF_INT: return (op_num == 2) ? IR_USE_SHOULD_BE_IN_REG : IR_USE_MUST_BE_IN_REG; case IR_CMP_FP: insn = &ctx->ir_base[ref]; if (insn->op == IR_LT || insn->op == IR_LE) { return (op_num == 1) ? IR_USE_SHOULD_BE_IN_REG : IR_USE_MUST_BE_IN_REG; } else { return (op_num == 2) ? IR_USE_SHOULD_BE_IN_REG : IR_USE_MUST_BE_IN_REG; } case IR_CALL: case IR_TAILCALL: if (op_num > 2) { insn = &ctx->ir_base[ref]; if (ir_get_arg_reg(ctx, insn, op_num) == IR_REG_NONE) { return IR_USE_MUST_BE_IN_REG; } return IR_USE_SHOULD_BE_IN_REG; } return IR_USE_SHOULD_BE_IN_REG; case IR_SKIP_REG: /* PARAM PHI PI */ insn = &ctx->ir_base[ref]; if (insn->op == IR_PARAM && op_num == 0) { if (ir_get_param_reg(ctx, ref) == IR_REG_NONE) { return IR_USE_MUST_BE_IN_REG; } } return IR_USE_SHOULD_BE_IN_REG; // case IR_VLOAD_INT: // case IR_VLOAD_FP: // case IR_VSTORE_INT: // case IR_VSTORE_FP: // return (op_num == 2) ? 0 : IR_USE_MUST_BE_IN_REG; case IR_SKIP: insn = &ctx->ir_base[ref]; switch (insn->op) { case IR_EQ: case IR_NE: case IR_LT: case IR_GE: case IR_LE: case IR_GT: case IR_ULT: case IR_UGE: case IR_ULE: case IR_UGT: if (IR_IS_TYPE_INT(ctx->ir_base[insn->op1].type)) { if (op_num == 2 || (op_num == 1 && IR_IS_CONST_REF(insn->op2))) { return IR_USE_SHOULD_BE_IN_REG; } } else { if (insn->op == IR_LT || insn->op == IR_LE) { return (op_num == 1) ? IR_USE_SHOULD_BE_IN_REG : IR_USE_MUST_BE_IN_REG; } else { return (op_num == 2) ? IR_USE_SHOULD_BE_IN_REG : IR_USE_MUST_BE_IN_REG; } } break; } break; } return IR_USE_MUST_BE_IN_REG; } int ir_get_temporary_regs(ir_ctx *ctx, ir_ref ref, ir_tmp_reg *tmp_regs) { ir_ref rule; ir_insn *insn; int n; rule = ctx->rules[ref]; switch (rule) { case IR_BINOP_INT: insn = &ctx->ir_base[ref]; if (IR_IS_CONST_REF(insn->op2) && insn->op1 != insn->op2) { insn = &ctx->ir_base[insn->op2]; if (ir_type_size[insn->type] == 8 && !IR_IS_32BIT(insn->type, insn->val)) { tmp_regs[0].num = 2; tmp_regs[0].type = insn->type; tmp_regs[0].start = IR_LOAD_SUB_REF; tmp_regs[0].end = IR_DEF_SUB_REF; return 1; } } break; case IR_DIV_INT: case IR_MOD_INT: insn = &ctx->ir_base[ref]; if (IR_IS_CONST_REF(insn->op2) && insn->op1 != insn->op2) { tmp_regs[0].num = 2; tmp_regs[0].type = insn->type; tmp_regs[0].start = IR_LOAD_SUB_REF; tmp_regs[0].end = IR_DEF_SUB_REF; return 1; } break; case IR_CMP_INT: insn = &ctx->ir_base[ref]; n = 0; cmp_int: if (IR_IS_CONST_REF(insn->op1)) { ir_insn *val_insn = &ctx->ir_base[insn->op1]; if (ir_type_size[val_insn->type] == 8 && !IR_IS_32BIT(val_insn->type, val_insn->val)) { tmp_regs[n].num = 1; tmp_regs[n].type = val_insn->type; tmp_regs[n].start = IR_LOAD_SUB_REF; tmp_regs[n].end = IR_DEF_SUB_REF; n++; } } if (IR_IS_CONST_REF(insn->op2) && insn->op1 != insn->op2) { ir_insn *val_insn = &ctx->ir_base[insn->op2]; if (ir_type_size[val_insn->type] == 8 && !IR_IS_32BIT(val_insn->type, val_insn->val)) { tmp_regs[n].num = 2; tmp_regs[n].type = val_insn->type; tmp_regs[n].start = IR_LOAD_SUB_REF; tmp_regs[n].end = IR_DEF_SUB_REF; n++; } } if (n == 0 && IR_IS_CONST_REF(insn->op1) && IR_IS_CONST_REF(insn->op2)) { insn = &ctx->ir_base[insn->op1]; tmp_regs[n].num = 2; tmp_regs[n].type = insn->type; tmp_regs[n].start = IR_LOAD_SUB_REF; tmp_regs[n].end = IR_DEF_SUB_REF; n++; } return n; case IR_CMP_FP: insn = &ctx->ir_base[ref]; tmp_regs[0].num = 3; tmp_regs[0].type = IR_BOOL; tmp_regs[0].start = IR_DEF_SUB_REF; tmp_regs[0].end = IR_SAVE_SUB_REF; n = 1; cmp_fp: if (insn->op == IR_LT || insn->op == IR_LE) { if (IR_IS_CONST_REF(insn->op2)) { ir_insn *val_insn = &ctx->ir_base[insn->op2]; tmp_regs[n].num = 2; tmp_regs[n].type = val_insn->type; tmp_regs[n].start = IR_LOAD_SUB_REF; tmp_regs[n].end = IR_DEF_SUB_REF; n++; } } else { if (IR_IS_CONST_REF(insn->op1)) { ir_insn *val_insn = &ctx->ir_base[insn->op1]; tmp_regs[n].num = 1; tmp_regs[n].type = val_insn->type; tmp_regs[n].start = IR_LOAD_SUB_REF; tmp_regs[n].end = IR_DEF_SUB_REF; n++; } } return n; case IR_BINOP_AVX: insn = &ctx->ir_base[ref]; if (IR_IS_CONST_REF(insn->op1)) { tmp_regs[0].num = 1; tmp_regs[0].type = insn->type; tmp_regs[0].start = IR_LOAD_SUB_REF; tmp_regs[0].end = IR_DEF_SUB_REF; return 1; } break; case IR_VSTORE_INT: case IR_STORE_INT: insn = &ctx->ir_base[ref]; if (IR_IS_CONST_REF(insn->op3)) { insn = &ctx->ir_base[insn->op3]; if (ir_type_size[insn->type] == 8 && !IR_IS_32BIT(insn->type, insn->val)) { tmp_regs[0].num = 3; tmp_regs[0].type = insn->type; tmp_regs[0].start = IR_LOAD_SUB_REF; tmp_regs[0].end = IR_DEF_SUB_REF; return 1; } } break; case IR_VSTORE_FP: case IR_STORE_FP: insn = &ctx->ir_base[ref]; if (IR_IS_CONST_REF(insn->op3)) { insn = &ctx->ir_base[insn->op3]; tmp_regs[0].num = 3; tmp_regs[0].type = insn->type; tmp_regs[0].start = IR_LOAD_SUB_REF; tmp_regs[0].end = IR_DEF_SUB_REF; return 1; } break; case IR_SWITCH: insn = &ctx->ir_base[ref]; n = 0; if (IR_IS_CONST_REF(insn->op2)) { insn = &ctx->ir_base[insn->op2]; tmp_regs[n].num = 2; tmp_regs[n].type = insn->type; tmp_regs[n].start = IR_LOAD_SUB_REF; tmp_regs[n].end = IR_DEF_SUB_REF; n++; } if (sizeof(void*) == 8) { tmp_regs[n].num = 3; tmp_regs[n].type = IR_ADDR; tmp_regs[n].start = IR_LOAD_SUB_REF; tmp_regs[n].end = IR_DEF_SUB_REF; n++; } return n; case IR_CALL: case IR_TAILCALL: if (ir_may_need_tmp_reg_for_int_args_swap(ctx, ref)) { tmp_regs[0].num = 1; tmp_regs[0].type = IR_ADDR; tmp_regs[0].start = IR_LOAD_SUB_REF; tmp_regs[0].end = IR_USE_SUB_REF; return 1; } break; case IR_SKIP: insn = &ctx->ir_base[ref]; switch (insn->op) { case IR_EQ: case IR_NE: case IR_LT: case IR_GE: case IR_LE: case IR_GT: case IR_ULT: case IR_UGE: case IR_ULE: case IR_UGT: if (IR_IS_TYPE_INT(ctx->ir_base[insn->op1].type)) { /* part of IR_CMP_AND_BRANCH_INT */ n = 0; goto cmp_int; } else { /* part of IR_CMP_AND_BRANCH_FP */ n = 0; goto cmp_fp; } break; } break; } return 0; } static uint32_t ir_match_insn(ir_ctx *ctx, ir_ref ref, ir_block *bb) { ir_insn *op2_insn; ir_insn *insn = &ctx->ir_base[ref]; switch (insn->op) { case IR_EQ: case IR_NE: case IR_LT: case IR_GE: case IR_LE: case IR_GT: case IR_ULT: case IR_UGE: case IR_ULE: case IR_UGT: if (IR_IS_TYPE_INT(ctx->ir_base[insn->op1].type)) { return IR_CMP_INT; } else { return IR_CMP_FP; } break; case IR_ADD: case IR_SUB: if (IR_IS_TYPE_INT(insn->type)) { if ((ctx->flags & IR_OPT_CODEGEN) && IR_IS_CONST_REF(insn->op2)) { op2_insn = &ctx->ir_base[insn->op2]; if (IR_IS_CONST_REF(insn->op1)) { // const } else if (op2_insn->val.i64 == 0) { return IR_COPY_INT; } else if (ir_type_size[insn->type] >= 4 && IR_IS_SIGNED_32BIT(op2_insn->val.i64)) { if (insn->op1 > bb->start && insn->op1 < ref && ctx->use_lists[insn->op1].count == 1) { if (!ctx->rules[insn->op1]) { ctx->rules[insn->op1] = ir_match_insn(ctx, insn->op1, bb); } } if (ctx->rules[insn->op1] == IR_LEA_SI) { ctx->rules[insn->op1] = IR_SKIP; return IR_LEA_SI_O; // lea ret, [op1.op1.reg*op1.op2.scale+op2.offset] } else if (ctx->rules[insn->op1] == IR_LEA_SIB) { ctx->rules[insn->op1] = IR_SKIP; return IR_LEA_SIB_O; // lea ret, [op1.op1.reg+op1.op1.reg*op1.op2.scale+op2.offset] } else if (ctx->rules[insn->op1] == IR_LEA_IB) { ctx->rules[insn->op1] = IR_SKIP; return IR_LEA_IB_O; // lea ret, [op1.op1.reg+op1.op2.reg+op2.offset] } return IR_LEA_OB; // lea ret, [op1.reg+op2.offset] } else if (op2_insn->val.i64 == 1 || op2_insn->val.i64 == -1) { if (insn->op == IR_ADD) { if (op2_insn->val.i64 == 1) { return IR_INC; // inc op1 } else { return IR_DEC; // dec op1 } } else { if (op2_insn->val.i64 == 1) { return IR_DEC; // dec op1 } else { return IR_INC; // inc op1 } } } } else if ((ctx->flags & IR_OPT_CODEGEN) && insn->op == IR_ADD && ir_type_size[insn->type] >= 4) { if (insn->op1 > bb->start && insn->op1 < ref && ctx->use_lists[insn->op1].count == 1) { if (!ctx->rules[insn->op1]) { ctx->rules[insn->op1] = ir_match_insn(ctx, insn->op1, bb); } } if (insn->op2 > bb->start && insn->op2 < ref && ctx->use_lists[insn->op2].count == 1) { if (!ctx->rules[insn->op2]) { ctx->rules[insn->op2] = ir_match_insn(ctx, insn->op2, bb); } } if (ctx->rules[insn->op1] == IR_LEA_OB) { ctx->rules[insn->op1] = IR_SKIP; if (ctx->rules[insn->op2] == IR_LEA_SI) { ctx->rules[insn->op2] = IR_SKIP; return IR_LEA_OB_SI; // lea ret, [op1.op1.reg+op1.op2.offset+op2.op1.reg*op2.op2.scale] } return IR_LEA_OB_I; // lea ret, [op1.op1.reg+op1.op2.offset+op2.reg] } if (ctx->rules[insn->op2] == IR_LEA_OB) { ctx->rules[insn->op2] = IR_SKIP; if (ctx->rules[insn->op1] == IR_LEA_SI) { ctx->rules[insn->op1] = IR_SKIP; return IR_LEA_SI_OB; // lea ret, [op1.op1.reg*op1.op2.scale+op2.op1.reg+op2.op2.offset] } return IR_LEA_I_OB; // lea ret, [op1.reg+op2.op1.reg+op2.op2.offset] } return IR_LEA_IB; // lea ret, [op1.reg+op2.reg] } return IR_BINOP_INT; } else if (ctx->flags & IR_AVX) { return IR_BINOP_AVX; } else { return IR_BINOP_SSE2; } break; case IR_MUL: if (IR_IS_TYPE_INT(insn->type)) { if ((ctx->flags & IR_OPT_CODEGEN) && IR_IS_CONST_REF(insn->op2)) { op2_insn = &ctx->ir_base[insn->op2]; if (IR_IS_CONST_REF(insn->op1)) { // const } else if (op2_insn->val.u64 == 0) { // 0 } else if (op2_insn->val.u64 == 1) { return IR_COPY_INT; } else if (ir_type_size[insn->type] >= 4 && (op2_insn->val.u64 == 2 || op2_insn->val.u64 == 4 || op2_insn->val.u64 == 8)) { return IR_LEA_SI; // lea ret, [op1.reg*op2.scale] } else if (ir_type_size[insn->type] >= 4 && (op2_insn->val.u64 == 3 || op2_insn->val.u64 == 5 || op2_insn->val.u64 == 9)) { return IR_LEA_SIB; // lea ret, [op1.reg+op1.reg*op2.scale] } else if (op2_insn->val.u64 == 2) { return IR_MUL_2; // add op1, op1 } else if (IR_IS_POWER_OF_TWO(op2_insn->val.u64)) { return IR_MUL_PWR2; // shl op1, IR_LOG2(op2_insn->val.u64) } } return (IR_IS_TYPE_SIGNED(insn->type) && ir_type_size[insn->type] != 1) ? IR_BINOP_INT : IR_MUL_INT; } else if (ctx->flags & IR_AVX) { return IR_BINOP_AVX; } else { return IR_BINOP_SSE2; } break; case IR_DIV: if (IR_IS_TYPE_INT(insn->type)) { if ((ctx->flags & IR_OPT_CODEGEN) && IR_IS_CONST_REF(insn->op2)) { op2_insn = &ctx->ir_base[insn->op2]; if (IR_IS_CONST_REF(insn->op1)) { // const } else if (op2_insn->val.u64 == 1) { return IR_COPY_INT; } else if (IR_IS_POWER_OF_TWO(op2_insn->val.u64)) { return IR_DIV_PWR2; // shr op1, IR_LOG2(op2_insn->val.u64) } } return IR_DIV_INT; } else if (ctx->flags & IR_AVX) { return IR_BINOP_AVX; } else { return IR_BINOP_SSE2; } break; case IR_MOD: if ((ctx->flags & IR_OPT_CODEGEN) && IR_IS_CONST_REF(insn->op2)) { op2_insn = &ctx->ir_base[insn->op2]; if (IR_IS_CONST_REF(insn->op1)) { // const } else if (IR_IS_POWER_OF_TWO(op2_insn->val.u64) && IR_IS_UNSIGNED_32BIT(op2_insn->val.u64 - 1)) { return IR_MOD_PWR2; // and op1, op2_insn->val.u64-1 } } return IR_MOD_INT; // case IR_CAST: // case IR_ADD_OV: // case IR_SUB_OV: // case IR_MUL_OV: // case IR_OVERFLOW: case IR_BSWAP: case IR_NOT: if (insn->type == IR_BOOL) { if (IR_IS_TYPE_INT(ctx->ir_base[insn->op1].type)) { return IR_BOOL_NOT_INT; } else { IR_ASSERT(0); // TODO: IR_BOOL_NOT_FP } } else if (IR_IS_TYPE_INT(insn->type)) { return IR_OP_INT; } else { IR_ASSERT(0); } break; case IR_NEG: if (IR_IS_TYPE_INT(insn->type)) { return IR_OP_INT; } else { return IR_OP_FP; } case IR_ABS: if (IR_IS_TYPE_INT(insn->type)) { return IR_ABS_INT; // movl %edi, %eax; negl %eax; cmovs %edi, %eax } else { return IR_OP_FP; } case IR_OR: if ((ctx->flags & IR_OPT_CODEGEN) && IR_IS_CONST_REF(insn->op2)) { op2_insn = &ctx->ir_base[insn->op2]; if (IR_IS_CONST_REF(insn->op1)) { // const } else if (op2_insn->val.i64 == 0) { return IR_COPY_INT; } else if (op2_insn->val.i64 == -1) { // -1 } } return IR_BINOP_INT; case IR_AND: if ((ctx->flags & IR_OPT_CODEGEN) && IR_IS_CONST_REF(insn->op2)) { op2_insn = &ctx->ir_base[insn->op2]; if (IR_IS_CONST_REF(insn->op1)) { // const } else if (op2_insn->val.i64 == 0) { // 0 } else if (op2_insn->val.i64 == -1) { return IR_COPY_INT; } } return IR_BINOP_INT; case IR_XOR: if ((ctx->flags & IR_OPT_CODEGEN) && IR_IS_CONST_REF(insn->op2)) { op2_insn = &ctx->ir_base[insn->op2]; if (IR_IS_CONST_REF(insn->op1)) { // const } } return IR_BINOP_INT; case IR_SHL: if (IR_IS_CONST_REF(insn->op2)) { if (ctx->flags & IR_OPT_CODEGEN) { op2_insn = &ctx->ir_base[insn->op2]; if (IR_IS_CONST_REF(insn->op1)) { // const } else if (op2_insn->val.u64 == 0) { return IR_COPY_INT; } else if (ir_type_size[insn->type] >= 4) { if (op2_insn->val.u64 == 1) { // lea [op1*2] } else if (op2_insn->val.u64 == 2) { // lea [op1*4] } else if (op2_insn->val.u64 == 3) { // lea [op1*8] } } } return IR_SHIFT_CONST; } return IR_SHIFT; case IR_SHR: case IR_SAR: case IR_ROL: case IR_ROR: if (IR_IS_CONST_REF(insn->op2)) { if (ctx->flags & IR_OPT_CODEGEN) { op2_insn = &ctx->ir_base[insn->op2]; if (IR_IS_CONST_REF(insn->op1)) { // const } else if (op2_insn->val.u64 == 0) { return IR_COPY_INT; } } return IR_SHIFT_CONST; } return IR_SHIFT; // case IR_MIN: // case IR_MAX: // case IR_COND: case IR_COPY: if (IR_IS_TYPE_INT(insn->type)) { return IR_COPY_INT; } else { return IR_COPY_FP; } break; case IR_PHI: case IR_PI: case IR_PARAM: return IR_SKIP_REG; case IR_CALL: ctx->flags |= IR_HAS_CALLS; return IR_CALL; case IR_VAR: if (ctx->use_lists[ref].count > 0) { return IR_VAR; } else { return IR_SKIP; } break; case IR_ALLOCA: ctx->flags |= IR_USE_FRAME_POINTER; return IR_ALLOCA; case IR_VLOAD: if (IR_IS_TYPE_INT(insn->type)) { return IR_VLOAD_INT; } else { return IR_VLOAD_FP; } break; case IR_VSTORE: if (IR_IS_TYPE_INT(ctx->ir_base[insn->op3].type)) { return IR_VSTORE_INT; } else { return IR_VSTORE_FP; } break; case IR_LOAD: if (IR_IS_TYPE_INT(ctx->ir_base[insn->op3].type)) { return IR_LOAD_INT; } else { return IR_LOAD_FP; } break; case IR_STORE: if (IR_IS_TYPE_INT(ctx->ir_base[insn->op3].type)) { return IR_STORE_INT; } else { return IR_STORE_FP; } break; case IR_START: case IR_BEGIN: // case IR_END: case IR_IF_TRUE: case IR_IF_FALSE: case IR_CASE_VAL: case IR_CASE_DEFAULT: case IR_MERGE: case IR_LOOP_BEGIN: // case IR_LOOP_END: case IR_LOOP_EXIT: case IR_UNREACHABLE: return IR_SKIP; case IR_RETURN: if (!insn->op2) { return IR_RETURN_VOID; } else if (IR_IS_TYPE_INT(ctx->ir_base[insn->op2].type)) { return IR_RETURN_INT; } else { return IR_RETURN_FP; } case IR_IF: if (insn->op2 > bb->start && insn->op2 < ref && ctx->use_lists[insn->op2].count == 1) { op2_insn = &ctx->ir_base[insn->op2]; if (op2_insn && op2_insn->op >= IR_EQ && op2_insn->op <= IR_UGT) { ctx->rules[insn->op2] = IR_SKIP; if (IR_IS_TYPE_INT(ctx->ir_base[op2_insn->op1].type)) { return IR_CMP_AND_BRANCH_INT; } else { return IR_CMP_AND_BRANCH_FP; } } } if (IR_IS_TYPE_INT(ctx->ir_base[insn->op2].type)) { return IR_IF_INT; } else { IR_ASSERT(0 && "NIY IR_IF_FP"); } default: break; } return insn->op; } int ir_match(ir_ctx *ctx) { int b, n; ir_ref i; ir_block *bb; ir_insn *insn; if (!ctx->prev_insn_len) { ctx->prev_insn_len = ir_mem_malloc(ctx->insns_count * sizeof(uint32_t)); for (b = 1, bb = ctx->cfg_blocks + b; b <= ctx->cfg_blocks_count; b++, bb++) { n = b; /* The first insn of BB keeps BB number in prev_insn_len[] */ for (i = bb->start, insn = ctx->ir_base + i; i <= bb->end;) { ctx->prev_insn_len[i] = n; /* The first insn of BB keeps BB number in prev_insn_len[] */ n = ir_operands_count(ctx, insn); n = 1 + (n >> 2); // support for multi-word instructions like MERGE and PHI i += n; insn += n; } } } ctx->rules = ir_mem_calloc(ctx->insns_count, sizeof(uint32_t)); for (b = ctx->cfg_blocks_count, bb = ctx->cfg_blocks + b; b > 0; b--, bb--) { for (i = bb->end; i >= bb->start; i -= ctx->prev_insn_len[i]) { insn = &ctx->ir_base[i]; if (!ctx->rules[i]) { // if (ctx->rules[i] != IR_SKIP) { ctx->rules[i] = ir_match_insn(ctx, i, bb); } } } return 1; } /* code genertion */ static int ir_skip_empty_blocks(ir_ctx *ctx, int b) { while (ctx->cfg_blocks[b].flags & IR_BB_MAY_SKIP) { b++; } return b; } static ir_reg ir_ref_spill_slot(ir_ctx *ctx, ir_ref ref) { ir_backend_data *data = ctx->data; int32_t offset; IR_ASSERT(ref >= 0); offset = ctx->live_intervals[ctx->vregs[ref]]->stack_spill_pos; IR_ASSERT(offset != -1); if (ctx->flags & IR_USE_FRAME_POINTER) { return offset - data->stack_frame_size - data->stack_frame_alignment; } else { return offset + data->stack_frame_alignment; } } static void ir_emit_load(ir_ctx *ctx, ir_type type, ir_reg reg, ir_ref src) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; int32_t offset; ir_reg fp; if (IR_IS_CONST_REF(src)) { if (IR_IS_TYPE_INT(type)) { ir_insn *insn = &ctx->ir_base[src]; if (insn->op == IR_STR) { int label = ctx->cfg_blocks_count - src; | lea Ra(reg), aword [=>label] insn->emit_const = 1; } else if (ir_type_size[type] == 8 && !IR_IS_SIGNED_32BIT(insn->val.i64)) { if (IR_IS_UNSIGNED_32BIT(insn->val.u64)) { | mov Rd(reg), insn->val.u32 // zero extended load } else { | mov64 Ra(reg), insn->val.i64 } } else if (insn->val.i64 == 0) { | ASM_REG_REG_OP xor, type, reg, reg } else { | ASM_REG_IMM_OP mov, type, reg, insn->val.i32 // sign extended load } } else { ir_insn *insn = &ctx->ir_base[src]; int label; if (insn->type == IR_FLOAT && insn->val.f == 0) { if (ctx->flags & IR_AVX) { | vxorps xmm(reg-IR_REG_XMM0), xmm(reg-IR_REG_XMM0), xmm(reg-IR_REG_XMM0) } else { | xorps xmm(reg-IR_REG_XMM0), xmm(reg-IR_REG_XMM0) } return; } else if (insn->type == IR_DOUBLE && insn->val.d == 0) { if (ctx->flags & IR_AVX) { | vxorpd xmm(reg-IR_REG_XMM0), xmm(reg-IR_REG_XMM0), xmm(reg-IR_REG_XMM0) } else { | xorpd xmm(reg-IR_REG_XMM0), xmm(reg-IR_REG_XMM0) } return; } label = ctx->cfg_blocks_count - src; | ASM_FP_REG_MEM_OP movss, movsd, vmovss, vmovsd, type, reg, [=>label] insn->emit_const = 1; } } else { offset = ctx->live_intervals[ctx->vregs[src]]->stack_spill_pos; IR_ASSERT(offset != -1); if (ctx->flags & IR_USE_FRAME_POINTER) { offset = offset - data->stack_frame_size - data->stack_frame_alignment; fp = IR_REG_RBP; } else { offset = offset + data->stack_frame_alignment; fp = IR_REG_RSP; } if (IR_IS_TYPE_INT(type)) { | ASM_REG_MEM_OP mov, type, reg, [Ra(fp)+offset] } else { | ASM_FP_REG_MEM_OP movss, movsd, vmovss, vmovsd, type, reg, [Ra(fp)+offset] } } } static void ir_emit_store(ir_ctx *ctx, ir_type type, ir_ref dst, ir_reg reg) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; int32_t offset; ir_reg fp; IR_ASSERT(dst >= 0); offset = ctx->live_intervals[ctx->vregs[dst]]->stack_spill_pos; IR_ASSERT(offset != -1); if (ctx->flags & IR_USE_FRAME_POINTER) { offset = offset - data->stack_frame_size - data->stack_frame_alignment; fp = IR_REG_RBP; } else { offset = offset + data->stack_frame_alignment; fp = IR_REG_RSP; } if (IR_IS_TYPE_INT(type)) { | ASM_MEM_REG_OP mov, type, [Ra(fp)+offset], reg } else { | ASM_FP_MEM_REG_OP movss, movsd, vmovss, vmovsd, type, [Ra(fp)+offset], reg } } static bool ir_is_same_mem(ir_ctx *ctx, ir_ref r1, ir_ref r2) { int32_t o1, o2; if (IR_IS_CONST_REF(r1) || IR_IS_CONST_REF(r2)) { return 0; } o1 = ctx->live_intervals[ctx->vregs[r1]]->stack_spill_pos; o2 = ctx->live_intervals[ctx->vregs[r2]]->stack_spill_pos; IR_ASSERT(o1 != -1 && o2 != -1); return o1 == o2 && o1; } static void ir_emit_mov(ir_ctx *ctx, ir_type type, ir_reg dst, ir_reg src) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; | ASM_REG_REG_OP mov, type, dst, src } static void ir_emit_fp_mov(ir_ctx *ctx, ir_type type, ir_reg dst, ir_reg src) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; | ASM_FP_REG_REG_OP movaps, movapd, vmovaps, vmovapd, type, dst, src } static void ir_emit_prologue(ir_ctx *ctx) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; if (ctx->flags & IR_USE_FRAME_POINTER) { | push rbp | mov rbp, rsp } if (data->stack_frame_size) { | sub rsp, data->stack_frame_size } if (data->used_preserved_regs) { int offset; uint32_t i; if (ctx->flags & IR_USE_FRAME_POINTER) { offset = -(int)sizeof(void*); } else { offset = data->stack_frame_size - (int)sizeof(void*); } for (i = 0; i < IR_REG_NUM; i++) { if (IR_REGSET_IN(data->used_preserved_regs, i)) { if (i < IR_REG_FP_FIRST) { ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | mov aword [Ra(fp)+offset], Ra(i) offset -= sizeof(void*); } else { IR_ASSERT(0 && "NIY FP register saing"); } } } } } static void ir_emit_epilogue(ir_ctx *ctx) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; if (data->used_preserved_regs) { int offset; uint32_t i; if (ctx->flags & IR_USE_FRAME_POINTER) { offset = -(int)sizeof(void*); } else { offset = data->stack_frame_size - (int)sizeof(void*); } for (i = 0; i < IR_REG_NUM; i++) { if (IR_REGSET_IN(data->used_preserved_regs, i)) { if (i < IR_REG_FP_FIRST) { ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | mov Ra(i), aword [Ra(fp)+offset] offset -= sizeof(void*); } else { IR_ASSERT(0 && "NIY FP register saing"); } } } } if (ctx->flags & IR_USE_FRAME_POINTER) { | mov rsp, rbp | pop rbp } else if (data->stack_frame_size) { | add rsp, data->stack_frame_size } } static void ir_emit_binop_int(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = insn->type; ir_ref op1 = insn->op1; ir_ref op2 = insn->op2; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; ir_reg op2_reg = ctx->regs[def][2]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (def_reg != op1_reg) { if (op1_reg >= 0) { ir_emit_mov(ctx, type, def_reg, op1_reg); } else { ir_emit_load(ctx, type, def_reg, op1); } } if (op2_reg == IR_REG_NONE && op1 == op2) { op2_reg = def_reg; } if (IR_IS_CONST_REF(op2) && ir_type_size[type] == 8 && !IR_IS_32BIT(type, ctx->ir_base[op2].val)) { /* Load 64-bit constant into a temporary register */ IR_ASSERT(op2_reg != IR_REG_NONE); ir_emit_load(ctx, type, op2_reg, op2); } if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; if (op1 != op2) { ir_emit_load(ctx, type, op2_reg, op2); } } switch (insn->op) { case IR_ADD: | ASM_REG_REG_OP add, type, def_reg, op2_reg break; case IR_SUB: | ASM_REG_REG_OP sub, type, def_reg, op2_reg break; case IR_MUL: | ASM_REG_REG_IMUL type, def_reg, op2_reg break; case IR_OR: | ASM_REG_REG_OP or, type, def_reg, op2_reg break; case IR_AND: | ASM_REG_REG_OP and, type, def_reg, op2_reg break; case IR_XOR: | ASM_REG_REG_OP xor, type, def_reg, op2_reg break; default: IR_ASSERT(0 && "NIY binary op"); break; } } else { switch (insn->op) { case IR_ADD: | ASM_REG_MREF_OP add, type, def_reg, op2 break; case IR_SUB: | ASM_REG_MREF_OP sub, type, def_reg, op2 break; case IR_MUL: | ASM_REG_MREF_IMUL type, def_reg, op2 break; case IR_OR: | ASM_REG_MREF_OP or, type, def_reg, op2 break; case IR_AND: | ASM_REG_MREF_OP and, type, def_reg, op2 break; case IR_XOR: | ASM_REG_MREF_OP xor, type, def_reg, op2 break; default: IR_ASSERT(0 && "NIY binary op"); break; } } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_mul_div_mod_pwr2(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = insn->type; ir_ref op1 = insn->op1; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (def_reg != op1_reg) { if (op1_reg >= 0) { ir_emit_mov(ctx, type, def_reg, op1_reg); } else { ir_emit_load(ctx, type, def_reg, op1); } } if (insn->op == IR_MUL) { uint32_t shift = IR_LOG2(ctx->ir_base[insn->op2].val.u64); if (shift == 1) { | ASM_REG_REG_OP add, insn->type, def_reg, def_reg } else { | ASM_REG_IMM_OP shl, insn->type, def_reg, shift } } else if (insn->op == IR_DIV) { uint32_t shift = IR_LOG2(ctx->ir_base[insn->op2].val.u64); | ASM_REG_IMM_OP shr, insn->type, def_reg, shift } else if (insn->op == IR_MOD) { uint64_t mask = ctx->ir_base[insn->op2].val.u64 - 1; IR_ASSERT(IR_IS_UNSIGNED_32BIT(mask)); | ASM_REG_IMM_OP and, insn->type, def_reg, mask } else { IR_ASSERT(0); } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_shift(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = insn->type; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; ir_reg op2_reg = ctx->regs[def][2]; IR_ASSERT(def_reg != IR_REG_NONE && def_reg != IR_REG_RCX); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, insn->op1); } if (op2_reg != IR_REG_NONE && (op2_reg & IR_REG_SPILL_LOAD)) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op2_reg, insn->op2); } if (op2_reg != IR_REG_RCX) { if (op2_reg >= 0) { ir_emit_mov(ctx, type, IR_REG_RCX, op2_reg); } else { ir_emit_load(ctx, type, IR_REG_RCX, insn->op2); } } if (def_reg != op1_reg) { if (op1_reg >= 0) { ir_emit_mov(ctx, type, def_reg, op1_reg); } else { ir_emit_load(ctx, type, def_reg, insn->op1); } } switch (insn->op) { case IR_SHL: | ASM_REG_IMM_OP shl, insn->type, def_reg, cl break; case IR_SHR: | ASM_REG_IMM_OP shr, insn->type, def_reg, cl break; case IR_SAR: | ASM_REG_IMM_OP sar, insn->type, def_reg, cl break; case IR_ROL: | ASM_REG_IMM_OP rol, insn->type, def_reg, cl break; case IR_ROR: | ASM_REG_IMM_OP ror, insn->type, def_reg, cl break; default: IR_ASSERT(0); } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_shift_const(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; uint32_t shift = ctx->ir_base[insn->op2].val.u64; ir_type type = insn->type; ir_ref op1 = insn->op1; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (def_reg != op1_reg) { if (op1_reg >= 0) { ir_emit_mov(ctx, type, def_reg, op1_reg); } else { ir_emit_load(ctx, type, def_reg, op1); } } switch (insn->op) { case IR_SHL: | ASM_REG_IMM_OP shl, insn->type, def_reg, shift break; case IR_SHR: | ASM_REG_IMM_OP shr, insn->type, def_reg, shift break; case IR_SAR: | ASM_REG_IMM_OP sar, insn->type, def_reg, shift break; case IR_ROL: | ASM_REG_IMM_OP rol, insn->type, def_reg, shift break; case IR_ROR: | ASM_REG_IMM_OP ror, insn->type, def_reg, shift break; default: IR_ASSERT(0); } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_op_int(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = insn->type; ir_ref op1 = insn->op1; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (def_reg != op1_reg) { if (op1_reg >= 0) { ir_emit_mov(ctx, type, def_reg, op1_reg); } else { ir_emit_load(ctx, type, def_reg, op1); } } if (insn->op == IR_ADD) { | ASM_REG_OP inc, insn->type, def_reg } else if (insn->op == IR_SUB) { | ASM_REG_OP dec, insn->type, def_reg } else if (insn->op == IR_NOT) { | ASM_REG_OP not, insn->type, def_reg } else if (insn->op == IR_NEG) { | ASM_REG_OP neg, insn->type, def_reg } else if (insn->op == IR_BSWAP) { switch (ir_type_size[insn->type]) { case 4: | bswap Rd(def_reg) break; case 8: | bswap Rq(def_reg) break; default: IR_ASSERT(0); } } else { IR_ASSERT(0); } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_abs_int(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = insn->type; ir_ref op1 = insn->op1; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; IR_ASSERT(def_reg != IR_REG_NONE && op1_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } IR_ASSERT(def_reg != op1_reg); ir_emit_mov(ctx, insn->type, def_reg, op1_reg); | ASM_REG_OP neg, insn->type, def_reg switch (ir_type_size[insn->type]) { case 1: case 2: | cmovs Rw(def_reg), Rw(op1_reg) break; case 4: | cmovs Rd(def_reg), Rd(op1_reg) break; case 8: | cmovs Rq(def_reg), Rq(op1_reg) break; default: IR_ASSERT(0); } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_bool_not_int(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = ctx->ir_base[insn->op1].type; ir_ref op1 = insn->op1; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (op1_reg >= 0) { | ASM_REG_REG_OP test, type, op1_reg, op1_reg } else { | ASM_MREF_IMM_OP cmp, type, op1, 0 } | setne Rb(def_reg) if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_mul_div_mod(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = insn->type; ir_ref op1 = insn->op1; ir_ref op2 = insn->op2; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; ir_reg op2_reg = ctx->regs[def][2]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (op1_reg != IR_REG_RAX) { if (op1_reg >= 0) { ir_emit_mov(ctx, type, IR_REG_RAX, op1_reg); } else { ir_emit_load(ctx, type, IR_REG_RAX, op1); } } if (op2_reg != IR_REG_NONE && (op2_reg & IR_REG_SPILL_LOAD)) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op2_reg, op2); } if (op2_reg == IR_REG_NONE && op1 == op2) { op2_reg = IR_REG_RAX; } else if (IR_IS_CONST_REF(op2)) { if (insn->op == IR_MUL) { op2_reg = IR_REG_RDX; } else { IR_ASSERT(op2_reg != IR_REG_NONE); } ir_emit_load(ctx, type, op2_reg, op2); } if (insn->op == IR_MUL) { if (IR_IS_TYPE_SIGNED(insn->type)) { if (op2_reg >= 0) { | ASM_REG_OP imul, type, op2_reg } else { | ASM_MREF_OP imul, type, op2 } } else { if (op2_reg >= 0) { | ASM_REG_OP mul, type, op2_reg } else { | ASM_MREF_OP mul, type, op2 } } } else { if (IR_IS_TYPE_SIGNED(type)) { | cdq if (op2_reg >= 0) { | ASM_REG_OP idiv, type, op2_reg } else { | ASM_MREF_OP idiv, type, op2 } } else { | ASM_REG_REG_OP xor, type, IR_REG_RDX, IR_REG_RDX if (op2_reg >= 0) { | ASM_REG_OP div, type, op2_reg } else { | ASM_MREF_OP div, type, op2 } } } if (insn->op == IR_MUL || insn->op == IR_DIV) { if (def_reg != IR_REG_RAX) { if (def_reg != IR_REG_NONE) { ir_emit_mov(ctx, type, def_reg, IR_REG_RAX); } else { ir_emit_store(ctx, type, def, IR_REG_RAX); } } } else if (insn->op == IR_MOD) { if (ir_type_size[type] == 1) { if (def_reg != IR_REG_NONE) { | mov al, ah | mov Rb(def_reg), al } else { int32_t offset = ir_ref_spill_slot(ctx, def); ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | mov byte [Ra(fp)+offset], ah } } else if (def_reg != IR_REG_RDX) { if (def_reg != IR_REG_NONE) { ir_emit_mov(ctx, type, def_reg, IR_REG_RDX); } else { ir_emit_store(ctx, type, def, IR_REG_RDX); } } } else { IR_ASSERT(0); } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_rodata(ir_ctx *ctx) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; |.rodata if (!data->rodata_label) { int label = data->rodata_label = ctx->cfg_blocks_count + ctx->consts_count + 2; |=>label: } } static void ir_emit_op_fp(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = insn->type; ir_ref op1 = insn->op1; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (def_reg != op1_reg) { if (op1_reg >= 0) { ir_emit_fp_mov(ctx, type, def_reg, op1_reg); } else { ir_emit_load(ctx, type, def_reg, op1); } } if (insn->op == IR_NEG) { if (insn->type == IR_DOUBLE) { if (!data->double_neg_const) { data->double_neg_const = 1; ir_rodata(ctx); |.align 8 |->double_neg_const: |.dword 0, 0x80000000, 0, 0 |.code } if (ctx->flags & IR_AVX) { | vxorpd xmm(def_reg-IR_REG_XMM0), xmm(def_reg-IR_REG_XMM0), [->double_neg_const] } else { | xorpd xmm(def_reg-IR_REG_XMM0), [->double_neg_const] } } else { IR_ASSERT(insn->type == IR_FLOAT); if (!data->float_neg_const) { data->float_neg_const = 1; ir_rodata(ctx); |.align 8 |->float_neg_const: |.dword 0x80000000, 0, 0, 0 |.code } if (ctx->flags & IR_AVX) { | vxorps xmm(def_reg-IR_REG_XMM0), xmm(def_reg-IR_REG_XMM0), [->float_neg_const] } else { | xorps xmm(def_reg-IR_REG_XMM0), [->float_neg_const] } } } else if (insn->op == IR_ABS) { if (insn->type == IR_DOUBLE) { if (!data->double_abs_const) { data->double_abs_const = 1; ir_rodata(ctx); |.align 8 |->double_abs_const: |.dword 0xffffffff, 0x7fffffff, 0, 0 |.code } if (ctx->flags & IR_AVX) { | vandpd xmm(def_reg-IR_REG_XMM0), xmm(def_reg-IR_REG_XMM0), [->double_abs_const] } else { | andpd xmm(def_reg-IR_REG_XMM0), [->double_abs_const] } } else { IR_ASSERT(insn->type == IR_FLOAT); if (!data->float_abs_const) { data->float_abs_const = 1; ir_rodata(ctx); |.align 8 |->float_abs_const: |.dword 0x7fffffff, 0, 0, 0 |.code } if (ctx->flags & IR_AVX) { | vandps xmm(def_reg-IR_REG_XMM0), xmm(def_reg-IR_REG_XMM0), [->float_abs_const] } else { | andps xmm(def_reg-IR_REG_XMM0), [->float_abs_const] } } } else { IR_ASSERT(0); } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } } static void ir_emit_binop_sse2(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = insn->type; ir_ref op1 = insn->op1; ir_ref op2 = insn->op2; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; ir_reg op2_reg = ctx->regs[def][2]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (def_reg != op1_reg) { if (op1_reg >= 0) { ir_emit_fp_mov(ctx, type, def_reg, op1_reg); } else { ir_emit_load(ctx, type, def_reg, op1); } } if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; if (op1 != op2) { ir_emit_load(ctx, type, op2_reg, op2); } } switch (insn->op) { case IR_ADD: | ASM_SSE2_REG_REG_OP addss, addsd, type, def_reg, op2_reg break; case IR_SUB: | ASM_SSE2_REG_REG_OP subss, subsd, type, def_reg, op2_reg break; case IR_MUL: | ASM_SSE2_REG_REG_OP mulss, mulsd, type, def_reg, op2_reg break; case IR_DIV: | ASM_SSE2_REG_REG_OP divss, divsd, type, def_reg, op2_reg break; default: IR_ASSERT(0 && "NIY binary op"); break; } } else { switch (insn->op) { case IR_ADD: | ASM_SSE2_REG_MREF_OP addss, addsd, type, def_reg, op2 break; case IR_SUB: | ASM_SSE2_REG_MREF_OP subss, subsd, type, def_reg, op2 break; case IR_MUL: | ASM_SSE2_REG_MREF_OP mulss, mulsd, type, def_reg, op2 break; case IR_DIV: | ASM_SSE2_REG_MREF_OP divss, divsd, type, def_reg, op2 break; default: IR_ASSERT(0 && "NIY binary op"); break; } } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } } static void ir_emit_binop_avx(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = insn->type; ir_ref op1 = insn->op1; ir_ref op2 = insn->op2; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; ir_reg op2_reg = ctx->regs[def][2]; IR_ASSERT(def_reg != IR_REG_NONE && op1_reg != IR_REG_NONE); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (IR_IS_CONST_REF(op1)) { ir_emit_load(ctx, type, op1_reg, op1); } if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; if (op1 != op2) { ir_emit_load(ctx, type, op2_reg, op2); } } switch (insn->op) { case IR_ADD: | ASM_AVX_REG_REG_REG_OP vaddss, vaddsd, type, def_reg, op1_reg, op2_reg break; case IR_SUB: | ASM_AVX_REG_REG_REG_OP vsubss, vsubsd, type, def_reg, op1_reg, op2_reg break; case IR_MUL: | ASM_AVX_REG_REG_REG_OP vmulss, vmulsd, type, def_reg, op1_reg, op2_reg break; case IR_DIV: | ASM_AVX_REG_REG_REG_OP vdivss, vdivsd, type, def_reg, op1_reg, op2_reg break; default: IR_ASSERT(0 && "NIY binary op"); break; } } else { switch (insn->op) { case IR_ADD: | ASM_AVX_REG_REG_MREF_OP vaddss, vaddsd, type, def_reg, op1_reg, op2 break; case IR_SUB: | ASM_AVX_REG_REG_MREF_OP vsubss, vsubsd, type, def_reg, op1_reg, op2 break; case IR_MUL: | ASM_AVX_REG_REG_MREF_OP vmulss, vmulsd, type, def_reg, op1_reg, op2 break; case IR_DIV: | ASM_AVX_REG_REG_MREF_OP vdivss, vdivsd, type, def_reg, op1_reg, op2 break; default: IR_ASSERT(0 && "NIY binary op"); break; } } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } } static void ir_emit_cmp_int(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = ctx->ir_base[insn->op1].type; ir_op op = insn->op; ir_ref op1 = insn->op1; ir_ref op2 = insn->op2; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; ir_reg op2_reg = ctx->regs[def][2]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (op1_reg != IR_REG_NONE && IR_IS_CONST_REF(op1)) { ir_emit_load(ctx, type, op1_reg, op1); } if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; if (op1 != op2) { ir_emit_load(ctx, type, op2_reg, op2); } } if (IR_IS_CONST_REF(op2)) { ir_emit_load(ctx, type, op2_reg, op2); } } if (op1_reg != IR_REG_NONE && op2_reg != IR_REG_NONE) { | ASM_REG_REG_OP cmp, type, op1_reg, op2_reg } else if (op1_reg != IR_REG_NONE) { if (IR_IS_CONST_REF(insn->op2) && ctx->ir_base[insn->op2].val.u64 == 0) { if (op == IR_ULT) { /* always false */ | xor Ra(def_reg), Ra(def_reg) if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } return; } else if (op == IR_UGE) { /* always true */ | ASM_REG_IMM_OP mov, insn->type, def_reg, 1 if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } return; } else if (op == IR_ULE) { op = IR_EQ; } else if (op == IR_UGT) { op = IR_NE; } | ASM_REG_REG_OP test, type, op1_reg, op1_reg } else { | ASM_REG_MREF_OP cmp, type, op1_reg, op2 } } else if (op2_reg != IR_REG_NONE) { | ASM_MREF_REG_OP cmp, type, op1, op2_reg } else { IR_ASSERT(!IR_IS_CONST_REF(op1)); IR_ASSERT(IR_IS_CONST_REF(op2)); IR_ASSERT(IR_IS_32BIT(ctx->ir_base[op2].type, ctx->ir_base[op2].val)); | ASM_MREF_IMM_OP cmp, type, op1, ctx->ir_base[op2].val.i32 } switch (op) { case IR_EQ: | sete Rb(def_reg) break; case IR_NE: | setne Rb(def_reg) break; case IR_LT: | setl Rb(def_reg) break; case IR_GE: | setge Rb(def_reg) break; case IR_LE: | setle Rb(def_reg) break; case IR_GT: | setg Rb(def_reg) break; case IR_ULT: | setb Rb(def_reg) break; case IR_UGE: | setae Rb(def_reg) break; case IR_ULE: | setbe Rb(def_reg) break; case IR_UGT: | seta Rb(def_reg) break; default: IR_ASSERT(0 && "NIY binary op"); break; } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } } static void ir_emit_cmp_fp(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type = ctx->ir_base[insn->op1].type; ir_op op = insn->op; ir_ref op1, op2; ir_reg def_reg, op1_reg, op2_reg, tmp_reg; if (op == IR_LT || op == IR_LE) { /* swap operands to avoid P flag check */ op ^= 3; op1 = insn->op2; op2 = insn->op1; op1_reg = ctx->regs[def][2]; op2_reg = ctx->regs[def][1]; } else { op1 = insn->op1; op2 = insn->op2; op1_reg = ctx->regs[def][1]; op2_reg = ctx->regs[def][2]; } def_reg = IR_REG_NUM(ctx->regs[def][0]); tmp_reg = ctx->regs[def][3]; IR_ASSERT(def_reg != IR_REG_NONE); if (op1_reg == IR_REG_NONE && op2_reg != IR_REG_NONE && (op == IR_EQ || op == IR_NE)) { ir_ref tmp; ir_reg tmp_reg; tmp = op1; op1 = op2; op2 = tmp; tmp_reg = op1_reg; op1_reg = op2_reg; op2_reg = tmp_reg; } IR_ASSERT(op1_reg != IR_REG_NONE); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (IR_IS_CONST_REF(op1)) { ir_emit_load(ctx, type, op1_reg, op1); } if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; if (op1 != op2) { ir_emit_load(ctx, type, op2_reg, op2); } } if (IR_IS_CONST_REF(op2)) { ir_emit_load(ctx, type, op2_reg, op2); } | ASM_FP_REG_REG_OP ucomiss, ucomisd, vucomiss, vucomisd, type, op1_reg, op2_reg } else { | ASM_FP_REG_MREF_OP ucomiss, ucomisd, vucomiss, vucomisd, type, op1_reg, op2 } switch (op) { case IR_EQ: //| sete Rb(reg) | setnp Rb(def_reg) | mov Rd(tmp_reg), 0 | cmovne Rd(def_reg), Rd(tmp_reg) break; case IR_NE: //| setne Rb(reg) | setp Rb(def_reg) | mov Rd(tmp_reg), 1 | cmovne Rd(def_reg), Rd(tmp_reg) break; case IR_LT: //| setb Rb(reg) | setnp Rb(def_reg) | mov Rd(tmp_reg), 0 | cmovae Rd(def_reg), Rd(tmp_reg) break; case IR_GE: | setae Rb(def_reg) break; case IR_LE: //| setbe Rb(reg) | setnp Rb(def_reg) | mov Rd(tmp_reg), 0 | cmova Rd(def_reg), Rd(tmp_reg) break; case IR_GT: | seta Rb(def_reg) break; default: IR_ASSERT(0 && "NIY binary op"); break; } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } } static void ir_emit_jmp_true(ir_ctx *ctx, int b, ir_ref def) { ir_use_list *use_list; ir_insn *use_insn; ir_ref i, *p, use, n; int true_block = 0, false_block = 0; ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; use_list = &ctx->use_lists[def]; n = use_list->count; for (i = 0, p = &ctx->use_edges[use_list->refs]; i < n; i++, p++) { use = *p; use_insn = &ctx->ir_base[use]; if (use_insn->op == IR_IF_TRUE) { true_block = ir_skip_empty_blocks(ctx, ctx->bb_num[use]); } else if (use_insn->op == IR_IF_FALSE) { false_block = ir_skip_empty_blocks(ctx, ctx->bb_num[use]); } else { IR_ASSERT(0); } } IR_ASSERT(true_block && false_block); if (b == ctx->cfg_blocks_count || true_block != ir_skip_empty_blocks(ctx, b + 1)) { | jmp =>true_block } } static void ir_emit_jmp_false(ir_ctx *ctx, int b, ir_ref def) { ir_use_list *use_list; ir_insn *use_insn; ir_ref i, *p, use, n; int true_block = 0, false_block = 0; ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; use_list = &ctx->use_lists[def]; n = use_list->count; for (i = 0, p = &ctx->use_edges[use_list->refs]; i < n; i++, p++) { use = *p; use_insn = &ctx->ir_base[use]; if (use_insn->op == IR_IF_TRUE) { true_block = ir_skip_empty_blocks(ctx, ctx->bb_num[use]); } else if (use_insn->op == IR_IF_FALSE) { false_block = ir_skip_empty_blocks(ctx, ctx->bb_num[use]); } else { IR_ASSERT(0); } } IR_ASSERT(true_block && false_block); if (b == ctx->cfg_blocks_count || true_block != ir_skip_empty_blocks(ctx, b + 1)) { | jmp =>false_block } } static void ir_emit_jcc(ir_ctx *ctx, uint8_t op, int b, ir_ref def, ir_insn *insn, bool int_cmp) { ir_use_list *use_list; ir_insn *use_insn; ir_ref i, *p, use, n; int true_block = 0, false_block = 0, next_block; ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; use_list = &ctx->use_lists[def]; n = use_list->count; for (i = 0, p = &ctx->use_edges[use_list->refs]; i < n; i++, p++) { use = *p; use_insn = &ctx->ir_base[use]; if (use_insn->op == IR_IF_TRUE) { true_block = ir_skip_empty_blocks(ctx, ctx->bb_num[use]); } else if (use_insn->op == IR_IF_FALSE) { false_block = ir_skip_empty_blocks(ctx, ctx->bb_num[use]); } else { IR_ASSERT(0); } } IR_ASSERT(true_block && false_block); next_block = b == ctx->cfg_blocks_count ? 0 : ir_skip_empty_blocks(ctx, b + 1); if (true_block == next_block) { if (int_cmp || (op != IR_GT && op != IR_GE)) { /* swap to avoid unconditional JMP if this doesn't introduce additional JP instruction */ if (op < IR_LT) { op ^= 1; // reverse } else { op ^= 3; // reverse } true_block = false_block; false_block = 0; } } else if (false_block == next_block) { false_block = 0; } if (int_cmp) { switch (op) { case IR_EQ: | je =>true_block break; case IR_NE: | jne =>true_block break; case IR_LT: | jl =>true_block break; case IR_GE: | jge =>true_block break; case IR_LE: | jle =>true_block break; case IR_GT: | jg =>true_block break; case IR_ULT: | jb =>true_block break; case IR_UGE: | jae =>true_block break; case IR_ULE: | jbe =>true_block break; case IR_UGT: | ja =>true_block break; default: IR_ASSERT(0 && "NIY binary op"); break; } } else { switch (op) { case IR_EQ: if (!false_block) { | jp >1 | je =>true_block |1: } else { | jp =>false_block | je =>true_block } break; case IR_NE: | jne =>true_block | jp =>true_block break; case IR_LT: if (!false_block) { | jp >1 | jb =>true_block |1: } else { | jp =>false_block | jb =>true_block } break; case IR_GE: | jae =>true_block break; case IR_LE: if (!false_block) { | jp >1 | jbe =>true_block |1: } else { | jp =>false_block | jbe =>true_block } break; case IR_GT: | ja =>true_block break; // case IR_ULT: fprintf(stderr, "\tjb .LL%d\n", true_block); break; // case IR_UGE: fprintf(stderr, "\tjae .LL%d\n", true_block); break; // case IR_ULE: fprintf(stderr, "\tjbe .LL%d\n", true_block); break; // case IR_UGT: fprintf(stderr, "\tja .LL%d\n", true_block); break; default: IR_ASSERT(0 && "NIY binary op"); break; } } if (false_block) { | jmp =>false_block } } static void ir_emit_cmp_and_branch_int(ir_ctx *ctx, int b, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_insn *cmp_insn = &ctx->ir_base[insn->op2]; ir_op op = cmp_insn->op; ir_type type = ctx->ir_base[cmp_insn->op1].type; ir_ref op1 = cmp_insn->op1; ir_ref op2 = cmp_insn->op2; ir_reg op1_reg = ctx->regs[insn->op2][1]; ir_reg op2_reg = ctx->regs[insn->op2][2]; if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (op1_reg != IR_REG_NONE && IR_IS_CONST_REF(op1)) { ir_emit_load(ctx, type, op1_reg, op1); } if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; if (op1 != op2) { ir_emit_load(ctx, type, op2_reg, op2); } } if (IR_IS_CONST_REF(op2)) { ir_emit_load(ctx, type, op2_reg, op2); } } if (op1_reg != IR_REG_NONE && op2_reg != IR_REG_NONE) { | ASM_REG_REG_OP cmp, type, op1_reg, op2_reg } else if (op1_reg != IR_REG_NONE) { if (IR_IS_CONST_REF(insn->op2) && ctx->ir_base[insn->op2].val.u64 == 0) { if (op == IR_ULT) { /* always false */ ir_emit_jmp_false(ctx, b, def); return; } else if (op == IR_UGE) { /* always true */ ir_emit_jmp_true(ctx, b, def); return; } else if (op == IR_ULE) { op = IR_EQ; } else if (op == IR_UGT) { op = IR_NE; } | ASM_REG_REG_OP test, type, op1_reg, op1_reg } else { | ASM_REG_MREF_OP cmp, type, op1_reg, op2 } } else if (op2_reg != IR_REG_NONE) { | ASM_MREF_REG_OP cmp, type, op1, op2_reg } else { IR_ASSERT(!IR_IS_CONST_REF(op1)); IR_ASSERT(IR_IS_CONST_REF(op2)); IR_ASSERT(IR_IS_32BIT(ctx->ir_base[op2].type, ctx->ir_base[op2].val)); | ASM_MREF_IMM_OP cmp, type, op1, ctx->ir_base[op2].val.i32 } ir_emit_jcc(ctx, op, b, def, insn, 1); } static void ir_emit_cmp_and_branch_fp(ir_ctx *ctx, int b, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_insn *cmp_insn = &ctx->ir_base[insn->op2]; ir_type type = ctx->ir_base[cmp_insn->op1].type; ir_op op = cmp_insn->op; ir_ref op1, op2; ir_reg op1_reg, op2_reg; if (op == IR_LT || op == IR_LE) { /* swap operands to avoid P flag check */ op ^= 3; op1 = cmp_insn->op2; op2 = cmp_insn->op1; op1_reg = ctx->regs[insn->op2][2]; op2_reg = ctx->regs[insn->op2][1]; } else { op1 = cmp_insn->op1; op2 = cmp_insn->op2; op1_reg = ctx->regs[insn->op2][1]; op2_reg = ctx->regs[insn->op2][2]; } if (op1_reg == IR_REG_NONE && op2_reg != IR_REG_NONE && (op == IR_EQ || op == IR_NE)) { ir_ref tmp; ir_reg tmp_reg; tmp = op1; op1 = op2; op2 = tmp; tmp_reg = op1_reg; op1_reg = op2_reg; op2_reg = tmp_reg; } IR_ASSERT(op1_reg != IR_REG_NONE); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, op1); } if (IR_IS_CONST_REF(op1)) { ir_emit_load(ctx, type, op1_reg, op1); } if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; if (op1 != op2) { ir_emit_load(ctx, type, op2_reg, op2); } } if (IR_IS_CONST_REF(op2)) { ir_emit_load(ctx, type, op2_reg, op2); } | ASM_FP_REG_REG_OP ucomiss, ucomisd, vucomiss, vucomisd, type, op1_reg, op2_reg } else { | ASM_FP_REG_MREF_OP ucomiss, ucomisd, vucomiss, vucomisd, type, op1_reg, op2 } ir_emit_jcc(ctx, op, b, def, insn, 0); } static void ir_emit_if_int(ir_ctx *ctx, int b, ir_ref def, ir_insn *insn) { ir_type type = ctx->ir_base[insn->op2].type; ir_reg op2_reg = ctx->regs[def][2]; ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op2_reg, insn->op2); } | ASM_REG_REG_OP test, type, op2_reg, op2_reg } else if (IR_IS_CONST_REF(insn->op2)) { IR_ASSERT(0); } else { | ASM_MREF_IMM_OP cmp, type, insn->op2, 0 } ir_emit_jcc(ctx, IR_NE, b, def, insn, 1); } static void ir_emit_return_void(ir_ctx *ctx) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_emit_epilogue(ctx); | ret } static void ir_emit_return_int(ir_ctx *ctx, ir_reg ref, ir_insn *insn) { ir_reg op2_reg = ctx->regs[ref][2]; if (op2_reg != IR_REG_INT_RET1) { ir_type type = ctx->ir_base[insn->op2].type; if (op2_reg != IR_REG_NONE) { if (!(op2_reg & IR_REG_SPILL_LOAD)) { ir_emit_mov(ctx, type, IR_REG_INT_RET1, op2_reg); } else { ir_emit_load(ctx, type, IR_REG_INT_RET1, insn->op2); } } else { ir_emit_load(ctx, type, IR_REG_INT_RET1, insn->op2); } } ir_emit_return_void(ctx); } static void ir_emit_return_fp(ir_ctx *ctx, ir_reg ref, ir_insn *insn) { ir_reg op2_reg = ctx->regs[ref][2]; if (op2_reg != IR_REG_FP_RET1) { ir_type type = ctx->ir_base[insn->op2].type; if (op2_reg != IR_REG_NONE) { if (!(op2_reg & IR_REG_SPILL_LOAD)) { ir_emit_fp_mov(ctx, type, IR_REG_FP_RET1, op2_reg); } else { ir_emit_load(ctx, type, IR_REG_FP_RET1, insn->op2); } } else { ir_emit_load(ctx, type, IR_REG_FP_RET1, insn->op2); } } ir_emit_return_void(ctx); } static void ir_emit_copy_int(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_ref type = insn->type; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; IR_ASSERT(def_reg != IR_REG_NONE || op1_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, insn->op1); } if (def_reg == op1_reg) { /* same reg */ } else if (def_reg != IR_REG_NONE && op1_reg != IR_REG_NONE) { ir_emit_mov(ctx, type, def_reg, op1_reg); } else if (def_reg != IR_REG_NONE) { ir_emit_load(ctx, type, def_reg, insn->op1); } else if (op1_reg != IR_REG_NONE) { ir_emit_store(ctx, type, def, op1_reg); } else { IR_ASSERT(0); } if (def_reg != IR_REG_NONE && (ctx->regs[def][0] & IR_REG_SPILL_STORE)) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_copy_fp(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_type type = insn->type; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); ir_reg op1_reg = ctx->regs[def][1]; IR_ASSERT(def_reg != IR_REG_NONE || op1_reg != IR_REG_NONE); if (op1_reg != IR_REG_NONE && (op1_reg & IR_REG_SPILL_LOAD)) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op1_reg, insn->op1); } if (def_reg == op1_reg) { /* same reg */ } else if (def_reg != IR_REG_NONE && op1_reg != IR_REG_NONE) { ir_emit_mov(ctx, type, def_reg, op1_reg); } else if (def_reg != IR_REG_NONE) { ir_emit_load(ctx, type, def_reg, insn->op1); } else if (op1_reg != IR_REG_NONE) { ir_emit_store(ctx, type, def, op1_reg); } else { IR_ASSERT(0); } if (def_reg != IR_REG_NONE && (ctx->regs[def][0] & IR_REG_SPILL_STORE)) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_vaddr(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_ref type = insn->type; ir_reg def_reg = ctx->regs[def][0]; int32_t offset; ir_reg fp; IR_ASSERT(def_reg != IR_REG_NONE); offset = ir_ref_spill_slot(ctx, insn->op1); fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | lea Ra(def_reg), aword [Ra(fp)+offset] if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_vload_int(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_ref type = insn->type; ir_reg def_reg = ctx->regs[def][0]; if (def_reg == IR_REG_NONE && ir_is_same_mem(ctx, insn->op2, def)) { return; // fake load } IR_ASSERT(def_reg != IR_REG_NONE); ir_emit_load(ctx, type, def_reg, insn->op2); if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_vload_fp(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_ref type = insn->type; ir_reg def_reg = ctx->regs[def][0]; if (def_reg == IR_REG_NONE && ir_is_same_mem(ctx, insn->op2, def)) { return; // fake load } IR_ASSERT(def_reg != IR_REG_NONE); ir_emit_load(ctx, type, def_reg, insn->op2); if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_vstore_int(ir_ctx *ctx, ir_ref ref, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_insn *val_insn = &ctx->ir_base[insn->op3]; ir_ref type = val_insn->type; ir_reg op3_reg = ctx->regs[ref][3]; int32_t offset; ir_reg fp; if (op3_reg == IR_REG_NONE && !IR_IS_CONST_REF(insn->op3) && ir_is_same_mem(ctx, insn->op3, insn->op2)) { return; // fake store } if (IR_IS_CONST_REF(insn->op3) && IR_IS_32BIT(type, val_insn->val)) { offset = ir_ref_spill_slot(ctx, insn->op2); fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | ASM_MEM_IMM_OP mov, type, [Ra(fp)+offset], val_insn->val.i32 } else { IR_ASSERT(op3_reg != IR_REG_NONE); if (op3_reg != IR_REG_NONE && (op3_reg & IR_REG_SPILL_LOAD)) { op3_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op3_reg, insn->op3); } if (IR_IS_CONST_REF(insn->op3)) { ir_emit_load(ctx, type, op3_reg, insn->op3); } ir_emit_store(ctx, type, insn->op2, op3_reg); } } static void ir_emit_vstore_fp(ir_ctx *ctx, ir_ref ref, ir_insn *insn) { ir_ref type = ctx->ir_base[insn->op3].type; ir_reg op3_reg = ctx->regs[ref][3]; if (op3_reg == IR_REG_NONE && !IR_IS_CONST_REF(insn->op3) && ir_is_same_mem(ctx, insn->op3, insn->op2)) { return; // fake store } IR_ASSERT(op3_reg != IR_REG_NONE); if (op3_reg != IR_REG_NONE && (op3_reg & IR_REG_SPILL_LOAD)) { op3_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op3_reg, insn->op3); } if (IR_IS_CONST_REF(insn->op3)) { ir_emit_load(ctx, type, op3_reg, insn->op3); } ir_emit_store(ctx, type, insn->op2, op3_reg); } static void ir_emit_load_int(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_ref type = insn->type; ir_reg op2_reg = ctx->regs[def][2]; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); IR_ASSERT(def_reg != IR_REG_NONE); if (op2_reg != IR_REG_NONE && (op2_reg & IR_REG_SPILL_LOAD)) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op2_reg, insn->op2); } if (op2_reg == IR_REG_NONE) { op2_reg = def_reg; ir_emit_load(ctx, type, op2_reg, insn->op2); } | ASM_REG_MEM_OP mov, type, def_reg, [Ra(op2_reg)] if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_load_fp(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_ref type = insn->type; ir_reg op2_reg = ctx->regs[def][2]; ir_reg def_reg = IR_REG_NUM(ctx->regs[def][0]); IR_ASSERT(def_reg != IR_REG_NONE && op2_reg != IR_REG_NONE); if (op2_reg != IR_REG_NONE && (op2_reg & IR_REG_SPILL_LOAD)) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op2_reg, insn->op2); } | ASM_FP_REG_MEM_OP movss, movsd, vmovss, vmovsd, type, def_reg, [Ra(op2_reg)] if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, type, def, def_reg); } } static void ir_emit_store_int(ir_ctx *ctx, ir_reg ref, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_insn *val_insn = &ctx->ir_base[insn->op3]; ir_ref type = val_insn->type; ir_reg op2_reg = ctx->regs[ref][2]; ir_reg op3_reg = ctx->regs[ref][3]; IR_ASSERT(op2_reg != IR_REG_NONE); if (op2_reg != IR_REG_NONE && (op2_reg & IR_REG_SPILL_LOAD)) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op2_reg, insn->op2); } if (IR_IS_CONST_REF(insn->op3) && IR_IS_32BIT(type, val_insn->val)) { | ASM_MEM_IMM_OP mov, type, [Ra(op2_reg)], val_insn->val.i32 } else { IR_ASSERT(op3_reg != IR_REG_NONE); if (op3_reg != IR_REG_NONE && (op3_reg & IR_REG_SPILL_LOAD)) { op3_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op3_reg, insn->op3); } if (IR_IS_CONST_REF(insn->op3)) { ir_emit_load(ctx, type, op3_reg, insn->op3); } | ASM_MEM_REG_OP mov, type, [Ra(op2_reg)], op3_reg } } static void ir_emit_store_fp(ir_ctx *ctx, ir_ref ref, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_ref type = ctx->ir_base[insn->op3].type; ir_reg op2_reg = ctx->regs[ref][2]; ir_reg op3_reg = ctx->regs[ref][3]; IR_ASSERT(op2_reg != IR_REG_NONE && op2_reg != IR_REG_NONE); if (op2_reg != IR_REG_NONE && (op2_reg & IR_REG_SPILL_LOAD)) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op2_reg, insn->op2); } if (op3_reg != IR_REG_NONE && (op3_reg & IR_REG_SPILL_LOAD)) { op3_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op3_reg, insn->op3); } if (IR_IS_CONST_REF(insn->op3)) { ir_emit_load(ctx, type, op3_reg, insn->op3); } | ASM_FP_MEM_REG_OP movss, movsd, vmovss, vmovsd, type, [Ra(op2_reg)], op3_reg } static void ir_emit_alloca(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_reg def_reg = ctx->regs[def][0]; if (IR_IS_CONST_REF(insn->op2)) { ir_insn *val = &ctx->ir_base[insn->op2]; int32_t size = val->val.i32; IR_ASSERT(IR_IS_TYPE_INT(val->type)); IR_ASSERT(IR_IS_TYPE_UNSIGNED(val->type) || val->val.i64 > 0); IR_ASSERT(IR_IS_SIGNED_32BIT(val->val.i64)); if (ctx->flags & IR_HAS_CALLS) { /* Stack must be 16 byte aligned */ size = IR_ALIGNED_SIZE(size, 16); } else { size = IR_ALIGNED_SIZE(size, 8); } | ASM_REG_IMM_OP sub, IR_ADDR, IR_REG_RSP, size } else { int32_t alignment = (ctx->flags & IR_HAS_CALLS) ? 16 : 8; ir_reg op2_reg = ctx->regs[def][2]; ir_type type = ctx->ir_base[insn->op2].type; IR_ASSERT(def_reg != IR_REG_NONE); if (op2_reg != IR_REG_NONE && (op2_reg & IR_REG_SPILL_LOAD)) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op2_reg, insn->op2); } if (def_reg != op2_reg) { if (op2_reg != IR_REG_NONE) { ir_emit_mov(ctx, type, def_reg, op2_reg); } else { ir_emit_load(ctx, type, def_reg, insn->op2); } } | ASM_REG_IMM_OP add, IR_ADDR, def_reg, (alignment-1) | ASM_REG_IMM_OP and, IR_ADDR, def_reg, ~(alignment-1) | ASM_REG_REG_OP sub, IR_ADDR, IR_REG_RSP, def_reg } if (def_reg != IR_REG_NONE) { | mov Ra(def_reg), rsp if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } } else { ir_emit_store(ctx, IR_ADDR, def, IR_REG_RSP); } } static void ir_emit_switch(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_type type; ir_use_list *use_list; ir_insn *use_insn, *val; ir_ref i, *p, use, n; int label, default_label = 0; int count = 0; ir_val min, max; int64_t offset; ir_reg op2_reg, tmp_reg; type = ctx->ir_base[insn->op2].type; if (IR_IS_TYPE_SIGNED(type)) { min.u64 = 0x7fffffffffffffff; max.u64 = 0x8000000000000000; } else { min.u64 = 0xffffffffffffffff; max.u64 = 0x0; } use_list = &ctx->use_lists[def]; n = use_list->count; for (i = 0, p = &ctx->use_edges[use_list->refs]; i < n; i++, p++) { use = *p; use_insn = &ctx->ir_base[use]; if (use_insn->op == IR_CASE_VAL) { val = &ctx->ir_base[use_insn->op2]; if (IR_IS_TYPE_SIGNED(type)) { IR_ASSERT(IR_IS_TYPE_SIGNED(val->type)); min.i64 = IR_MIN(min.i64, val->val.i64); max.i64 = IR_MAX(max.i64, val->val.i64); } else { IR_ASSERT(!IR_IS_TYPE_SIGNED(val->type)); min.u64 = (int64_t)IR_MIN(min.u64, val->val.u64); max.u64 = (int64_t)IR_MAX(max.u64, val->val.u64); } count++; } else if (use_insn->op == IR_CASE_DEFAULT) { default_label = ir_skip_empty_blocks(ctx, ctx->bb_num[use]); } else { IR_ASSERT(0); } } op2_reg = ctx->regs[def][2]; tmp_reg = ctx->regs[def][3]; IR_ASSERT(op2_reg != IR_REG_NONE); IR_ASSERT(tmp_reg != IR_REG_NONE || sizeof(void*) != 8); if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, op2_reg, insn->op2); } else if (IR_IS_CONST_REF(insn->op2)) { ir_emit_load(ctx, type, op2_reg, insn->op2); } /* Generate a table jmp or a seqence of calls */ if ((max.i64-min.i64) < count * 8) { int *labels = ir_mem_malloc(sizeof(int) * (max.i64 - min.i64 + 1)); for (i = 0; i <= (max.i64 - min.i64 + 1); i++) { labels[i] = default_label; } for (i = 0, p = &ctx->use_edges[use_list->refs]; i < n; i++, p++) { use = *p; use_insn = &ctx->ir_base[use]; if (use_insn->op == IR_CASE_VAL) { val = &ctx->ir_base[use_insn->op2]; label = ir_skip_empty_blocks(ctx, ctx->bb_num[use]); labels[val->val.i64 - min.i64] = label; } } if (IR_IS_32BIT(type, max)) { | ASM_REG_IMM_OP cmp, type, op2_reg, max.i32 } else { | mov64 Rq(tmp_reg), max.i64 | ASM_REG_REG_OP cmp, type, op2_reg, tmp_reg } if (IR_IS_TYPE_SIGNED(type)) { | jg =>default_label } else { | ja =>default_label } if (IR_IS_32BIT(type, min)) { offset = -min.i64 * 8; if (IR_IS_SIGNED_32BIT(offset)) { | ASM_REG_IMM_OP cmp, type, op2_reg, min.i32 } else { | ASM_REG_REG_OP sub, type, op2_reg, (int32_t)offset // TODO: reg clobbering offset = 0; } } else { | mov64 Rq(tmp_reg), min.i64 | ASM_REG_REG_OP sub, type, op2_reg, tmp_reg // TODO: reg clobbering offset = 0; } if (IR_IS_TYPE_SIGNED(type)) { | jl =>default_label } else { | jb =>default_label } if (sizeof(void*) == 8) { |.if X64 switch (ir_type_size[type]) { case 1: if (IR_IS_TYPE_SIGNED(type)) { | movsx Ra(op2_reg), Rb(op2_reg) } else { | movzx Ra(op2_reg), Rb(op2_reg) } break; case 2: if (IR_IS_TYPE_SIGNED(type)) { | movsx Ra(op2_reg), Rw(op2_reg) } else { | movzx Ra(op2_reg), Rw(op2_reg) } break; case 4: if (IR_IS_TYPE_SIGNED(type)) { | movsxd Ra(op2_reg), Rd(op2_reg) } else { | mov Rd(op2_reg), Rd(op2_reg) } break; case 8: break; default: IR_ASSERT(0); } | lea Ra(tmp_reg), aword [>1] | jmp aword [Ra(tmp_reg)+Ra(op2_reg)*8+(int32_t)offset] |.endif } else { |.if not X64 switch (ir_type_size[type]) { case 1: if (IR_IS_TYPE_SIGNED(type)) { | movsx Ra(op2_reg), Rb(op2_reg) } else { | movzx Ra(op2_reg), Rb(op2_reg) } break; case 2: if (IR_IS_TYPE_SIGNED(type)) { | movsx Ra(op2_reg), Rw(op2_reg) } else { | movzx Ra(op2_reg), Rw(op2_reg) } break; case 4: break; default: IR_ASSERT(0 && "Unsupported type size"); } | jmp aword [(int32_t)offset + Ra(op2_reg) * 4 + >1] |.endif } |.jmp_table if (!data->jmp_table_label) { data->jmp_table_label = ctx->cfg_blocks_count + ctx->consts_count + 3; |=>data->jmp_table_label: } |.align aword |1: for (i = 0; i <= (max.i64 - min.i64 + 1); i++) { | .aword =>labels[i] } |.code ir_mem_free(labels); } else { for (i = 0, p = &ctx->use_edges[use_list->refs]; i < n; i++, p++) { use = *p; use_insn = &ctx->ir_base[use]; if (use_insn->op == IR_CASE_VAL) { val = &ctx->ir_base[use_insn->op2]; label = ir_skip_empty_blocks(ctx, ctx->bb_num[use]); if (IR_IS_32BIT(type, val->val)) { | ASM_REG_IMM_OP cmp, type, op2_reg, val->val.i32 } else { | mov64 Ra(tmp_reg), val->val.i64 | ASM_REG_REG_OP cmp, type, op2_reg, tmp_reg } | je =>label } } if (default_label) { | jmp =>default_label } } } #define IS_32BIT(addr) (((uintptr_t)(addr)) <= 0x7fffffff) #define IS_SIGNED_32BIT(val) ((((intptr_t)(val)) <= 0x7fffffff) && (((intptr_t)(val)) >= (-2147483647 - 1))) #include static void *ir_resolve_sym_name(const char *name) { void *handle = NULL; void *addr; #ifdef RTLD_DEFAULT handle = RTLD_DEFAULT; #endif addr = dlsym(handle, name); IR_ASSERT(addr != NULL); return addr; } typedef struct _ir_copy { ir_type type; ir_reg from; ir_reg to; } ir_copy; static int ir_parallel_copy(ir_ctx *ctx, ir_copy *copies, int count, ir_reg tmp_reg, ir_reg tmp_fp_reg) { int i; int8_t *pred, *loc, *types; ir_reg to, from_reg; ir_type type; ir_regset todo, ready; loc = ir_mem_malloc(IR_REG_NUM * 3 * sizeof(int8_t)); pred = loc + IR_REG_NUM; types = pred + IR_REG_NUM; memset(loc, IR_REG_NONE, IR_REG_NUM * 2 * sizeof(int8_t)); todo = IR_REGSET_EMPTY; ready = IR_REGSET_EMPTY; for (i = 0; i < count; i++) { from_reg = copies[i].from; to = copies[i].to; if (from_reg != to) { loc[from_reg] = from_reg; pred[to] = from_reg; types[to] = copies[i].type; IR_REGSET_INCL(todo, to); } } IR_REGSET_FOREACH(todo, i) { if (loc[i] == IR_REG_NONE) { IR_REGSET_INCL(ready, i); } } IR_REGSET_FOREACH_END(); while (todo != IR_REGSET_EMPTY) { ir_ref /*a, b,*/ c; while (ready != IR_REGSET_EMPTY) { to = IR_REGSET_FIRST(ready); IR_REGSET_EXCL(ready, to); from_reg = pred[to]; type = types[to]; c = loc[from_reg]; if (IR_IS_TYPE_INT(type)) { ir_emit_mov(ctx, type, to, c); } else { ir_emit_fp_mov(ctx, type, to, c); } loc[from_reg] = to; if (from_reg == c && pred[from_reg] != IR_REG_NONE) { IR_REGSET_INCL(ready, from_reg); } } to = IR_REGSET_FIRST(todo); IR_REGSET_EXCL(todo, to); from_reg = pred[to]; if (to != loc[from_reg]) { type = types[to]; if (IR_IS_TYPE_INT(type)) { IR_ASSERT(tmp_reg != IR_REG_NONE); IR_ASSERT(tmp_reg >= IR_REG_GP_FIRST && tmp_reg <= IR_REG_GP_LAST); ir_emit_mov(ctx, type, tmp_reg, to); loc[to] = tmp_reg; } else { IR_ASSERT(tmp_fp_reg != IR_REG_NONE); IR_ASSERT(tmp_fp_reg >= IR_REG_FP_FIRST && tmp_fp_reg <= IR_REG_FP_LAST); ir_emit_fp_mov(ctx, type, tmp_fp_reg, to); loc[to] = tmp_fp_reg; } IR_REGSET_INCL(ready, to); } } ir_mem_free(loc); return 1; } static int32_t ir_call_used_stack(ir_ctx *ctx, ir_insn *insn) { int j, n; ir_type type; int int_param = 0; int fp_param = 0; int int_reg_params_count = IR_REG_INT_ARGS; int fp_reg_params_count = IR_REG_FP_ARGS; int32_t used_stack = 0; n = ir_input_edges_count(ctx, insn); for (j = 3; j <= n; j++) { type = ctx->ir_base[insn->ops[j]].type; if (IR_IS_TYPE_INT(type)) { if (int_param >= int_reg_params_count) { used_stack += sizeof(void*); } int_param++; } else if (IR_IS_TYPE_FP(type)) { if (fp_param >= fp_reg_params_count) { used_stack += sizeof(void*); } fp_param++; } else { IR_ASSERT(0); } } if (used_stack % 16 != 0) { /* Stack must be 16 byte aligned */ used_stack += 8; } return used_stack; } static int32_t ir_emit_arguments(ir_ctx *ctx, ir_ref def, ir_insn *insn, ir_reg tmp_reg) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; int j, n; ir_ref arg; ir_insn *arg_insn; uint8_t type; ir_reg src_reg, dst_reg; int int_param = 0; int fp_param = 0; int count = 0; int int_reg_params_count = IR_REG_INT_ARGS; int fp_reg_params_count = IR_REG_FP_ARGS; const int8_t *int_reg_params = _ir_int_reg_params; const int8_t *fp_reg_params = _ir_fp_reg_params; int32_t used_stack, stack_offset = 0; ir_copy *copies; bool has_mem_const_args = 0; ir_reg tmp_fp_reg = IR_REG_FP_LAST; /* Temporary register for FP loads and swap */ if (0) { // TODO: support for preallocated stack used_stack = 0; } else { used_stack = ir_call_used_stack(ctx, insn); if (used_stack) { | lea rsp, [rsp-used_stack] } } n = ir_input_edges_count(ctx, insn); copies = ir_mem_malloc((n - 2) * sizeof(ir_copy)); for (j = 3; j <= n; j++) { arg = insn->ops[j]; src_reg = ctx->regs[def][j]; arg_insn = &ctx->ir_base[arg]; type = arg_insn->type; if (IR_IS_TYPE_INT(type)) { if (int_param < int_reg_params_count) { dst_reg = int_reg_params[int_param]; } else { dst_reg = IR_REG_NONE; /* pass argument through stack */ } int_param++; } else if (IR_IS_TYPE_FP(type)) { if (fp_param < fp_reg_params_count) { dst_reg = fp_reg_params[fp_param]; } else { dst_reg = IR_REG_NONE; /* pass argument through stack */ } fp_param++; } else { IR_ASSERT(0); } if (dst_reg != IR_REG_NONE && (IR_IS_CONST_REF(arg) || src_reg == IR_REG_NONE)) { /* delay constant and memory arguments for second pass */ has_mem_const_args = 1; continue; } if (dst_reg != IR_REG_NONE) { IR_ASSERT(src_reg != IR_REG_NONE); if (src_reg & IR_REG_SPILL_LOAD) { src_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, src_reg, arg); } if (src_reg != dst_reg) { copies[count].type = type; copies[count].from = src_reg; copies[count].to = dst_reg; count++; } } else { /* Pass argument through stack */ if (IR_IS_TYPE_INT(type)) { if (IR_IS_CONST_REF(arg)) { ir_val *val = &ctx->ir_base[arg].val; if (IR_IS_SIGNED_32BIT(val->i64)) { if (ir_type_size[type] <= 4) { | mov dword [rsp+stack_offset], val->i32 } else { | mov qword [rsp+stack_offset], val->i32 } } else { IR_ASSERT(tmp_reg != IR_REG_NONE); | mov64 Ra(tmp_reg), val->i64 | mov [rsp+stack_offset], Ra(tmp_reg) } } else { IR_ASSERT(src_reg != IR_REG_NONE); if (src_reg & IR_REG_SPILL_LOAD) { src_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, src_reg, arg); } | ASM_MEM_REG_OP, mov, type, [rsp+stack_offset], src_reg } } else { if (IR_IS_CONST_REF(arg)) { ir_emit_load(ctx, type, tmp_fp_reg, arg); src_reg = tmp_fp_reg; } else { IR_ASSERT(src_reg != IR_REG_NONE); if (src_reg & IR_REG_SPILL_LOAD) { src_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, src_reg, arg); } } | ASM_FP_MEM_REG_OP movss, movsd, vmovss, vmovsd, type, [rsp+stack_offset], src_reg } stack_offset += sizeof(void*); } } if (count) { ir_parallel_copy(ctx, copies, count, tmp_reg, tmp_fp_reg); } ir_mem_free(copies); if (has_mem_const_args) { int_param = 0; fp_param = 0; for (j = 3; j <= n; j++) { arg = insn->ops[j]; src_reg = ctx->regs[def][j]; arg_insn = &ctx->ir_base[arg]; type = arg_insn->type; if (IR_IS_TYPE_INT(type)) { if (int_param < int_reg_params_count) { dst_reg = int_reg_params[int_param]; } else { dst_reg = IR_REG_NONE; /* argument already passed through stack */ } int_param++; } else if (IR_IS_TYPE_FP(type)) { if (fp_param < fp_reg_params_count) { dst_reg = fp_reg_params[fp_param]; } else { dst_reg = IR_REG_NONE; /* argument already passed through stack */ } fp_param++; } else { IR_ASSERT(0); } if (dst_reg != IR_REG_NONE && (IR_IS_CONST_REF(arg) || src_reg == IR_REG_NONE)) { if (IR_IS_TYPE_INT(type)) { if (IR_IS_CONST_REF(arg) && ir_type_size[type] == 1) { type = IR_ADDR; } ir_emit_load(ctx, type, dst_reg, arg); } else { ir_emit_load(ctx, type, dst_reg, arg); } } } } return used_stack; } static void ir_emit_call(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_reg def_reg; int32_t used_stack = ir_emit_arguments(ctx, def, insn, ctx->regs[def][1]); if (IR_IS_CONST_REF(insn->op2)) { const char *name = ir_get_str(ctx, ctx->ir_base[insn->op2].val.addr); void *addr = ir_resolve_sym_name(name); // if (IS_SIGNED_32BIT(addr)) { // TODO: 32-bit IP relative or 64-bit absolute address // | call qword &addr // } else { | mov64 rax, ((ptrdiff_t)addr) // 0x48 0xb8 | call rax // } } else { ir_reg op2_reg = ctx->regs[def][2]; if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, IR_ADDR, op2_reg, insn->op2); } | call Ra(op2_reg) } else { int32_t offset = ir_ref_spill_slot(ctx, insn->op2); ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | call aword [Ra(fp)+offset] } } if (used_stack) { | lea rsp, [rsp+used_stack] } if (insn->type != IR_VOID) { if (IR_IS_TYPE_INT(insn->type)) { def_reg = IR_REG_NUM(ctx->regs[def][0]); if (def_reg != IR_REG_NONE) { if (def_reg != IR_REG_INT_RET1) { ir_emit_mov(ctx, insn->type, def_reg, IR_REG_INT_RET1); } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } } else { ir_emit_store(ctx, insn->type, def, IR_REG_INT_RET1); } } else if (IR_IS_TYPE_FP(insn->type)) { def_reg = IR_REG_NUM(ctx->regs[def][0]); if (def_reg != IR_REG_NONE) { if (def_reg != IR_REG_FP_RET1) { ir_emit_fp_mov(ctx, insn->type, def_reg, IR_REG_FP_RET1); } if (ctx->regs[def][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, def, def_reg); } } else { ir_emit_store(ctx, insn->type, def, IR_REG_FP_RET1); } } else { IR_ASSERT(0); } } } static void ir_emit_tailcall(ir_ctx *ctx, ir_ref def, ir_insn *insn) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; int32_t used_stack = ir_emit_arguments(ctx, def, insn, ctx->regs[def][1]); IR_ASSERT(used_stack == 0); ir_emit_epilogue(ctx); if (IR_IS_CONST_REF(insn->op2)) { const char *name = ir_get_str(ctx, ctx->ir_base[insn->op2].val.addr); void *addr = ir_resolve_sym_name(name); // if (IS_SIGNED_32BIT(addr)) { // TODO: 32-bit IP relative or 64-bit absolute address // | call qword &addr // } else { | mov64 rax, ((ptrdiff_t)addr) // 0x48 0xb8 | jmp rax // } } else { ir_reg op2_reg = ctx->regs[def][2]; if (op2_reg != IR_REG_NONE) { if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, IR_ADDR, op2_reg, insn->op2); } | jmp Ra(op2_reg) } else { int32_t offset = ir_ref_spill_slot(ctx, insn->op2); ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; | jmp aword [Ra(fp)+offset] } } } static int ir_emit_dessa_move(ir_ctx *ctx, uint8_t type, ir_ref from, ir_ref to) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; uint32_t from_block = data->dessa_from_block; ir_block *from_bb = &ctx->cfg_blocks[from_block]; ir_ref ref = from_bb->end; int8_t to_reg, from_reg; ir_block *to_bb; int j, k = 0; ir_ref phi = 0; bool spill_store = 0; IR_ASSERT(from_bb->successors_count == 1); to_bb = &ctx->cfg_blocks[ctx->cfg_edges[from_bb->successors]]; for (j = 0; j < to_bb->predecessors_count; j++) { if (ctx->cfg_edges[to_bb->predecessors + j] == from_block) { k = j + 2; break; } } IR_ASSERT(k != 0); if (to) { phi = to; } else { ir_use_list *use_list = &ctx->use_lists[from]; for (j = 0; j < use_list->count; j++) { ir_ref use = ctx->use_edges[use_list->refs + j]; ir_insn *insn = &ctx->ir_base[use]; if (insn->op == IR_PHI && insn->ops[k] == from && insn->op1 == to_bb->start) { phi = use; break; } } } IR_ASSERT(phi != 0); if (IR_IS_TYPE_INT(type)) { to_reg = to ? ctx->regs[phi][0] : ctx->regs[ref][0]; /* temporary register for integer swap (see ir_fix_dessa_tmps) */ from_reg = from ? ctx->regs[phi][k] : ctx->regs[ref][0]; /* temporary register for integer swap (see ir_fix_dessa_tmps) */ if (from_reg != IR_REG_NONE && (from_reg & IR_REG_SPILL_LOAD)) { from_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, from_reg, from); } if (to_reg != IR_REG_NONE && (to_reg & IR_REG_SPILL_STORE)) { to_reg &= ~IR_REG_SPILL_STORE; spill_store = 1; } if (from_reg >= 0 && to_reg >= 0) { if (from_reg != to_reg) { ir_emit_mov(ctx, type, to_reg, from_reg); } } else if (to_reg >= 0) { ir_emit_load(ctx, type, to_reg, from); } else if (from_reg >= 0) { ir_emit_store(ctx, type, to, from_reg); } else if (IR_IS_CONST_REF(from) && (ir_type_size[type] != 8 || IR_IS_SIGNED_32BIT(ctx->ir_base[from].val.i64))) { | ASM_MREF_IMM_OP mov, type, to, ctx->ir_base[from].val.i32 } else if (IR_IS_CONST_REF(from) && ir_type_size[type] == 8 && IR_IS_UNSIGNED_32BIT(ctx->ir_base[from].val.u64)) { | mov Rd(to), ctx->ir_base[from].val.u32 } else if (IR_IS_CONST_REF(from) || !ir_is_same_mem(ctx, from, to)) { from_reg = ctx->regs[ref][1]; /* temporary register for int mem->mem (see ir_fix_dessa_tmps) */ IR_ASSERT(from_reg != IR_REG_NONE); ir_emit_load(ctx, type, from_reg, from); ir_emit_store(ctx, type, to, from_reg); } if (spill_store) { ir_emit_store(ctx, type, to, to_reg); } } else { to_reg = to ? ctx->regs[phi][0] : ctx->regs[ref][2]; /* temporary register for fp swap (see ir_fix_dessa_tmps) */ from_reg = from ? ctx->regs[phi][k] : ctx->regs[ref][2]; /* temporary register for fp swap (see ir_fix_dessa_tmps) */ if (from_reg != IR_REG_NONE && (from_reg & IR_REG_SPILL_LOAD)) { from_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, type, from_reg, from); } if (to_reg != IR_REG_NONE && (to_reg & IR_REG_SPILL_STORE)) { to_reg &= ~IR_REG_SPILL_STORE; spill_store = 1; } if (from_reg >= 0 && to_reg >= 0) { if (from_reg != to_reg) { ir_emit_fp_mov(ctx, type, to_reg, from_reg); } } else if (to_reg >= 0) { ir_emit_load(ctx, type, to_reg, from); } else if (from_reg >= 0) { ir_emit_store(ctx, type, to, from_reg); } else if (IR_IS_CONST_REF(from) || !ir_is_same_mem(ctx, from, to)) { from_reg = ctx->regs[ref][3]; /* temporary register for fp mem->mem (see ir_fix_dessa_tmps) */ IR_ASSERT(from_reg != IR_REG_NONE); ir_emit_load(ctx, type, from_reg, from); ir_emit_store(ctx, type, to, from_reg); } if (spill_store) { ir_emit_store(ctx, type, to, to_reg); } } return 1; } static void ir_emit_param_move(ir_ctx *ctx, uint8_t type, ir_reg from_reg, ir_reg to_reg, ir_ref to) { ir_backend_data *data = ctx->data; dasm_State **Dst = &data->dasm_state; ir_reg fp = (ctx->flags & IR_USE_FRAME_POINTER) ? IR_REG_RBP : IR_REG_RSP; int32_t offset; IR_ASSERT(from_reg != IR_REG_NONE || to_reg != IR_REG_NONE); if (IR_IS_TYPE_INT(type)) { if (from_reg != IR_REG_NONE) { if (to_reg != IR_REG_NONE) { ir_emit_mov(ctx, type, to_reg, from_reg); } else { ir_emit_store(ctx, type, to, from_reg); } } else { offset = (ctx->flags & IR_USE_FRAME_POINTER) ? -8 : -8; // TODO: | ASM_REG_MEM_OP mov, type, to_reg, [Ra(fp)+offset] } } else { if (from_reg != IR_REG_NONE) { if (to_reg != IR_REG_NONE) { ir_emit_fp_mov(ctx, type, to_reg, from_reg); } else { ir_emit_store(ctx, type, to, from_reg); } } else { offset = (ctx->flags & IR_USE_FRAME_POINTER) ? -8 : -8; // TODO: | ASM_FP_REG_MEM_OP movss, movsd, vmovss, vmovsd, type, to_reg, [Ra(fp)+offset] } } } static void ir_emit_load_params(ir_ctx *ctx) { ir_use_list *use_list = &ctx->use_lists[1]; ir_insn *insn; ir_ref i, n, *p, use; int int_param_num = 0; int fp_param_num = 0; int src_reg; /* negative values represent arguments on CPU stack */ int dst_reg; // TODO: Calling convention specific int int_reg_params_count = IR_REG_INT_ARGS; int fp_reg_params_count = IR_REG_FP_ARGS; const int8_t *int_reg_params = _ir_int_reg_params; const int8_t *fp_reg_params = _ir_fp_reg_params; n = use_list->count; for (i = 0, p = &ctx->use_edges[use_list->refs]; i < n; i++, p++) { use = *p; insn = &ctx->ir_base[use]; if (insn->op == IR_PARAM) { if (ctx->vregs[use]) { if (IR_IS_TYPE_INT(insn->type)) { if (int_param_num < int_reg_params_count) { src_reg = int_reg_params[int_param_num]; } else { // TODO: replace IR_REG_NONE by stack slot src_reg = IR_REG_NONE; } int_param_num++; } else { if (fp_param_num < fp_reg_params_count) { src_reg = fp_reg_params[fp_param_num]; } else { // TODO: replace IR_REG_NONE by stack slot src_reg = IR_REG_NONE; } fp_param_num++; } dst_reg = IR_REG_NUM(ctx->regs[use][0]); if (src_reg != dst_reg) { // TODO: parallel move ir_emit_param_move(ctx, insn->type, src_reg, dst_reg, use); } if (dst_reg != IR_REG_NONE && (ctx->regs[use][0] & IR_REG_SPILL_STORE)) { ir_emit_store(ctx, insn->type, use, dst_reg); } } else { if (IR_IS_TYPE_INT(insn->type)) { int_param_num++; } else { fp_param_num++; } } } } } static ir_reg ir_get_free_reg(ir_type type, ir_regset available) { if (IR_IS_TYPE_INT(type)) { available = IR_REGSET_INTERSECTION(available, IR_REGSET_GP); } else { IR_ASSERT(IR_IS_TYPE_FP(type)); available = IR_REGSET_INTERSECTION(available, IR_REGSET_FP); } IR_ASSERT(!IR_REGSET_IS_EMPTY(available)); return IR_REGSET_FIRST(available); } static int ir_fix_dessa_tmps(ir_ctx *ctx, uint8_t type, ir_ref from, ir_ref to) { ir_backend_data *data = ctx->data; ir_ref ref = ctx->cfg_blocks[data->dessa_from_block].end; if (to == 0) { if (IR_IS_TYPE_INT(type)) { if (ctx->regs[ref][0] == IR_REG_NONE) { ctx->regs[ref][0] = ctx->regs[ref][1] != IR_REG_RAX ? IR_REG_RAX : IR_REG_RDX; } } else if (IR_IS_TYPE_FP(type)) { if (ctx->regs[ref][2] == IR_REG_NONE) { ctx->regs[ref][2] = ctx->regs[ref][3] != IR_REG_XMM0 ? IR_REG_XMM0 : IR_REG_XMM1; } } else { IR_ASSERT(0); return 0; } } else if (from != 0) { if (IR_IS_TYPE_INT(type)) { if (ctx->regs[ref][1] == IR_REG_NONE) { ctx->regs[ref][1] = ctx->regs[ref][0] != IR_REG_RAX ? IR_REG_RAX : IR_REG_RDX; } } else if (IR_IS_TYPE_FP(type)) { if (ctx->regs[ref][3] == IR_REG_NONE) { ctx->regs[ref][3] = ctx->regs[ref][2] != IR_REG_XMM0 ? IR_REG_XMM0 : IR_REG_XMM1; } } else { IR_ASSERT(0); return 0; } } return 1; } static void ir_allocate_spill_slot(ir_ctx *ctx, ir_live_interval *ival, ir_backend_data *data) { uint8_t size = ir_type_size[ival->type]; if (size == 8) { ival->stack_spill_pos = data->stack_frame_size; data->stack_frame_size += 8; } else if (size == 4) { if (data->unused_slot_4) { ival->stack_spill_pos = data->unused_slot_4; data->unused_slot_4 = 0; } else { ival->stack_spill_pos = data->stack_frame_size; data->unused_slot_4 = data->stack_frame_size + 4; data->stack_frame_size += 8; } } else if (size == 2) { if (data->unused_slot_2) { ival->stack_spill_pos = data->unused_slot_2; data->unused_slot_2 = 0; } else if (data->unused_slot_4) { ival->stack_spill_pos = data->unused_slot_4; data->unused_slot_2 = data->unused_slot_4 + 2; data->unused_slot_4 = 0; } else { ival->stack_spill_pos = data->stack_frame_size; data->unused_slot_2 = data->stack_frame_size + 2; data->unused_slot_4 = data->stack_frame_size + 4; data->stack_frame_size += 8; } } else if (size == 1) { if (data->unused_slot_1) { ival->stack_spill_pos = data->unused_slot_1; data->unused_slot_1 = 0; } else if (data->unused_slot_2) { ival->stack_spill_pos = data->unused_slot_2; data->unused_slot_1 = data->unused_slot_2 + 1; data->unused_slot_2 = 0; } else if (data->unused_slot_4) { ival->stack_spill_pos = data->unused_slot_4; data->unused_slot_1 = data->unused_slot_4 + 1; data->unused_slot_2 = data->unused_slot_4 + 2; data->unused_slot_4 = 0; } else { ival->stack_spill_pos = data->stack_frame_size; data->unused_slot_1 = data->stack_frame_size + 1; data->unused_slot_2 = data->stack_frame_size + 2; data->unused_slot_4 = data->stack_frame_size + 4; data->stack_frame_size += 8; } } else { IR_ASSERT(0); } } static void ir_allocate_unique_spill_slots(ir_ctx *ctx) { int b; ir_block *bb; ir_insn *insn; ir_ref i, n, j, *p; uint32_t flags, *rule, insn_flags; ir_backend_data *data = ctx->data; ir_regset available = 0; ir_tmp_reg tmp_regs[4]; ctx->regs = ir_mem_malloc(sizeof(ir_regs) * ctx->insns_count); memset(ctx->regs, IR_REG_NONE, sizeof(ir_regs) * ctx->insns_count); ctx->live_intervals = ir_mem_calloc(ctx->vregs_count + 1 + IR_REG_NUM, sizeof(ir_live_interval*)); for (b = 1, bb = ctx->cfg_blocks + b; b <= ctx->cfg_blocks_count; b++, bb++) { bb->flags &= ~IR_BB_MAY_SKIP; flags = IR_BB_MAY_SKIP; if (bb->successors_count != 1 || ctx->cfg_edges[bb->successors] != b + 1 || (bb->flags & IR_BB_DESSA_MOVES)) { flags = 0; } for (i = bb->start, insn = ctx->ir_base + i, rule = ctx->rules + i; i <= bb->end;) { switch (ctx->rules ? *rule : insn->op) { case IR_START: case IR_BEGIN: case IR_END: case IR_IF_TRUE: case IR_IF_FALSE: case IR_CASE_VAL: case IR_CASE_DEFAULT: case IR_MERGE: case IR_LOOP_BEGIN: case IR_LOOP_END: case IR_LOOP_EXIT: break; default: flags = 0; case IR_VAR: case IR_SKIP: case IR_SKIP_REG: /* PARAM PHI PI */ /* skip */ if (ctx->rules && *rule != IR_CMP_AND_BRANCH_INT && *rule != IR_CMP_AND_BRANCH_FP) { available = IR_REGSET_SCRATCH; } if (ctx->vregs[i]) { ir_reg reg = ir_uses_fixed_reg(ctx, i, 0); if (reg != IR_REG_NONE && IR_REGSET_IN(available, reg)) { IR_REGSET_EXCL(available, reg); ctx->regs[i][0] = reg | IR_REG_SPILL_STORE; } else if (ir_get_use_flags(ctx, i, 0) & IR_USE_MUST_BE_IN_REG) { if (insn->op == IR_VLOAD && ctx->live_intervals[ctx->vregs[i]] && ctx->live_intervals[ctx->vregs[i]]->stack_spill_pos != -1) { /* pass */ } else { reg = ir_get_free_reg(insn->type, available); IR_REGSET_EXCL(available, reg); ctx->regs[i][0] = reg | IR_REG_SPILL_STORE; } } if (!ctx->live_intervals[ctx->vregs[i]]) { ir_live_interval *ival = ir_mem_calloc(1, sizeof(ir_live_interval)); ctx->live_intervals[ctx->vregs[i]] = ival; ival->type = insn->type; ival->reg = IR_REG_NONE; ival->vreg = ctx->vregs[i]; ir_allocate_spill_slot(ctx, ival, data); ival->top = ival; if (insn->op == IR_VAR) { ir_use_list *use_list = &ctx->use_lists[i]; ir_reg i, n, *p, use; ir_insn *use_insn; int32_t stack_spill_pos = ival->stack_spill_pos; n = use_list->count; for (i = 0, p = &ctx->use_edges[use_list->refs]; i < n; i++, p++) { use = *p; use_insn = &ctx->ir_base[use]; if (use_insn->op == IR_VLOAD) { if (ctx->vregs[use] && !ctx->live_intervals[ctx->vregs[use]]) { ir_live_interval *ival = ir_mem_calloc(1, sizeof(ir_live_interval)); ctx->live_intervals[ctx->vregs[use]] = ival; ival->type = insn->type; ival->reg = IR_REG_NONE; ival->vreg = ctx->vregs[use]; ival->stack_spill_pos = stack_spill_pos; ival->top = ival; } } else if (use_insn->op == IR_VSTORE) { if (!IR_IS_CONST_REF(use_insn->op3) && ctx->vregs[use_insn->op3] && !ctx->live_intervals[ctx->vregs[use_insn->op3]]) { ir_live_interval *ival = ir_mem_calloc(1, sizeof(ir_live_interval)); ctx->live_intervals[ctx->vregs[insn->op3]] = ival; ival->type = insn->type; ival->reg = IR_REG_NONE; ival->vreg = ctx->vregs[insn->op3]; ival->stack_spill_pos = stack_spill_pos; ival->top = ival; } } } } } else if (insn->op == IR_PARAM) { IR_ASSERT(0 && "unexpected PARAM"); return; } } insn_flags = ir_op_flags[insn->op]; n = ir_input_edges_count(ctx, insn); if (n > 0) { ir_live_pos start, end; ir_regset scratch = ir_get_scratch_regset(ctx, i, &start, &end); if (!IR_REGSET_IS_EMPTY(scratch) && start == IR_LOAD_SUB_REF) { available = IR_REGSET_DIFFERENCE(available, scratch); } } for (j = 1, p = insn->ops + 1; j <= n; j++, p++) { ir_ref input = *p; if (IR_OPND_KIND(insn_flags, j) == IR_OPND_DATA && input > 0 && ctx->vregs[input]) { if (j == 1 && ir_result_reuses_op1_reg(ctx, i)) { ir_reg reg = ctx->regs[i][0] & ~IR_REG_SPILL_STORE; ctx->regs[i][1] = reg | IR_REG_SPILL_LOAD; } else { ir_reg reg = ir_uses_fixed_reg(ctx, i, j); if (reg != IR_REG_NONE && IR_REGSET_IN(available, reg)) { IR_REGSET_EXCL(available, reg); ctx->regs[i][j] = reg | IR_REG_SPILL_LOAD; } else if (ir_get_use_flags(ctx, i, j) & IR_USE_MUST_BE_IN_REG) { reg = ir_get_free_reg(ctx->ir_base[input].type, available); IR_REGSET_EXCL(available, reg); ctx->regs[i][j] = reg | IR_REG_SPILL_LOAD; } else if (j > 1 && input == insn->op1 && ctx->regs[i][1] != IR_REG_NONE) { ctx->regs[i][j] = ctx->regs[i][1]; } } } } n = ir_get_temporary_regs(ctx, i, tmp_regs); if (n) { ir_reg reg; do { n--; reg = ir_get_free_reg(tmp_regs[n].type, available); IR_REGSET_EXCL(available, reg); ctx->regs[i][tmp_regs[n].num] = reg; } while (n); } break; } n = ir_operands_count(ctx, insn); n = 1 + (n >> 2); // support for multi-word instructions like MERGE and PHI i += n; insn += n; rule += n; } bb->flags |= flags; if (bb->flags & IR_BB_DESSA_MOVES) { data->dessa_from_block = b; ir_gen_dessa_moves(ctx, b, ir_fix_dessa_tmps); } } if (ctx->flags & IR_HAS_CALLS) { /* Stack must be 16 byte aligned */ if (ctx->flags & IR_USE_FRAME_POINTER) { while (IR_ALIGNED_SIZE(data->stack_frame_size + sizeof(void*) * 2, 16) != data->stack_frame_size + sizeof(void*) * 2) { data->stack_frame_size += 8; data->stack_frame_alignment += 8; } } else { while (IR_ALIGNED_SIZE(data->stack_frame_size + sizeof(void*), 16) != data->stack_frame_size + sizeof(void*)) { data->stack_frame_size += 8; data->stack_frame_alignment += 8; } } } } static void ir_mark_empty_blocks(ir_ctx *ctx) { int b; ir_block *bb; ir_insn *insn; ir_ref i, n; uint32_t flags, *rule; for (b = 1, bb = ctx->cfg_blocks + b; b <= ctx->cfg_blocks_count; b++, bb++) { bb->flags &= ~IR_BB_MAY_SKIP; if (bb->successors_count == 1 && ctx->cfg_edges[bb->successors] == b + 1 && !(bb->flags & IR_BB_DESSA_MOVES)) { flags = IR_BB_MAY_SKIP; for (i = bb->start, insn = ctx->ir_base + i, rule = ctx->rules + i; i < bb->end;) { if (*rule != IR_SKIP && *rule != IR_SKIP_REG) { flags = 0; break; } n = ir_operands_count(ctx, insn); n = 1 + (n >> 2); // support for multi-word instructions like MERGE and PHI i += n; insn += n; rule += n; } bb->flags |= flags; } } } static void ir_calc_stack_frame_size(ir_ctx *ctx, ir_backend_data *data) { int i; ir_live_interval **p, *ival; uint32_t additional_size = 0; for (i = 1, p = ctx->live_intervals + i; i <= ctx->vregs_count; i++, p++) { ival = *p; if (ival) { if (ival->stack_spill_pos != -1) { if (ival->stack_spill_pos + ir_type_size[ival->type] > data->stack_frame_size) { data->stack_frame_size = ival->stack_spill_pos + ir_type_size[ival->type]; } } if (ival->reg >= 0) { if (!IR_REGSET_IN(data->used_preserved_regs, ival->reg) && IR_REGSET_IN(IR_REGSET_PRESERVED, ival->reg)) { IR_REGSET_INCL(data->used_preserved_regs, ival->reg); additional_size += 8; } } } } data->stack_frame_size = IR_ALIGNED_SIZE(data->stack_frame_size, 8); data->stack_frame_size += additional_size; if (ctx->flags & IR_HAS_CALLS) { /* Stack must be 16 byte aligned */ if (ctx->flags & IR_USE_FRAME_POINTER) { while (IR_ALIGNED_SIZE(data->stack_frame_size + sizeof(void*) * 2, 16) != data->stack_frame_size + sizeof(void*) * 2) { data->stack_frame_size += 8; data->stack_frame_alignment += 8; } } else { while (IR_ALIGNED_SIZE(data->stack_frame_size + sizeof(void*), 16) != data->stack_frame_size + sizeof(void*)) { data->stack_frame_size += 8; data->stack_frame_alignment += 8; } } } } static void* dasm_labels[ir_lb_MAX]; void *ir_emit(ir_ctx *ctx, size_t *size) { int b, n, target; ir_block *bb; ir_ref i; ir_insn *insn; uint32_t *rule; ir_backend_data data; dasm_State **Dst; int ret; void *entry; ctx->data = &data; data.stack_frame_size = 0; data.unused_slot_4 = 0; data.unused_slot_2 = 0; data.unused_slot_1 = 0; data.stack_frame_alignment = 0; data.used_preserved_regs = 0; data.rodata_label = 0; data.jmp_table_label = 0; data.double_neg_const = 0; data.float_neg_const = 0; data.double_abs_const = 0; data.float_abs_const = 0; if (!ctx->live_intervals) { ir_allocate_unique_spill_slots(ctx); } else { ir_mark_empty_blocks(ctx); ir_calc_stack_frame_size(ctx, &data); } Dst = &data.dasm_state; data.dasm_state = NULL; dasm_init(&data.dasm_state, DASM_MAXSECTION); dasm_setupglobal(&data.dasm_state, dasm_labels, ir_lb_MAX); dasm_setup(&data.dasm_state, dasm_actions); dasm_growpc(&data.dasm_state, ctx->cfg_blocks_count + 1 + ctx->consts_count + 1 + 1 + 1); if (ctx->flags & IR_FUNCTION) { ir_emit_prologue(ctx); ir_emit_load_params(ctx); } for (b = 1, bb = ctx->cfg_blocks + b; b <= ctx->cfg_blocks_count; b++, bb++) { // if (bb->flags & IR_BB_MAY_SKIP) { // continue; // } // if (ir_needs_block_label(ctx, b)) { |=>b: // } for (i = bb->start, insn = ctx->ir_base + i, rule = ctx->rules + i; i <= bb->end;) { switch (*rule) { case IR_SKIP: case IR_SKIP_REG: case IR_VAR: break; case IR_LEA_OB: { ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[i][1]; int32_t offset = ctx->ir_base[insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, insn->op1); } if (insn->op == IR_SUB) { offset = -offset; } if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+offset] } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_SI: { ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[i][1]; int32_t scale = ctx->ir_base[insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, insn->op1); } if (scale == 2) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)*2] } else { | lea Ra(def_reg), aword [Ra(op1_reg)*2] } } else if (scale == 4) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)*4] } else { | lea Ra(def_reg), aword [Ra(op1_reg)*4] } } else if (scale == 8) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)*8] } else { | lea Ra(def_reg), aword [Ra(op1_reg)*8] } } else { IR_ASSERT(0); } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_SIB: { ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[i][1]; int32_t scale = ctx->ir_base[insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, insn->op1); } if (scale == 3) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op1_reg)*2] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op1_reg)*2] } } else if (scale == 5) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op1_reg)*4] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op1_reg)*4] } } else if (scale == 9) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op1_reg)*8] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op1_reg)*8] } } else { IR_ASSERT(0); } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_IB: { ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[i][1]; ir_reg op2_reg = ctx->regs[i][2]; IR_ASSERT(def_reg >= 0 && op1_reg >= 0 && op2_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, insn->op1); } if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op2_reg, insn->op2); } if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)] } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_OB_I: { ir_insn *op1_insn = &ctx->ir_base[insn->op1]; ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[insn->op1][1]; ir_reg op2_reg = ctx->regs[i][2]; int32_t offset = ctx->ir_base[op1_insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0 && op2_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, op1_insn->op1); } if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op2_reg, insn->op2); } if (op1_insn->op == IR_SUB) { offset = -offset; } if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)+offset] } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_I_OB: { ir_insn *op2_insn = &ctx->ir_base[insn->op2]; ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[i][1]; ir_reg op2_reg = ctx->regs[insn->op2][1]; int32_t offset = ctx->ir_base[op2_insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0 && op2_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, insn->op1); } if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op2_reg, op2_insn->op1); } if (op2_insn->op == IR_SUB) { offset = -offset; } if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)+offset] } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_SI_O: { ir_insn *op1_insn = &ctx->ir_base[insn->op1]; ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[insn->op1][1]; int32_t scale = ctx->ir_base[op1_insn->op2].val.i32; int32_t offset = ctx->ir_base[insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, op1_insn->op1); } if (insn->op == IR_SUB) { offset = -offset; } if (scale == 2) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)*2+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)*2+offset] } } else if (scale == 4) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)*4+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)*4+offset] } } else if (scale == 8) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)*8+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)*8+offset] } } else { IR_ASSERT(0); } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_SIB_O: { ir_insn *op1_insn = &ctx->ir_base[insn->op1]; ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[insn->op1][1]; int32_t scale = ctx->ir_base[op1_insn->op2].val.i32; int32_t offset = ctx->ir_base[insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, op1_insn->op1); } if (insn->op == IR_SUB) { offset = -offset; } if (scale == 3) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op1_reg)*2+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op1_reg)*2+offset] } } else if (scale == 5) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op1_reg)*4+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op1_reg)*4+offset] } } else if (scale == 6) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op1_reg)*8+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op1_reg)*8+offset] } } else { IR_ASSERT(0); } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_IB_O: { ir_insn *op1_insn = &ctx->ir_base[insn->op1]; ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[insn->op1][1]; ir_reg op2_reg = ctx->regs[insn->op1][2]; int32_t offset = ctx->ir_base[insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0 && op2_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, op1_insn->op1); } if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op2_reg, op1_insn->op2); } if (insn->op == IR_SUB) { offset = -offset; } if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)+offset] } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_OB_SI: { ir_insn *op1_insn = &ctx->ir_base[insn->op1]; ir_insn *op2_insn = &ctx->ir_base[insn->op2]; ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[insn->op1][1]; ir_reg op2_reg = ctx->regs[insn->op2][1]; int32_t offset = ctx->ir_base[op1_insn->op2].val.i32; int32_t scale = ctx->ir_base[op2_insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0 && op2_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, op1_insn->op1); } if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op2_reg, op2_insn->op1); } if (op1_insn->op == IR_SUB) { offset = -offset; } if (scale == 2) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)*2+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)*2+offset] } } else if (scale == 4) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)*4+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)*4+offset] } } else if (scale == 8) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)*8+offset] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)*8+offset] } } else { IR_ASSERT(0); } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_SI_OB: { ir_insn *op1_insn = &ctx->ir_base[insn->op1]; ir_insn *op2_insn = &ctx->ir_base[insn->op2]; ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[insn->op1][1]; ir_reg op2_reg = ctx->regs[insn->op2][1]; int32_t scale = ctx->ir_base[op1_insn->op2].val.i32; int32_t offset = ctx->ir_base[op2_insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0 && op2_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, op1_insn->op1); } if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op2_reg, op2_insn->op1); } if (op1_insn->op == IR_SUB) { offset = -offset; } if (scale == 2) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op2_reg)+Rd(op1_reg)*2+offset] } else { | lea Ra(def_reg), aword [Ra(op2_reg)+Ra(op1_reg)*2+offset] } } else if (scale == 4) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op2_reg)+Rd(op1_reg)*4+offset] } else { | lea Ra(def_reg), aword [Ra(op2_reg)+Ra(op1_reg)*4+offset] } } else if (scale == 8) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op2_reg)+Rd(op1_reg)*8+offset] } else { | lea Ra(def_reg), aword [Ra(op2_reg)+Ra(op1_reg)*8+offset] } } else { IR_ASSERT(0); } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_B_SI: { ir_insn *op2_insn = &ctx->ir_base[insn->op2]; ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[i][1]; ir_reg op2_reg = ctx->regs[insn->op2][1]; int32_t scale = ctx->ir_base[op2_insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0 && op2_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, insn->op1); } if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op2_reg, op2_insn->op1); } if (scale == 2) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)*2] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)*2] } } else if (scale == 4) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)*4] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)*4] } } else if (scale == 8) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op1_reg)+Rd(op2_reg)*8] } else { | lea Ra(def_reg), aword [Ra(op1_reg)+Ra(op2_reg)*8] } } else { IR_ASSERT(0); } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_LEA_SI_B: { ir_insn *op1_insn = &ctx->ir_base[insn->op1]; ir_reg def_reg = IR_REG_NUM(ctx->regs[i][0]); ir_reg op1_reg = ctx->regs[insn->op1][1]; ir_reg op2_reg = ctx->regs[i][2]; int32_t scale = ctx->ir_base[op1_insn->op2].val.i32; IR_ASSERT(def_reg >= 0 && op1_reg >= 0 && op2_reg >= 0); if (op1_reg & IR_REG_SPILL_LOAD) { op1_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op1_reg, op1_insn->op1); } if (op2_reg & IR_REG_SPILL_LOAD) { op2_reg &= ~IR_REG_SPILL_LOAD; ir_emit_load(ctx, insn->type, op2_reg, insn->op2); } if (scale == 2) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op2_reg)+Rd(op1_reg)*2] } else { | lea Ra(def_reg), aword [Ra(op2_reg)+Ra(op1_reg)*2] } } else if (scale == 4) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), dword [Rd(op2_reg)+Rd(op1_reg)*4] } else { | lea Ra(def_reg), aword [Ra(op2_reg)+Ra(op1_reg)*4] } } else if (scale == 8) { if (ir_type_size[insn->type] == 4) { | lea Rd(def_reg), aword [Rd(op2_reg)+Rd(op1_reg)*8] } else { | lea Ra(def_reg), aword [Ra(op2_reg)+Ra(op1_reg)*8] } } else { IR_ASSERT(0); } if (ctx->regs[i][0] & IR_REG_SPILL_STORE) { ir_emit_store(ctx, insn->type, i, def_reg); } } break; case IR_MUL_2: case IR_MUL_PWR2: case IR_DIV_PWR2: case IR_MOD_PWR2: ir_emit_mul_div_mod_pwr2(ctx, i, insn); break; case IR_SHIFT: ir_emit_shift(ctx, i, insn); break; case IR_SHIFT_CONST: ir_emit_shift_const(ctx, i, insn); break; case IR_INC: case IR_DEC: case IR_OP_INT: ir_emit_op_int(ctx, i, insn); break; case IR_ABS_INT: ir_emit_abs_int(ctx, i, insn); break; case IR_BOOL_NOT_INT: ir_emit_bool_not_int(ctx, i, insn); break; case IR_OP_FP: ir_emit_op_fp(ctx, i, insn); break; case IR_BINOP_INT: ir_emit_binop_int(ctx, i, insn); break; case IR_BINOP_SSE2: ir_emit_binop_sse2(ctx, i, insn); break; case IR_BINOP_AVX: ir_emit_binop_avx(ctx, i, insn); break; case IR_MUL_INT: case IR_DIV_INT: case IR_MOD_INT: ir_emit_mul_div_mod(ctx, i, insn); break; case IR_CMP_INT: ir_emit_cmp_int(ctx, i, insn); break; case IR_CMP_FP: ir_emit_cmp_fp(ctx, i, insn); break; case IR_COPY_INT: ir_emit_copy_int(ctx, i, insn); break; case IR_COPY_FP: ir_emit_copy_fp(ctx, i, insn); break; case IR_CMP_AND_BRANCH_INT: ir_emit_cmp_and_branch_int(ctx, b, i, insn); break; case IR_CMP_AND_BRANCH_FP: ir_emit_cmp_and_branch_fp(ctx, b, i, insn); break; case IR_IF_INT: ir_emit_if_int(ctx, b, i, insn); break; case IR_SWITCH: ir_emit_switch(ctx, i, insn); break; case IR_END: case IR_LOOP_END: if (bb->flags & IR_BB_DESSA_MOVES) { data.dessa_from_block = b; ir_gen_dessa_moves(ctx, b, ir_emit_dessa_move); } IR_ASSERT(bb->successors_count == 1); target = ir_skip_empty_blocks(ctx, ctx->cfg_edges[bb->successors]); if (b == ctx->cfg_blocks_count || target != ir_skip_empty_blocks(ctx, b + 1)) { | jmp =>target } break; case IR_RETURN_VOID: ir_emit_return_void(ctx); break; case IR_RETURN_INT: ir_emit_return_int(ctx, i, insn); break; case IR_RETURN_FP: ir_emit_return_fp(ctx, i, insn); break; case IR_CALL: ir_emit_call(ctx, i, insn); break; case IR_TAILCALL: ir_emit_tailcall(ctx, i, insn); break; case IR_VADDR: ir_emit_vaddr(ctx, i, insn); break; case IR_VLOAD_INT: ir_emit_vload_int(ctx, i, insn); break; case IR_VLOAD_FP: ir_emit_vload_fp(ctx, i, insn); break; case IR_VSTORE_INT: ir_emit_vstore_int(ctx, i, insn); break; case IR_VSTORE_FP: ir_emit_vstore_fp(ctx, i, insn); break; case IR_LOAD_INT: ir_emit_load_int(ctx, i, insn); break; case IR_LOAD_FP: ir_emit_load_fp(ctx, i, insn); break; case IR_STORE_INT: ir_emit_store_int(ctx, i, insn); break; case IR_STORE_FP: ir_emit_store_fp(ctx, i, insn); break; case IR_ALLOCA: ir_emit_alloca(ctx, i, insn); break; default: IR_ASSERT(0 && "NIY rule/insruction"); break; } n = ir_operands_count(ctx, insn); n = 1 + (n >> 2); // support for multi-word instructions like MERGE and PHI i += n; insn += n; rule += n; } } if (data.rodata_label) { |.rodata } for (i = IR_UNUSED + 1, insn = ctx->ir_base - i; i < ctx->consts_count; i++, insn--) { if (insn->emit_const) { if (IR_IS_TYPE_FP(insn->type)) { int label = ctx->cfg_blocks_count + i; if (!data.rodata_label) { data.rodata_label = ctx->cfg_blocks_count + ctx->consts_count + 2; |.rodata |=>data.rodata_label: } if (insn->type == IR_DOUBLE) { |.align 8 |=>label: |.dword insn->val.u32, insn->val.u32_hi } else if (insn->type == IR_FLOAT) { |.align 4 |=>label: |.dword insn->val.u32 } else { IR_ASSERT(0); } } else if (insn->op == IR_STR) { int label = ctx->cfg_blocks_count + i; const char *str = ir_get_str(ctx, insn->val.addr); int i = 0; if (!data.rodata_label) { data.rodata_label = ctx->cfg_blocks_count + ctx->consts_count + 2; |.rodata |=>data.rodata_label: } |.align 8 |=>label: while (str[i]) { char c = str[i]; if (c == '\\') { if (str[i+1] == '\\') { i++; c = '\\'; } else if (str[i+1] == 'n') { i++; c = '\n'; } else if (str[i+1] == 'r') { i++; c = '\r'; } else if (str[i+1] == 't') { i++; c = '\t'; } } |.byte c i++; } |.byte 0 } else { IR_ASSERT(0); } } } if (data.rodata_label) { |.code } ret = dasm_link(&data.dasm_state, size); if (ret != DASM_S_OK) { return 0; } entry = ir_mem_mmap(4096); ir_mem_unprotect(entry, 4096); ret = dasm_encode(&data.dasm_state, entry); if (data.rodata_label) { int rodata = dasm_getpclabel(&data.dasm_state, data.rodata_label); *size = rodata; } else if (data.jmp_table_label) { int rodata = dasm_getpclabel(&data.dasm_state, data.jmp_table_label); *size = rodata; } if (ret != DASM_S_OK) { IR_ASSERT(0); ir_mem_unmap(entry, 4096); return NULL; } dasm_free(&data.dasm_state); ir_mem_flush(entry, 4096); ir_mem_protect(entry, 4096); return entry; }