diff --git a/py/emitviperx64.c b/py/emitviperx64.c index c2d2acee7d..250dd146c9 100644 --- a/py/emitviperx64.c +++ b/py/emitviperx64.c @@ -66,8 +66,8 @@ struct _emit_t { need_to_push_t need_to_push; vtype_kind_t last_vtype; - int last_r64; - int64_t last_i64; + int last_reg; + int64_t last_imm; scope_t *scope; @@ -173,13 +173,13 @@ static void stack_settle(emit_t *emit) { case NEED_TO_PUSH_R64: emit->stack_vtype[emit->stack_size] = emit->last_vtype; - asm_x64_mov_r64_to_local(emit->as, emit->last_r64, emit->stack_start + emit->stack_size); + asm_x64_mov_r64_to_local(emit->as, emit->last_reg, emit->stack_start + emit->stack_size); adjust_stack(emit, 1); break; case NEED_TO_PUSH_I64: emit->stack_vtype[emit->stack_size] = emit->last_vtype; - asm_x64_mov_i64_to_r64_optimised(emit->as, emit->last_i64, REG_RAX); + asm_x64_mov_i64_to_r64_optimised(emit->as, emit->last_imm, REG_RAX); asm_x64_mov_r64_to_local(emit->as, REG_RAX, emit->stack_start + emit->stack_size); adjust_stack(emit, 1); break; @@ -212,40 +212,40 @@ static vtype_kind_t peek_vtype(emit_t *emit) { } } -static void emit_pre_pop_r64(emit_t *emit, vtype_kind_t *vtype, int r64) { +static void emit_pre_pop_reg(emit_t *emit, vtype_kind_t *vtype, int reg) { switch (emit->need_to_push) { case NEED_TO_PUSH_NOTHING: *vtype = emit->stack_vtype[emit->stack_size - 1]; - asm_x64_mov_local_to_r64(emit->as, emit->stack_start + emit->stack_size - 1, r64); + asm_x64_mov_local_to_r64(emit->as, emit->stack_start + emit->stack_size - 1, reg); emit_pre_raw(emit, -1); break; case NEED_TO_PUSH_R64: emit_pre_raw(emit, 0); *vtype = emit->last_vtype; - if (emit->last_r64 != r64) { - asm_x64_mov_r64_to_r64(emit->as, emit->last_r64, r64); + if (emit->last_reg != reg) { + asm_x64_mov_r64_to_r64(emit->as, emit->last_reg, reg); } break; case NEED_TO_PUSH_I64: emit_pre_raw(emit, 0); *vtype = emit->last_vtype; - asm_x64_mov_i64_to_r64_optimised(emit->as, emit->last_i64, r64); + asm_x64_mov_i64_to_r64_optimised(emit->as, emit->last_imm, reg); break; } emit->need_to_push = NEED_TO_PUSH_NOTHING; } -static void emit_pre_pop_r64_r64(emit_t *emit, vtype_kind_t *vtypea, int r64a, vtype_kind_t *vtypeb, int r64b) { - emit_pre_pop_r64(emit, vtypea, r64a); +static void emit_pre_pop_reg_reg(emit_t *emit, vtype_kind_t *vtypea, int r64a, vtype_kind_t *vtypeb, int r64b) { + emit_pre_pop_reg(emit, vtypea, r64a); *vtypeb = emit->stack_vtype[emit->stack_size - 1]; asm_x64_mov_local_to_r64(emit->as, emit->stack_start + emit->stack_size - 1, r64b); adjust_stack(emit, -1); } -static void emit_pre_pop_r64_r64_r64(emit_t *emit, vtype_kind_t *vtypea, int r64a, vtype_kind_t *vtypeb, int r64b, vtype_kind_t *vtypec, int r64c) { - emit_pre_pop_r64(emit, vtypea, r64a); +static void emit_pre_pop_reg_reg_reg(emit_t *emit, vtype_kind_t *vtypea, int r64a, vtype_kind_t *vtypeb, int r64b, vtype_kind_t *vtypec, int r64c) { + emit_pre_pop_reg(emit, vtypea, r64a); *vtypeb = emit->stack_vtype[emit->stack_size - 1]; asm_x64_mov_local_to_r64(emit->as, emit->stack_start + emit->stack_size - 1, r64b); *vtypec = emit->stack_vtype[emit->stack_size - 2]; @@ -256,28 +256,28 @@ static void emit_pre_pop_r64_r64_r64(emit_t *emit, vtype_kind_t *vtypea, int r64 static void emit_post(emit_t *emit) { } -static void emit_post_push_r64(emit_t *emit, vtype_kind_t vtype, int r64) { +static void emit_post_push_reg(emit_t *emit, vtype_kind_t vtype, int reg) { emit->need_to_push = NEED_TO_PUSH_R64; emit->last_vtype = vtype; - emit->last_r64 = r64; + emit->last_reg = reg; } -static void emit_post_push_i64(emit_t *emit, vtype_kind_t vtype, int64_t i64) { +static void emit_post_push_imm(emit_t *emit, vtype_kind_t vtype, int64_t imm) { emit->need_to_push = NEED_TO_PUSH_I64; emit->last_vtype = vtype; - emit->last_i64 = i64; + emit->last_imm = imm; } -static void emit_post_push_r64_r64(emit_t *emit, vtype_kind_t vtypea, int r64a, vtype_kind_t vtypeb, int r64b) { +static void emit_post_push_reg_reg(emit_t *emit, vtype_kind_t vtypea, int r64a, vtype_kind_t vtypeb, int r64b) { emit->stack_vtype[emit->stack_size] = vtypea; asm_x64_mov_r64_to_local(emit->as, r64a, emit->stack_start + emit->stack_size); emit->need_to_push = NEED_TO_PUSH_R64; emit->last_vtype = vtypeb; - emit->last_r64 = r64b; + emit->last_reg = r64b; adjust_stack(emit, 1); } -static void emit_post_push_r64_r64_r64(emit_t *emit, vtype_kind_t vtypea, int r64a, vtype_kind_t vtypeb, int r64b, vtype_kind_t vtypec, int r64c) { +static void emit_post_push_reg_reg_reg(emit_t *emit, vtype_kind_t vtypea, int r64a, vtype_kind_t vtypeb, int r64b, vtype_kind_t vtypec, int r64c) { emit->stack_vtype[emit->stack_size] = vtypea; asm_x64_mov_r64_to_local(emit->as, r64a, emit->stack_start + emit->stack_size); emit->stack_vtype[emit->stack_size + 1] = vtypeb; @@ -287,7 +287,7 @@ static void emit_post_push_r64_r64_r64(emit_t *emit, vtype_kind_t vtypea, int r6 adjust_stack(emit, 3); } -static void emit_post_push_r64_r64_r64_r64(emit_t *emit, vtype_kind_t vtypea, int r64a, vtype_kind_t vtypeb, int r64b, vtype_kind_t vtypec, int r64c, vtype_kind_t vtyped, int r64d) { +static void emit_post_push_reg_reg_reg_reg(emit_t *emit, vtype_kind_t vtypea, int r64a, vtype_kind_t vtypeb, int r64b, vtype_kind_t vtypec, int r64c, vtype_kind_t vtyped, int r64d) { emit->stack_vtype[emit->stack_size] = vtypea; asm_x64_mov_r64_to_local(emit->as, r64a, emit->stack_start + emit->stack_size); emit->stack_vtype[emit->stack_size + 1] = vtypeb; @@ -300,17 +300,17 @@ static void emit_post_push_r64_r64_r64_r64(emit_t *emit, vtype_kind_t vtypea, in } // vtype of all n_pop objects is VTYPE_PYOBJ -static void emit_get_stack_pointer_to_r64_for_pop(emit_t *emit, int r64, int n_pop) { - asm_x64_mov_local_addr_to_r64(emit->as, emit->stack_start + emit->stack_size - 1, r64); +static void emit_get_stack_pointer_to_reg_for_pop(emit_t *emit, int reg, int n_pop) { + asm_x64_mov_local_addr_to_r64(emit->as, emit->stack_start + emit->stack_size - 1, reg); adjust_stack(emit, -n_pop); } // vtype of all n_push objects is VTYPE_PYOBJ -static void emit_get_stack_pointer_to_r64_for_push(emit_t *emit, int r64, int n_push) { +static void emit_get_stack_pointer_to_reg_for_push(emit_t *emit, int reg, int n_push) { for (int i = 0; i < n_push; i++) { emit->stack_vtype[emit->stack_size + i] = VTYPE_PYOBJ; } - asm_x64_mov_local_addr_to_r64(emit->as, emit->stack_start + emit->stack_size + n_push - 1, r64); + asm_x64_mov_local_addr_to_r64(emit->as, emit->stack_start + emit->stack_size + n_push - 1, reg); adjust_stack(emit, n_push); } @@ -318,8 +318,8 @@ static void emit_call(emit_t *emit, void *fun) { asm_x64_call_ind(emit->as, fun, REG_RAX); } -static void emit_call_with_i64_arg(emit_t *emit, void *fun, int64_t arg_val, int arg_r64) { - asm_x64_mov_i64_to_r64_optimised(emit->as, arg_val, arg_r64); +static void emit_call_with_imm_arg(emit_t *emit, void *fun, int64_t arg_val, int arg_reg) { + asm_x64_mov_i64_to_r64_optimised(emit->as, arg_val, arg_reg); asm_x64_call_ind(emit->as, fun, REG_RAX); } @@ -374,12 +374,12 @@ static void emit_viper_x64_load_const_tok(emit_t *emit, py_token_kind_t tok) { case PY_TOKEN_KW_TRUE: vtype = VTYPE_BOOL; val = 1; break; default: assert(0); vtype = 0; val = 0; // shouldn't happen } - emit_post_push_i64(emit, vtype, val); + emit_post_push_imm(emit, vtype, val); } static void emit_viper_x64_load_const_small_int(emit_t *emit, int arg) { emit_pre(emit); - emit_post_push_i64(emit, VTYPE_INT, arg); + emit_post_push_imm(emit, VTYPE_INT, arg); } static void emit_viper_x64_load_const_int(emit_t *emit, qstr qstr) { @@ -403,7 +403,7 @@ static void emit_viper_x64_load_const_str(emit_t *emit, qstr qstr, bool bytes) { // load a pointer to the asciiz string? assert(0); emit_pre(emit); - emit_post_push_i64(emit, VTYPE_PTR, (machine_uint_t)qstr_str(qstr)); + emit_post_push_imm(emit, VTYPE_PTR, (machine_uint_t)qstr_str(qstr)); } static void emit_viper_x64_load_const_verbatim_start(emit_t *emit) { @@ -442,24 +442,24 @@ static void emit_viper_x64_load_fast(emit_t *emit, qstr qstr, int local_num) { } if (local_num == 0) { emit_pre(emit); - emit_post_push_r64(emit, emit->local_vtype[local_num], REG_LOCAL_1); + emit_post_push_reg(emit, emit->local_vtype[local_num], REG_LOCAL_1); } else { emit_pre(emit); asm_x64_mov_local_to_r64(emit->as, local_num - 1, REG_RAX); - emit_post_push_r64(emit, emit->local_vtype[local_num], REG_RAX); + emit_post_push_reg(emit, emit->local_vtype[local_num], REG_RAX); } } static void emit_viper_x64_load_name(emit_t *emit, qstr qstr) { emit_pre(emit); - emit_call_with_i64_arg(emit, rt_load_name, qstr, REG_ARG_1); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); + emit_call_with_imm_arg(emit, rt_load_name, qstr, REG_ARG_1); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); } static void emit_viper_x64_load_global(emit_t *emit, qstr qstr) { emit_pre(emit); - emit_call_with_i64_arg(emit, rt_load_global, qstr, REG_ARG_1); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); + emit_call_with_imm_arg(emit, rt_load_global, qstr, REG_ARG_1); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); } static void emit_viper_x64_load_deref(emit_t *emit, qstr qstr) { @@ -479,18 +479,18 @@ static void emit_viper_x64_load_attr(emit_t *emit, qstr qstr) { // - pointer to structure: get member, quite easy // - Python object: call rt_load_attr, and needs to be typed to convert result vtype_kind_t vtype_base; - emit_pre_pop_r64(emit, &vtype_base, REG_ARG_1); // arg1 = base + emit_pre_pop_reg(emit, &vtype_base, REG_ARG_1); // arg1 = base assert(vtype_base == VTYPE_PYOBJ); - emit_call_with_i64_arg(emit, rt_load_attr, qstr, REG_ARG_2); // arg2 = attribute name - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); + emit_call_with_imm_arg(emit, rt_load_attr, qstr, REG_ARG_2); // arg2 = attribute name + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); } static void emit_viper_x64_load_method(emit_t *emit, qstr qstr) { vtype_kind_t vtype_base; - emit_pre_pop_r64(emit, &vtype_base, REG_ARG_1); // arg1 = base + emit_pre_pop_reg(emit, &vtype_base, REG_ARG_1); // arg1 = base assert(vtype_base == VTYPE_PYOBJ); - emit_get_stack_pointer_to_r64_for_push(emit, REG_ARG_3, 2); // arg3 = dest ptr - emit_call_with_i64_arg(emit, rt_load_method, qstr, REG_ARG_2); // arg2 = method name + emit_get_stack_pointer_to_reg_for_push(emit, REG_ARG_3, 2); // arg3 = dest ptr + emit_call_with_imm_arg(emit, rt_load_method, qstr, REG_ARG_2); // arg2 = method name } static void emit_viper_x64_load_build_class(emit_t *emit) { @@ -501,10 +501,10 @@ static void emit_viper_x64_load_build_class(emit_t *emit) { static void emit_viper_x64_store_fast(emit_t *emit, qstr qstr, int local_num) { vtype_kind_t vtype; if (local_num == 0) { - emit_pre_pop_r64(emit, &vtype, REG_LOCAL_1); + emit_pre_pop_reg(emit, &vtype, REG_LOCAL_1); emit_post(emit); } else { - emit_pre_pop_r64(emit, &vtype, REG_RAX); + emit_pre_pop_reg(emit, &vtype, REG_RAX); asm_x64_mov_r64_to_local(emit->as, REG_RAX, local_num - 1); emit_post(emit); } @@ -522,9 +522,9 @@ static void emit_viper_x64_store_fast(emit_t *emit, qstr qstr, int local_num) { static void emit_viper_x64_store_name(emit_t *emit, qstr qstr) { // rt_store_name, but needs conversion of object (maybe have rt_viper_store_name(obj, type)) vtype_kind_t vtype; - emit_pre_pop_r64(emit, &vtype, REG_ARG_2); + emit_pre_pop_reg(emit, &vtype, REG_ARG_2); assert(vtype == VTYPE_PYOBJ); - emit_call_with_i64_arg(emit, rt_store_name, qstr, REG_ARG_1); // arg1 = name + emit_call_with_imm_arg(emit, rt_store_name, qstr, REG_ARG_1); // arg1 = name emit_post(emit); } @@ -554,7 +554,7 @@ static void emit_viper_x64_store_subscr(emit_t *emit) { // - pointer to integers: store as per array // - Python object: call runtime with converted object or type info vtype_kind_t vtype_index, vtype_base, vtype_value; - emit_pre_pop_r64_r64_r64(emit, &vtype_index, REG_ARG_2, &vtype_base, REG_ARG_1, &vtype_value, REG_ARG_3); // index, base, value to store + emit_pre_pop_reg_reg_reg(emit, &vtype_index, REG_ARG_2, &vtype_base, REG_ARG_1, &vtype_value, REG_ARG_3); // index, base, value to store assert(vtype_index == VTYPE_PYOBJ); assert(vtype_base == VTYPE_PYOBJ); assert(vtype_value == VTYPE_PYOBJ); @@ -596,19 +596,19 @@ static void emit_viper_x64_delete_subscr(emit_t *emit) { static void emit_viper_x64_dup_top(emit_t *emit) { vtype_kind_t vtype; - emit_pre_pop_r64(emit, &vtype, REG_RAX); - emit_post_push_r64_r64(emit, vtype, REG_RAX, vtype, REG_RAX); + emit_pre_pop_reg(emit, &vtype, REG_RAX); + emit_post_push_reg_reg(emit, vtype, REG_RAX, vtype, REG_RAX); } static void emit_viper_x64_dup_top_two(emit_t *emit) { vtype_kind_t vtype1, vtype2; - emit_pre_pop_r64_r64(emit, &vtype1, REG_RAX, &vtype2, REG_RDI); - emit_post_push_r64_r64_r64_r64(emit, vtype2, REG_RDI, vtype1, REG_RAX, vtype2, REG_RDI, vtype1, REG_RAX); + emit_pre_pop_reg_reg(emit, &vtype1, REG_RAX, &vtype2, REG_RDI); + emit_post_push_reg_reg_reg_reg(emit, vtype2, REG_RDI, vtype1, REG_RAX, vtype2, REG_RDI, vtype1, REG_RAX); } static void emit_viper_x64_pop_top(emit_t *emit) { vtype_kind_t vtype; - emit_pre_pop_r64(emit, &vtype, REG_RAX); + emit_pre_pop_reg(emit, &vtype, REG_RAX); emit_post(emit); } @@ -618,8 +618,8 @@ static void emit_viper_x64_rot_two(emit_t *emit) { static void emit_viper_x64_rot_three(emit_t *emit) { vtype_kind_t vtype_rax, vtype_rdi, vtype_rsi; - emit_pre_pop_r64_r64_r64(emit, &vtype_rax, REG_RAX, &vtype_rdi, REG_RDI, &vtype_rsi, REG_RSI); - emit_post_push_r64_r64_r64(emit, vtype_rax, REG_RAX, vtype_rsi, REG_RSI, vtype_rdi, REG_RDI); + emit_pre_pop_reg_reg_reg(emit, &vtype_rax, REG_RAX, &vtype_rdi, REG_RDI, &vtype_rsi, REG_RSI); + emit_post_push_reg_reg_reg(emit, vtype_rax, REG_RAX, vtype_rsi, REG_RSI, vtype_rdi, REG_RDI); } static void emit_viper_x64_jump(emit_t *emit, int label) { @@ -631,12 +631,12 @@ static void emit_viper_x64_jump(emit_t *emit, int label) { static void emit_viper_x64_pop_jump_if_false(emit_t *emit, int label) { vtype_kind_t vtype = peek_vtype(emit); if (vtype == VTYPE_BOOL) { - emit_pre_pop_r64(emit, &vtype, REG_RAX); + emit_pre_pop_reg(emit, &vtype, REG_RAX); asm_x64_test_r8_with_r8(emit->as, REG_RAX, REG_RAX); asm_x64_jcc_label(emit->as, JCC_JZ, label); emit_post(emit); } else if (vtype == VTYPE_PYOBJ) { - emit_pre_pop_r64(emit, &vtype, REG_ARG_1); + emit_pre_pop_reg(emit, &vtype, REG_ARG_1); emit_call(emit, rt_is_true); asm_x64_test_r8_with_r8(emit->as, REG_RET, REG_RET); asm_x64_jcc_label(emit->as, JCC_JZ, label); @@ -707,22 +707,22 @@ static void emit_viper_x64_pop_except(emit_t *emit) { static void emit_viper_x64_unary_op(emit_t *emit, rt_unary_op_t op) { vtype_kind_t vtype; - emit_pre_pop_r64(emit, &vtype, REG_ARG_2); + emit_pre_pop_reg(emit, &vtype, REG_ARG_2); assert(vtype == VTYPE_PYOBJ); - emit_call_with_i64_arg(emit, rt_unary_op, op, REG_ARG_1); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); + emit_call_with_imm_arg(emit, rt_unary_op, op, REG_ARG_1); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); } static void emit_viper_x64_binary_op(emit_t *emit, rt_binary_op_t op) { vtype_kind_t vtype_lhs, vtype_rhs; - emit_pre_pop_r64_r64(emit, &vtype_rhs, REG_ARG_3, &vtype_lhs, REG_ARG_2); + emit_pre_pop_reg_reg(emit, &vtype_rhs, REG_ARG_3, &vtype_lhs, REG_ARG_2); if (vtype_lhs == VTYPE_INT && vtype_rhs == VTYPE_INT) { assert(op == RT_BINARY_OP_ADD); asm_x64_add_r64_to_r64(emit->as, REG_ARG_3, REG_ARG_2); - emit_post_push_r64(emit, VTYPE_INT, REG_ARG_2); + emit_post_push_reg(emit, VTYPE_INT, REG_ARG_2); } else if (vtype_lhs == VTYPE_PYOBJ && vtype_rhs == VTYPE_PYOBJ) { - emit_call_with_i64_arg(emit, rt_binary_op, op, REG_ARG_1); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); + emit_call_with_imm_arg(emit, rt_binary_op, op, REG_ARG_1); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); } else { printf("ViperTypeError: can't do binary op between types %d and %d\n", vtype_lhs, vtype_rhs); assert(0); @@ -731,16 +731,16 @@ static void emit_viper_x64_binary_op(emit_t *emit, rt_binary_op_t op) { static void emit_viper_x64_compare_op(emit_t *emit, rt_compare_op_t op) { vtype_kind_t vtype_lhs, vtype_rhs; - emit_pre_pop_r64_r64(emit, &vtype_rhs, REG_ARG_3, &vtype_lhs, REG_ARG_2); + emit_pre_pop_reg_reg(emit, &vtype_rhs, REG_ARG_3, &vtype_lhs, REG_ARG_2); if (vtype_lhs == VTYPE_INT && vtype_rhs == VTYPE_INT) { assert(op == RT_COMPARE_OP_LESS); asm_x64_xor_r64_to_r64(emit->as, REG_RET, REG_RET); asm_x64_cmp_r64_with_r64(emit->as, REG_ARG_3, REG_ARG_2); asm_x64_setcc_r8(emit->as, JCC_JL, REG_RET); - emit_post_push_r64(emit, VTYPE_BOOL, REG_RET); + emit_post_push_reg(emit, VTYPE_BOOL, REG_RET); } else if (vtype_lhs == VTYPE_PYOBJ && vtype_rhs == VTYPE_PYOBJ) { - emit_call_with_i64_arg(emit, rt_compare_op, op, REG_ARG_1); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); + emit_call_with_imm_arg(emit, rt_compare_op, op, REG_ARG_1); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); } else { printf("ViperTypeError: can't do comparison between types %d and %d\n", vtype_lhs, vtype_rhs); assert(0); @@ -755,9 +755,9 @@ static void emit_viper_x64_build_tuple(emit_t *emit, int n_args) { static void emit_viper_x64_build_list(emit_t *emit, int n_args) { emit_pre(emit); - emit_get_stack_pointer_to_r64_for_pop(emit, REG_ARG_2, n_args); // pointer to items in reverse order - emit_call_with_i64_arg(emit, rt_build_list, n_args, REG_ARG_1); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); // new list + emit_get_stack_pointer_to_reg_for_pop(emit, REG_ARG_2, n_args); // pointer to items in reverse order + emit_call_with_imm_arg(emit, rt_build_list, n_args, REG_ARG_1); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); // new list } static void emit_viper_x64_list_append(emit_t *emit, int list_index) { @@ -767,18 +767,18 @@ static void emit_viper_x64_list_append(emit_t *emit, int list_index) { static void emit_viper_x64_build_map(emit_t *emit, int n_args) { emit_pre(emit); - emit_call_with_i64_arg(emit, rt_build_map, n_args, REG_ARG_1); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); // new map + emit_call_with_imm_arg(emit, rt_build_map, n_args, REG_ARG_1); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); // new map } static void emit_viper_x64_store_map(emit_t *emit) { vtype_kind_t vtype_key, vtype_value, vtype_map; - emit_pre_pop_r64_r64_r64(emit, &vtype_key, REG_ARG_2, &vtype_value, REG_ARG_3, &vtype_map, REG_ARG_1); // key, value, map + emit_pre_pop_reg_reg_reg(emit, &vtype_key, REG_ARG_2, &vtype_value, REG_ARG_3, &vtype_map, REG_ARG_1); // key, value, map assert(vtype_key == VTYPE_PYOBJ); assert(vtype_value == VTYPE_PYOBJ); assert(vtype_map == VTYPE_PYOBJ); emit_call(emit, rt_store_map); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); // map + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); // map } static void emit_viper_x64_map_add(emit_t *emit, int map_index) { @@ -787,9 +787,9 @@ static void emit_viper_x64_map_add(emit_t *emit, int map_index) { static void emit_viper_x64_build_set(emit_t *emit, int n_args) { emit_pre(emit); - emit_get_stack_pointer_to_r64_for_pop(emit, REG_ARG_2, n_args); // pointer to items in reverse order - emit_call_with_i64_arg(emit, rt_build_set, n_args, REG_ARG_1); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); // new set + emit_get_stack_pointer_to_reg_for_pop(emit, REG_ARG_2, n_args); // pointer to items in reverse order + emit_call_with_imm_arg(emit, rt_build_set, n_args, REG_ARG_1); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); // new set } static void emit_viper_x64_set_add(emit_t *emit, int set_index) { @@ -810,8 +810,8 @@ static void emit_viper_x64_make_function(emit_t *emit, scope_t *scope, int n_dic // call runtime, with type info for args, or don't support dict/default params, or only support Python objects for them assert(n_default_params == 0 && n_dict_params == 0); emit_pre(emit); - emit_call_with_i64_arg(emit, rt_make_function_from_id, scope->unique_code_id, REG_ARG_1); - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); + emit_call_with_imm_arg(emit, rt_make_function_from_id, scope->unique_code_id, REG_ARG_1); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); } static void emit_viper_x64_make_closure(emit_t *emit, scope_t *scope, int n_dict_params, int n_default_params) { @@ -823,18 +823,18 @@ static void emit_viper_x64_call_function(emit_t *emit, int n_positional, int n_k assert(n_keyword == 0 && !have_star_arg && !have_dbl_star_arg); if (n_positional == 0) { vtype_kind_t vtype_fun; - emit_pre_pop_r64(emit, &vtype_fun, REG_ARG_1); // the function + emit_pre_pop_reg(emit, &vtype_fun, REG_ARG_1); // the function assert(vtype_fun == VTYPE_PYOBJ); emit_call(emit, rt_call_function_0); } else if (n_positional == 1) { vtype_kind_t vtype_fun, vtype_arg1; - emit_pre_pop_r64_r64(emit, &vtype_arg1, REG_ARG_2, &vtype_fun, REG_ARG_1); // the single argument, the function + emit_pre_pop_reg_reg(emit, &vtype_arg1, REG_ARG_2, &vtype_fun, REG_ARG_1); // the single argument, the function assert(vtype_fun == VTYPE_PYOBJ); assert(vtype_arg1 == VTYPE_PYOBJ); emit_call(emit, rt_call_function_1); } else if (n_positional == 2) { vtype_kind_t vtype_fun, vtype_arg1, vtype_arg2; - emit_pre_pop_r64_r64_r64(emit, &vtype_arg2, REG_ARG_3, &vtype_arg1, REG_ARG_2, &vtype_fun, REG_ARG_1); // the second argument, the first argument, the function + emit_pre_pop_reg_reg_reg(emit, &vtype_arg2, REG_ARG_3, &vtype_arg1, REG_ARG_2, &vtype_fun, REG_ARG_1); // the second argument, the first argument, the function assert(vtype_fun == VTYPE_PYOBJ); assert(vtype_arg1 == VTYPE_PYOBJ); assert(vtype_arg2 == VTYPE_PYOBJ); @@ -842,20 +842,20 @@ static void emit_viper_x64_call_function(emit_t *emit, int n_positional, int n_k } else { assert(0); } - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); } static void emit_viper_x64_call_method(emit_t *emit, int n_positional, int n_keyword, bool have_star_arg, bool have_dbl_star_arg) { assert(n_keyword == 0 && !have_star_arg && !have_dbl_star_arg); if (n_positional == 0) { vtype_kind_t vtype_meth, vtype_self; - emit_pre_pop_r64_r64(emit, &vtype_self, REG_ARG_2, &vtype_meth, REG_ARG_1); // the self object (or NULL), the method + emit_pre_pop_reg_reg(emit, &vtype_self, REG_ARG_2, &vtype_meth, REG_ARG_1); // the self object (or NULL), the method assert(vtype_meth == VTYPE_PYOBJ); assert(vtype_self == VTYPE_PYOBJ); emit_call(emit, rt_call_method_1); } else if (n_positional == 1) { vtype_kind_t vtype_meth, vtype_self, vtype_arg1; - emit_pre_pop_r64_r64_r64(emit, &vtype_arg1, REG_ARG_3, &vtype_self, REG_ARG_2, &vtype_meth, REG_ARG_1); // the first argument, the self object (or NULL), the method + emit_pre_pop_reg_reg_reg(emit, &vtype_arg1, REG_ARG_3, &vtype_self, REG_ARG_2, &vtype_meth, REG_ARG_1); // the first argument, the self object (or NULL), the method assert(vtype_meth == VTYPE_PYOBJ); assert(vtype_self == VTYPE_PYOBJ); assert(vtype_arg1 == VTYPE_PYOBJ); @@ -863,14 +863,14 @@ static void emit_viper_x64_call_method(emit_t *emit, int n_positional, int n_key } else { assert(0); } - emit_post_push_r64(emit, VTYPE_PYOBJ, REG_RET); + emit_post_push_reg(emit, VTYPE_PYOBJ, REG_RET); } static void emit_viper_x64_return_value(emit_t *emit) { // easy. since we don't know who we return to, just return the raw value. // runtime needs then to know our type signature, but I think that's possible. vtype_kind_t vtype; - emit_pre_pop_r64(emit, &vtype, REG_RAX); + emit_pre_pop_reg(emit, &vtype, REG_RAX); assert(vtype == VTYPE_PTR_NONE); emit->last_emit_was_return_value = true; //asm_x64_call_ind(emit->as, 0, REG_RAX); to seg fault for debugging with gdb