/* Copyright (C) 2001, 2009, 2010, 2011, 2012, 2013, * 2014, 2015 Free Software Foundation, Inc. * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public License * as published by the Free Software Foundation; either version 3 of * the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA * 02110-1301 USA */ /* This file is included in vm.c multiple times. */ #define UNPACK_8_8_8(op,a,b,c) \ do \ { \ a = (op >> 8) & 0xff; \ b = (op >> 16) & 0xff; \ c = op >> 24; \ } \ while (0) #define UNPACK_8_16(op,a,b) \ do \ { \ a = (op >> 8) & 0xff; \ b = op >> 16; \ } \ while (0) #define UNPACK_16_8(op,a,b) \ do \ { \ a = (op >> 8) & 0xffff; \ b = op >> 24; \ } \ while (0) #define UNPACK_12_12(op,a,b) \ do \ { \ a = (op >> 8) & 0xfff; \ b = op >> 20; \ } \ while (0) #define UNPACK_24(op,a) \ do \ { \ a = op >> 8; \ } \ while (0) /* Assign some registers by hand. There used to be a bigger list here, but it was never tested, and in the case of x86-32, was a source of compilation failures. It can be revived if it's useful, but my naive hope is that simply annotating the locals with "register" will be a sufficient hint to the compiler. */ #ifdef __GNUC__ # if defined __x86_64__ /* GCC 4.6 chooses %rbp for IP_REG and %rbx for SP_REG, which works well. Tell it to keep the jump table in a r12, which is callee-saved. */ # define JT_REG asm ("r12") # endif #endif #ifndef IP_REG # define IP_REG #endif #ifndef FP_REG # define FP_REG #endif #ifndef JT_REG # define JT_REG #endif #define VM_ASSERT(condition, handler) \ do { \ if (SCM_UNLIKELY (!(condition))) \ { \ SYNC_IP(); \ handler; \ } \ } while (0) #ifdef VM_ENABLE_ASSERTIONS # define ASSERT(condition) VM_ASSERT (condition, abort()) #else # define ASSERT(condition) #endif #if VM_USE_HOOKS #define RUN_HOOK(exp) \ do { \ if (SCM_UNLIKELY (vp->trace_level > 0)) \ { \ SYNC_IP (); \ exp; \ CACHE_SP (); \ } \ } while (0) #else #define RUN_HOOK(exp) #endif #define RUN_HOOK0(h) RUN_HOOK (vm_dispatch_##h##_hook (vp)) #define RUN_HOOK1(h, arg) RUN_HOOK (vm_dispatch_##h##_hook (vp, arg)) #define APPLY_HOOK() \ RUN_HOOK0 (apply) #define PUSH_CONTINUATION_HOOK() \ RUN_HOOK0 (push_continuation) #define POP_CONTINUATION_HOOK(old_fp) \ RUN_HOOK1 (pop_continuation, old_fp) #define NEXT_HOOK() \ RUN_HOOK0 (next) #define ABORT_CONTINUATION_HOOK() \ RUN_HOOK0 (abort) /* Virtual Machine The VM has three state bits: the instruction pointer (IP), the frame pointer (FP), and the stack pointer (SP). We cache the IP in a machine register, local to the VM, because it is used extensively by the VM. We do the same for SP. The FP is used more by code outside the VM than by the VM itself, we don't bother caching it locally. Keeping vp->ip in sync with the local IP would be a big lose, as it is updated so often. Instead of updating vp->ip all the time, we call SYNC_IP whenever we would need to know the IP of the top frame. In practice, we need to SYNC_IP whenever we call out of the VM to a function that would like to walk the stack, perhaps as the result of an exception. On the other hand, we do always keep vp->sp in sync with the local SP. One more thing. We allow the stack to move, when it expands. Therefore if you call out to a C procedure that could call Scheme code, or otherwise push anything on the stack, you will need to CACHE_SP afterwards to restore the possibly-changed stack pointer. */ #define SYNC_IP() vp->ip = (ip) #define CACHE_SP() sp = vp->sp #define CACHE_REGISTER() \ do { \ ip = vp->ip; \ CACHE_SP (); \ } while (0) /* Reserve stack space for a frame. Will check that there is sufficient stack space for N locals, including the procedure. Invoke after preparing the new frame and setting the fp and ip. If there is not enough space for this frame, we try to expand the stack, possibly relocating it somewhere else in the address space. Because of the possible relocation, no pointer into the stack besides FP is valid across an ALLOC_FRAME call. Be careful! */ #define ALLOC_FRAME(n) \ do { \ sp = vp->fp - (n); \ if (sp < vp->sp_min_since_gc) \ { \ if (SCM_UNLIKELY (sp < vp->stack_limit)) \ { \ SYNC_IP (); \ vm_expand_stack (vp, sp); \ CACHE_SP (); \ } \ else \ vp->sp_min_since_gc = vp->sp = sp; \ } \ else \ vp->sp = sp; \ } while (0) /* Reset the current frame to hold N locals. Used when we know that no stack expansion is needed. */ #define RESET_FRAME(n) \ do { \ vp->sp = sp = vp->fp - (n); \ if (sp < vp->sp_min_since_gc) \ vp->sp_min_since_gc = sp; \ } while (0) /* Compute the number of locals in the frame. At a call, this is equal to the number of actual arguments when a function is first called, plus one for the function. */ #define FRAME_LOCALS_COUNT() (vp->fp - sp) #define FRAME_LOCALS_COUNT_FROM(slot) (FRAME_LOCALS_COUNT () - slot) /* Restore registers after returning from a frame. */ #define RESTORE_FRAME() \ do { \ } while (0) #ifdef HAVE_LABELS_AS_VALUES # define BEGIN_DISPATCH_SWITCH /* */ # define END_DISPATCH_SWITCH /* */ # define NEXT(n) \ do \ { \ ip += n; \ NEXT_HOOK (); \ op = *ip; \ goto *jump_table[op & 0xff]; \ } \ while (0) # define VM_DEFINE_OP(opcode, tag, name, meta) \ op_##tag: #else # define BEGIN_DISPATCH_SWITCH \ vm_start: \ NEXT_HOOK (); \ op = *ip; \ switch (op & 0xff) \ { # define END_DISPATCH_SWITCH \ } # define NEXT(n) \ do \ { \ ip += n; \ goto vm_start; \ } \ while (0) # define VM_DEFINE_OP(opcode, tag, name, meta) \ op_##tag: \ case opcode: #endif #define FP_SLOT(i) SCM_FRAME_SLOT (vp->fp, i) #define FP_REF(i) SCM_FRAME_LOCAL (vp->fp, i) #define FP_SET(i,o) SCM_FRAME_LOCAL (vp->fp, i) = o #define SP_REF_SLOT(i) (sp[i]) #define SP_SET_SLOT(i,o) (sp[i] = o) #define SP_REF(i) (sp[i].as_scm) #define SP_SET(i,o) (sp[i].as_scm = o) #define SP_REF_F64(i) (sp[i].as_f64) #define SP_SET_F64(i,o) (sp[i].as_f64 = o) #define SP_REF_U64(i) (sp[i].as_u64) #define SP_SET_U64(i,o) (sp[i].as_u64 = o) #define SP_REF_S64(i) (sp[i].as_s64) #define SP_SET_S64(i,o) (sp[i].as_s64 = o) #define VARIABLE_REF(v) SCM_VARIABLE_REF (v) #define VARIABLE_SET(v,o) SCM_VARIABLE_SET (v, o) #define VARIABLE_BOUNDP(v) (!scm_is_eq (VARIABLE_REF (v), SCM_UNDEFINED)) #define BR_NARGS(rel) \ scm_t_uint32 expected; \ UNPACK_24 (op, expected); \ if (FRAME_LOCALS_COUNT() rel expected) \ { \ scm_t_int32 offset = ip[1]; \ offset >>= 8; /* Sign-extending shift. */ \ NEXT (offset); \ } \ NEXT (2) #define BR_UNARY(x, exp) \ scm_t_uint32 test; \ SCM x; \ UNPACK_24 (op, test); \ x = SP_REF (test); \ if ((ip[1] & 0x1) ? !(exp) : (exp)) \ { \ scm_t_int32 offset = ip[1]; \ offset >>= 8; /* Sign-extending shift. */ \ NEXT (offset); \ } \ NEXT (2) #define BR_BINARY(x, y, exp) \ scm_t_uint32 a, b; \ SCM x, y; \ UNPACK_24 (op, a); \ UNPACK_24 (ip[1], b); \ x = SP_REF (a); \ y = SP_REF (b); \ if ((ip[2] & 0x1) ? !(exp) : (exp)) \ { \ scm_t_int32 offset = ip[2]; \ offset >>= 8; /* Sign-extending shift. */ \ NEXT (offset); \ } \ NEXT (3) #define BR_ARITHMETIC(crel,srel) \ { \ scm_t_uint32 a, b; \ SCM x, y; \ UNPACK_24 (op, a); \ UNPACK_24 (ip[1], b); \ x = SP_REF (a); \ y = SP_REF (b); \ if (SCM_I_INUMP (x) && SCM_I_INUMP (y)) \ { \ scm_t_signed_bits x_bits = SCM_UNPACK (x); \ scm_t_signed_bits y_bits = SCM_UNPACK (y); \ if ((ip[2] & 0x1) ? !(x_bits crel y_bits) : (x_bits crel y_bits)) \ { \ scm_t_int32 offset = ip[2]; \ offset >>= 8; /* Sign-extending shift. */ \ NEXT (offset); \ } \ NEXT (3); \ } \ else \ { \ SCM res; \ SYNC_IP (); \ res = srel (x, y); \ CACHE_SP (); \ if ((ip[2] & 0x1) ? scm_is_false (res) : scm_is_true (res)) \ { \ scm_t_int32 offset = ip[2]; \ offset >>= 8; /* Sign-extending shift. */ \ NEXT (offset); \ } \ NEXT (3); \ } \ } #define BR_U64_ARITHMETIC(crel) \ { \ scm_t_uint32 a, b; \ scm_t_uint64 x, y; \ UNPACK_24 (op, a); \ UNPACK_24 (ip[1], b); \ x = SP_REF_U64 (a); \ y = SP_REF_U64 (b); \ if ((ip[2] & 0x1) ? !(x crel y) : (x crel y)) \ { \ scm_t_int32 offset = ip[2]; \ offset >>= 8; /* Sign-extending shift. */ \ NEXT (offset); \ } \ NEXT (3); \ } #define BR_F64_ARITHMETIC(crel) \ { \ scm_t_uint32 a, b; \ double x, y; \ UNPACK_24 (op, a); \ UNPACK_24 (ip[1], b); \ x = SP_REF_F64 (a); \ y = SP_REF_F64 (b); \ if ((ip[2] & 0x1) ? !(x crel y) : (x crel y)) \ { \ scm_t_int32 offset = ip[2]; \ offset >>= 8; /* Sign-extending shift. */ \ NEXT (offset); \ } \ NEXT (3); \ } #define ARGS1(a1) \ scm_t_uint16 dst, src; \ SCM a1; \ UNPACK_12_12 (op, dst, src); \ a1 = SP_REF (src) #define ARGS2(a1, a2) \ scm_t_uint8 dst, src1, src2; \ SCM a1, a2; \ UNPACK_8_8_8 (op, dst, src1, src2); \ a1 = SP_REF (src1); \ a2 = SP_REF (src2) #define RETURN(x) \ do { SP_SET (dst, x); NEXT (1); } while (0) #define RETURN_EXP(exp) \ do { SCM __x; SYNC_IP (); __x = exp; CACHE_SP (); RETURN (__x); } while (0) /* The maximum/minimum tagged integers. */ #define INUM_MAX \ ((scm_t_signed_bits) SCM_UNPACK (SCM_I_MAKINUM (SCM_MOST_POSITIVE_FIXNUM))) #define INUM_MIN \ ((scm_t_signed_bits) SCM_UNPACK (SCM_I_MAKINUM (SCM_MOST_NEGATIVE_FIXNUM))) #define INUM_STEP \ ((scm_t_signed_bits) SCM_UNPACK (SCM_INUM1) \ - (scm_t_signed_bits) SCM_UNPACK (SCM_INUM0)) #define BINARY_INTEGER_OP(CFUNC,SFUNC) \ { \ ARGS2 (x, y); \ if (SCM_I_INUMP (x) && SCM_I_INUMP (y)) \ { \ scm_t_int64 n = SCM_I_INUM (x) CFUNC SCM_I_INUM (y); \ if (SCM_FIXABLE (n)) \ RETURN (SCM_I_MAKINUM (n)); \ } \ RETURN_EXP (SFUNC (x, y)); \ } #define VM_VALIDATE(x, pred, proc, what) \ VM_ASSERT (pred (x), vm_error_not_a_ ## what (proc, x)) #define VM_VALIDATE_ATOMIC_BOX(x, proc) \ VM_VALIDATE (x, scm_is_atomic_box, proc, atomic_box) #define VM_VALIDATE_BYTEVECTOR(x, proc) \ VM_VALIDATE (x, SCM_BYTEVECTOR_P, proc, bytevector) #define VM_VALIDATE_CHAR(x, proc) \ VM_VALIDATE (x, SCM_CHARP, proc, char) #define VM_VALIDATE_PAIR(x, proc) \ VM_VALIDATE (x, scm_is_pair, proc, pair) #define VM_VALIDATE_STRING(obj, proc) \ VM_VALIDATE (obj, scm_is_string, proc, string) #define VM_VALIDATE_STRUCT(obj, proc) \ VM_VALIDATE (obj, SCM_STRUCTP, proc, struct) #define VM_VALIDATE_VARIABLE(obj, proc) \ VM_VALIDATE (obj, SCM_VARIABLEP, proc, variable) #define VM_VALIDATE_VECTOR(obj, proc) \ VM_VALIDATE (obj, SCM_I_IS_VECTOR, proc, vector) #define VM_VALIDATE_INDEX(u64, size, proc) \ VM_ASSERT (u64 < size, vm_error_out_of_range_uint64 (proc, u64)) /* Return true (non-zero) if PTR has suitable alignment for TYPE. */ #define ALIGNED_P(ptr, type) \ ((scm_t_uintptr) (ptr) % alignof_type (type) == 0) static SCM VM_NAME (scm_i_thread *thread, struct scm_vm *vp, scm_i_jmp_buf *registers, int resume) { /* Instruction pointer: A pointer to the opcode that is currently running. */ register scm_t_uint32 *ip IP_REG; /* Stack pointer: A pointer to the hot end of the stack, off of which we index arguments and local variables. Pushed at function calls, popped on returns. */ register union scm_vm_stack_element *sp FP_REG; /* Current opcode: A cache of *ip. */ register scm_t_uint32 op; #ifdef HAVE_LABELS_AS_VALUES static const void *jump_table_[256] = { #define LABEL_ADDR(opcode, tag, name, meta) &&op_##tag, FOR_EACH_VM_OPERATION(LABEL_ADDR) #undef LABEL_ADDR }; register const void **jump_table JT_REG; /* Attempt to keep JUMP_TABLE_POINTER in a register. This saves one load instruction at each instruction dispatch. */ jump_table = jump_table_; #endif /* Load VM registers. */ CACHE_REGISTER (); /* Usually a call to the VM happens on application, with the boot continuation on the next frame. Sometimes it happens after a non-local exit however; in that case the VM state is all set up, and we have but to jump to the next opcode. */ if (SCM_UNLIKELY (resume)) NEXT (0); if (SCM_LIKELY (SCM_PROGRAM_P (FP_REF (0)))) ip = SCM_PROGRAM_CODE (FP_REF (0)); else ip = (scm_t_uint32 *) vm_apply_non_program_code; APPLY_HOOK (); NEXT (0); BEGIN_DISPATCH_SWITCH; /* * Call and return */ /* halt _:24 * * Bring the VM to a halt, returning all the values from the stack. */ VM_DEFINE_OP (0, halt, "halt", OP1 (X32)) { /* Boot closure in r0, empty frame in r1/r2, proc in r3, values from r4. */ scm_t_uint32 nvals = FRAME_LOCALS_COUNT_FROM (4); SCM ret; if (nvals == 1) ret = FP_REF (4); else { scm_t_uint32 n; ret = SCM_EOL; SYNC_IP (); for (n = nvals; n > 0; n--) ret = scm_inline_cons (thread, FP_REF (4 + n - 1), ret); ret = scm_values (ret); } vp->ip = SCM_FRAME_RETURN_ADDRESS (vp->fp); vp->sp = SCM_FRAME_PREVIOUS_SP (vp->fp); vp->fp = SCM_FRAME_DYNAMIC_LINK (vp->fp); return ret; } /* call proc:24 _:8 nlocals:24 * * Call a procedure. PROC is the local corresponding to a procedure. * The two values below PROC will be overwritten by the saved call * frame data. The new frame will have space for NLOCALS locals: one * for the procedure, and the rest for the arguments which should * already have been pushed on. * * When the call returns, execution proceeds with the next * instruction. There may be any number of values on the return * stack; the precise number can be had by subtracting the address of * PROC from the post-call SP. */ VM_DEFINE_OP (1, call, "call", OP2 (X8_F24, X8_C24)) { scm_t_uint32 proc, nlocals; union scm_vm_stack_element *old_fp; UNPACK_24 (op, proc); UNPACK_24 (ip[1], nlocals); PUSH_CONTINUATION_HOOK (); old_fp = vp->fp; vp->fp = SCM_FRAME_SLOT (old_fp, proc - 1); SCM_FRAME_SET_DYNAMIC_LINK (vp->fp, old_fp); SCM_FRAME_SET_RETURN_ADDRESS (vp->fp, ip + 2); RESET_FRAME (nlocals); if (SCM_LIKELY (SCM_PROGRAM_P (FP_REF (0)))) ip = SCM_PROGRAM_CODE (FP_REF (0)); else ip = (scm_t_uint32 *) vm_apply_non_program_code; APPLY_HOOK (); NEXT (0); } /* call-label proc:24 _:8 nlocals:24 label:32 * * Call a procedure in the same compilation unit. * * This instruction is just like "call", except that instead of * dereferencing PROC to find the call target, the call target is * known to be at LABEL, a signed 32-bit offset in 32-bit units from * the current IP. Since PROC is not dereferenced, it may be some * other representation of the closure. */ VM_DEFINE_OP (2, call_label, "call-label", OP3 (X8_F24, X8_C24, L32)) { scm_t_uint32 proc, nlocals; scm_t_int32 label; union scm_vm_stack_element *old_fp; UNPACK_24 (op, proc); UNPACK_24 (ip[1], nlocals); label = ip[2]; PUSH_CONTINUATION_HOOK (); old_fp = vp->fp; vp->fp = SCM_FRAME_SLOT (old_fp, proc - 1); SCM_FRAME_SET_DYNAMIC_LINK (vp->fp, old_fp); SCM_FRAME_SET_RETURN_ADDRESS (vp->fp, ip + 3); RESET_FRAME (nlocals); ip += label; APPLY_HOOK (); NEXT (0); } /* tail-call nlocals:24 * * Tail-call a procedure. Requires that the procedure and all of the * arguments have already been shuffled into position. Will reset the * frame to NLOCALS. */ VM_DEFINE_OP (3, tail_call, "tail-call", OP1 (X8_C24)) { scm_t_uint32 nlocals; UNPACK_24 (op, nlocals); RESET_FRAME (nlocals); if (SCM_LIKELY (SCM_PROGRAM_P (FP_REF (0)))) ip = SCM_PROGRAM_CODE (FP_REF (0)); else ip = (scm_t_uint32 *) vm_apply_non_program_code; APPLY_HOOK (); NEXT (0); } /* tail-call-label nlocals:24 label:32 * * Tail-call a known procedure. As call is to call-label, tail-call * is to tail-call-label. */ VM_DEFINE_OP (4, tail_call_label, "tail-call-label", OP2 (X8_C24, L32)) { scm_t_uint32 nlocals; scm_t_int32 label; UNPACK_24 (op, nlocals); label = ip[1]; RESET_FRAME (nlocals); ip += label; APPLY_HOOK (); NEXT (0); } /* tail-call/shuffle from:24 * * Tail-call a procedure. The procedure should already be set to slot * 0. The rest of the args are taken from the frame, starting at * FROM, shuffled down to start at slot 0. This is part of the * implementation of the call-with-values builtin. */ VM_DEFINE_OP (5, tail_call_shuffle, "tail-call/shuffle", OP1 (X8_F24)) { scm_t_uint32 n, from, nlocals; UNPACK_24 (op, from); VM_ASSERT (from > 0, abort ()); nlocals = FRAME_LOCALS_COUNT (); for (n = 0; from + n < nlocals; n++) FP_SET (n + 1, FP_REF (from + n)); RESET_FRAME (n + 1); if (SCM_LIKELY (SCM_PROGRAM_P (FP_REF (0)))) ip = SCM_PROGRAM_CODE (FP_REF (0)); else ip = (scm_t_uint32 *) vm_apply_non_program_code; APPLY_HOOK (); NEXT (0); } /* receive dst:12 proc:12 _:8 nlocals:24 * * Receive a single return value from a call whose procedure was in * PROC, asserting that the call actually returned at least one * value. Afterwards, resets the frame to NLOCALS locals. */ VM_DEFINE_OP (6, receive, "receive", OP2 (X8_F12_F12, X8_C24) | OP_DST) { scm_t_uint16 dst, proc; scm_t_uint32 nlocals; UNPACK_12_12 (op, dst, proc); UNPACK_24 (ip[1], nlocals); VM_ASSERT (FRAME_LOCALS_COUNT () > proc + 1, vm_error_no_values ()); FP_SET (dst, FP_REF (proc + 1)); RESET_FRAME (nlocals); NEXT (2); } /* receive-values proc:24 allow-extra?:1 _:7 nvalues:24 * * Receive a return of multiple values from a call whose procedure was * in PROC. If fewer than NVALUES values were returned, signal an * error. Unless ALLOW-EXTRA? is true, require that the number of * return values equals NVALUES exactly. After receive-values has * run, the values can be copied down via `mov'. */ VM_DEFINE_OP (7, receive_values, "receive-values", OP2 (X8_F24, B1_X7_C24)) { scm_t_uint32 proc, nvalues; UNPACK_24 (op, proc); UNPACK_24 (ip[1], nvalues); if (ip[1] & 0x1) VM_ASSERT (FRAME_LOCALS_COUNT () > proc + nvalues, vm_error_not_enough_values ()); else VM_ASSERT (FRAME_LOCALS_COUNT () == proc + 1 + nvalues, vm_error_wrong_number_of_values (nvalues)); NEXT (2); } VM_DEFINE_OP (8, unused_8, NULL, NOP) { vm_error_bad_instruction (op); abort (); /* never reached */ } /* return-values nlocals:24 * * Return a number of values from a call frame. This opcode * corresponds to an application of `values' in tail position. As * with tail calls, we expect that the values have already been * shuffled down to a contiguous array starting at slot 1. * If NLOCALS is not zero, we also reset the frame to hold NLOCALS * values. */ VM_DEFINE_OP (9, return_values, "return-values", OP1 (X8_C24)) { union scm_vm_stack_element *old_fp; scm_t_uint32 nlocals; UNPACK_24 (op, nlocals); if (nlocals) RESET_FRAME (nlocals); old_fp = vp->fp; ip = SCM_FRAME_RETURN_ADDRESS (vp->fp); vp->fp = SCM_FRAME_DYNAMIC_LINK (vp->fp); /* Clear stack frame. */ old_fp[0].as_scm = SCM_BOOL_F; old_fp[1].as_scm = SCM_BOOL_F; POP_CONTINUATION_HOOK (old_fp); NEXT (0); } /* * Specialized call stubs */ /* subr-call _:24 * * Call a subr, passing all locals in this frame as arguments. Return * from the calling frame. This instruction is part of the * trampolines created in gsubr.c, and is not generated by the * compiler. */ VM_DEFINE_OP (10, subr_call, "subr-call", OP1 (X32)) { SCM ret; SYNC_IP (); ret = scm_apply_subr (sp, FRAME_LOCALS_COUNT ()); CACHE_SP (); if (SCM_UNLIKELY (SCM_VALUESP (ret))) { SCM vals = scm_struct_ref (ret, SCM_INUM0); long len = scm_ilength (vals); ALLOC_FRAME (1 + len); while (len--) { SP_SET (len, SCM_CAR (vals)); vals = SCM_CDR (vals); } NEXT (1); } else { ALLOC_FRAME (2); SP_SET (0, ret); NEXT (1); } } /* foreign-call cif-idx:12 ptr-idx:12 * * Call a foreign function. Fetch the CIF and foreign pointer from * CIF-IDX and PTR-IDX, both free variables. Return from the calling * frame. Arguments are taken from the stack. This instruction is * part of the trampolines created by the FFI, and is not generated by * the compiler. */ VM_DEFINE_OP (11, foreign_call, "foreign-call", OP1 (X8_C12_C12)) { scm_t_uint16 cif_idx, ptr_idx; int err = 0; SCM closure, cif, pointer, ret; UNPACK_12_12 (op, cif_idx, ptr_idx); closure = FP_REF (0); cif = SCM_PROGRAM_FREE_VARIABLE_REF (closure, cif_idx); pointer = SCM_PROGRAM_FREE_VARIABLE_REF (closure, ptr_idx); SYNC_IP (); ret = scm_i_foreign_call (cif, pointer, &err, sp); CACHE_SP (); ALLOC_FRAME (3); SP_SET (1, ret); SP_SET (0, scm_from_int (err)); NEXT (1); } /* continuation-call contregs:24 * * Return to a continuation, nonlocally. The arguments to the * continuation are taken from the stack. CONTREGS is a free variable * containing the reified continuation. This instruction is part of * the implementation of undelimited continuations, and is not * generated by the compiler. */ VM_DEFINE_OP (12, continuation_call, "continuation-call", OP1 (X8_C24)) { SCM contregs; scm_t_uint32 contregs_idx; UNPACK_24 (op, contregs_idx); contregs = SCM_PROGRAM_FREE_VARIABLE_REF (FP_REF (0), contregs_idx); SYNC_IP (); scm_i_check_continuation (contregs); vm_return_to_continuation (scm_i_contregs_vp (contregs), scm_i_contregs_vm_cont (contregs), FRAME_LOCALS_COUNT_FROM (1), sp); scm_i_reinstate_continuation (contregs); /* no NEXT */ abort (); } /* compose-continuation cont:24 * * Compose a partial continuation with the current continuation. The * arguments to the continuation are taken from the stack. CONT is a * free variable containing the reified continuation. This * instruction is part of the implementation of partial continuations, * and is not generated by the compiler. */ VM_DEFINE_OP (13, compose_continuation, "compose-continuation", OP1 (X8_C24)) { SCM vmcont; scm_t_uint32 cont_idx; UNPACK_24 (op, cont_idx); vmcont = SCM_PROGRAM_FREE_VARIABLE_REF (FP_REF (0), cont_idx); SYNC_IP (); VM_ASSERT (SCM_VM_CONT_REWINDABLE_P (vmcont), vm_error_continuation_not_rewindable (vmcont)); vm_reinstate_partial_continuation (vp, vmcont, FRAME_LOCALS_COUNT_FROM (1), &thread->dynstack, registers); CACHE_REGISTER (); NEXT (0); } /* tail-apply _:24 * * Tail-apply the procedure in local slot 0 to the rest of the * arguments. This instruction is part of the implementation of * `apply', and is not generated by the compiler. */ VM_DEFINE_OP (14, tail_apply, "tail-apply", OP1 (X32)) { int i, list_idx, list_len, nlocals; SCM list; nlocals = FRAME_LOCALS_COUNT (); // At a minimum, there should be apply, f, and the list. VM_ASSERT (nlocals >= 3, abort ()); list_idx = nlocals - 1; list = FP_REF (list_idx); list_len = scm_ilength (list); VM_ASSERT (list_len >= 0, vm_error_apply_to_non_list (list)); nlocals = nlocals - 2 + list_len; ALLOC_FRAME (nlocals); for (i = 1; i < list_idx; i++) FP_SET (i - 1, FP_REF (i)); /* Null out these slots, just in case there are less than 2 elements in the list. */ FP_SET (list_idx - 1, SCM_UNDEFINED); FP_SET (list_idx, SCM_UNDEFINED); for (i = 0; i < list_len; i++, list = SCM_CDR (list)) FP_SET (list_idx - 1 + i, SCM_CAR (list)); if (SCM_LIKELY (SCM_PROGRAM_P (FP_REF (0)))) ip = SCM_PROGRAM_CODE (FP_REF (0)); else ip = (scm_t_uint32 *) vm_apply_non_program_code; APPLY_HOOK (); NEXT (0); } /* call/cc _:24 * * Capture the current continuation, and tail-apply the procedure in * local slot 1 to it. This instruction is part of the implementation * of `call/cc', and is not generated by the compiler. */ VM_DEFINE_OP (15, call_cc, "call/cc", OP1 (X32)) { SCM vm_cont, cont; scm_t_dynstack *dynstack; int first; SYNC_IP (); dynstack = scm_dynstack_capture_all (&thread->dynstack); vm_cont = scm_i_vm_capture_stack (vp->stack_top, SCM_FRAME_DYNAMIC_LINK (vp->fp), SCM_FRAME_PREVIOUS_SP (vp->fp), SCM_FRAME_RETURN_ADDRESS (vp->fp), dynstack, 0); /* FIXME: Seems silly to capture the registers here, when they are already captured in the registers local, which here we are copying out to the heap; and likewise, the setjmp(®isters) code already has the non-local return handler. But oh well! */ cont = scm_i_make_continuation (&first, vp, vm_cont); if (first) { RESET_FRAME (2); SP_SET (1, SP_REF (0)); SP_SET (0, cont); if (SCM_LIKELY (SCM_PROGRAM_P (SP_REF (1)))) ip = SCM_PROGRAM_CODE (SP_REF (1)); else ip = (scm_t_uint32 *) vm_apply_non_program_code; APPLY_HOOK (); NEXT (0); } else { CACHE_REGISTER (); ABORT_CONTINUATION_HOOK (); NEXT (0); } } /* abort _:24 * * Abort to a prompt handler. The tag is expected in r1, and the rest * of the values in the frame are returned to the prompt handler. * This corresponds to a tail application of abort-to-prompt. */ VM_DEFINE_OP (16, abort, "abort", OP1 (X32)) { scm_t_uint32 nlocals = FRAME_LOCALS_COUNT (); ASSERT (nlocals >= 2); /* FIXME: Really we should capture the caller's registers. Until then, manually advance the IP so that when the prompt resumes, it continues with the next instruction. */ ip++; SYNC_IP (); vm_abort (vp, FP_REF (1), nlocals - 2, registers); /* vm_abort should not return */ abort (); } /* builtin-ref dst:12 idx:12 * * Load a builtin stub by index into DST. */ VM_DEFINE_OP (17, builtin_ref, "builtin-ref", OP1 (X8_S12_C12) | OP_DST) { scm_t_uint16 dst, idx; UNPACK_12_12 (op, dst, idx); SP_SET (dst, scm_vm_builtin_ref (idx)); NEXT (1); } /* * Function prologues */ /* br-if-nargs-ne expected:24 _:8 offset:24 * br-if-nargs-lt expected:24 _:8 offset:24 * br-if-nargs-gt expected:24 _:8 offset:24 * * If the number of actual arguments is not equal, less than, or greater * than EXPECTED, respectively, add OFFSET, a signed 24-bit number, to * the current instruction pointer. */ VM_DEFINE_OP (18, br_if_nargs_ne, "br-if-nargs-ne", OP2 (X8_C24, X8_L24)) { BR_NARGS (!=); } VM_DEFINE_OP (19, br_if_nargs_lt, "br-if-nargs-lt", OP2 (X8_C24, X8_L24)) { BR_NARGS (<); } VM_DEFINE_OP (20, br_if_nargs_gt, "br-if-nargs-gt", OP2 (X8_C24, X8_L24)) { BR_NARGS (>); } /* assert-nargs-ee expected:24 * assert-nargs-ge expected:24 * assert-nargs-le expected:24 * * If the number of actual arguments is not ==, >=, or <= EXPECTED, * respectively, signal an error. */ VM_DEFINE_OP (21, assert_nargs_ee, "assert-nargs-ee", OP1 (X8_C24)) { scm_t_uint32 expected; UNPACK_24 (op, expected); VM_ASSERT (FRAME_LOCALS_COUNT () == expected, vm_error_wrong_num_args (FP_REF (0))); NEXT (1); } VM_DEFINE_OP (22, assert_nargs_ge, "assert-nargs-ge", OP1 (X8_C24)) { scm_t_uint32 expected; UNPACK_24 (op, expected); VM_ASSERT (FRAME_LOCALS_COUNT () >= expected, vm_error_wrong_num_args (FP_REF (0))); NEXT (1); } VM_DEFINE_OP (23, assert_nargs_le, "assert-nargs-le", OP1 (X8_C24)) { scm_t_uint32 expected; UNPACK_24 (op, expected); VM_ASSERT (FRAME_LOCALS_COUNT () <= expected, vm_error_wrong_num_args (FP_REF (0))); NEXT (1); } /* alloc-frame nlocals:24 * * Ensure that there is space on the stack for NLOCALS local variables, * setting them all to SCM_UNDEFINED, except those nargs values that * were passed as arguments and procedure. */ VM_DEFINE_OP (24, alloc_frame, "alloc-frame", OP1 (X8_C24)) { scm_t_uint32 nlocals, nargs; UNPACK_24 (op, nlocals); nargs = FRAME_LOCALS_COUNT (); ALLOC_FRAME (nlocals); while (nlocals-- > nargs) FP_SET (nlocals, SCM_UNDEFINED); NEXT (1); } /* reset-frame nlocals:24 * * Like alloc-frame, but doesn't check that the stack is big enough. * Used to reset the frame size to something less than the size that * was previously set via alloc-frame. */ VM_DEFINE_OP (25, reset_frame, "reset-frame", OP1 (X8_C24)) { scm_t_uint32 nlocals; UNPACK_24 (op, nlocals); RESET_FRAME (nlocals); NEXT (1); } /* push src:24 * * Push SRC onto the stack. */ VM_DEFINE_OP (26, push, "push", OP1 (X8_S24)) { scm_t_uint32 src; union scm_vm_stack_element val; /* FIXME: The compiler currently emits "push" for SCM, F64, U64, and S64 variables. However SCM values are the usual case, and on a 32-bit machine it might be cheaper to move a SCM than to move a 64-bit number. */ UNPACK_24 (op, src); val = SP_REF_SLOT (src); ALLOC_FRAME (FRAME_LOCALS_COUNT () + 1); SP_SET_SLOT (0, val); NEXT (1); } /* pop dst:24 * * Pop the stack, storing to DST. */ VM_DEFINE_OP (27, pop, "pop", OP1 (X8_S24) | OP_DST) { scm_t_uint32 dst; union scm_vm_stack_element val; /* FIXME: The compiler currently emits "pop" for SCM, F64, U64, and S64 variables. However SCM values are the usual case, and on a 32-bit machine it might be cheaper to move a SCM than to move a 64-bit number. */ UNPACK_24 (op, dst); val = SP_REF_SLOT (0); vp->sp = sp = sp + 1; SP_SET_SLOT (dst, val); NEXT (1); } /* drop count:24 * * Drop some number of values from the stack. */ VM_DEFINE_OP (28, drop, "drop", OP1 (X8_C24)) { scm_t_uint32 count; UNPACK_24 (op, count); vp->sp = sp = sp + count; NEXT (1); } /* assert-nargs-ee/locals expected:12 nlocals:12 * * Equivalent to a sequence of assert-nargs-ee and reserve-locals. The * number of locals reserved is EXPECTED + NLOCALS. */ VM_DEFINE_OP (29, assert_nargs_ee_locals, "assert-nargs-ee/locals", OP1 (X8_C12_C12)) { scm_t_uint16 expected, nlocals; UNPACK_12_12 (op, expected, nlocals); VM_ASSERT (FRAME_LOCALS_COUNT () == expected, vm_error_wrong_num_args (FP_REF (0))); ALLOC_FRAME (expected + nlocals); while (nlocals--) SP_SET (nlocals, SCM_UNDEFINED); NEXT (1); } /* br-if-npos-gt nreq:24 _:8 npos:24 _:8 offset:24 * * Find the first positional argument after NREQ. If it is greater * than NPOS, jump to OFFSET. * * This instruction is only emitted for functions with multiple * clauses, and an earlier clause has keywords and no rest arguments. * See "Case-lambda" in the manual, for more on how case-lambda * chooses the clause to apply. */ VM_DEFINE_OP (30, br_if_npos_gt, "br-if-npos-gt", OP3 (X8_C24, X8_C24, X8_L24)) { scm_t_uint32 nreq, npos; UNPACK_24 (op, nreq); UNPACK_24 (ip[1], npos); /* We can only have too many positionals if there are more arguments than NPOS. */ if (FRAME_LOCALS_COUNT() > npos) { scm_t_uint32 n; for (n = nreq; n < npos; n++) if (scm_is_keyword (FP_REF (n))) break; if (n == npos && !scm_is_keyword (FP_REF (n))) { scm_t_int32 offset = ip[2]; offset >>= 8; /* Sign-extending shift. */ NEXT (offset); } } NEXT (3); } /* bind-kwargs nreq:24 flags:8 nreq-and-opt:24 _:8 ntotal:24 kw-offset:32 * * flags := allow-other-keys:1 has-rest:1 _:6 * * Find the last positional argument, and shuffle all the rest above * NTOTAL. Initialize the intervening locals to SCM_UNDEFINED. Then * load the constant at KW-OFFSET words from the current IP, and use it * to bind keyword arguments. If HAS-REST, collect all shuffled * arguments into a list, and store it in NREQ-AND-OPT. Finally, clear * the arguments that we shuffled up. * * A macro-mega-instruction. */ VM_DEFINE_OP (31, bind_kwargs, "bind-kwargs", OP4 (X8_C24, C8_C24, X8_C24, N32)) { scm_t_uint32 nreq, nreq_and_opt, ntotal, npositional, nkw, n, nargs; scm_t_int32 kw_offset; scm_t_bits kw_bits; SCM kw; char allow_other_keys, has_rest; UNPACK_24 (op, nreq); allow_other_keys = ip[1] & 0x1; has_rest = ip[1] & 0x2; UNPACK_24 (ip[1], nreq_and_opt); UNPACK_24 (ip[2], ntotal); kw_offset = ip[3]; kw_bits = (scm_t_bits) (ip + kw_offset); VM_ASSERT (!(kw_bits & 0x7), abort()); kw = SCM_PACK (kw_bits); nargs = FRAME_LOCALS_COUNT (); /* look in optionals for first keyword or last positional */ /* starting after the last required positional arg */ npositional = nreq; while (/* while we have args */ npositional < nargs /* and we still have positionals to fill */ && npositional < nreq_and_opt /* and we haven't reached a keyword yet */ && !scm_is_keyword (FP_REF (npositional))) /* bind this optional arg (by leaving it in place) */ npositional++; nkw = nargs - npositional; /* shuffle non-positional arguments above ntotal */ ALLOC_FRAME (ntotal + nkw); n = nkw; while (n--) FP_SET (ntotal + n, FP_REF (npositional + n)); /* and fill optionals & keyword args with SCM_UNDEFINED */ n = npositional; while (n < ntotal) FP_SET (n++, SCM_UNDEFINED); /* Now bind keywords, in the order given. */ for (n = 0; n < nkw; n++) if (scm_is_keyword (FP_REF (ntotal + n))) { SCM walk; for (walk = kw; scm_is_pair (walk); walk = SCM_CDR (walk)) if (scm_is_eq (SCM_CAAR (walk), FP_REF (ntotal + n))) { SCM si = SCM_CDAR (walk); if (n + 1 < nkw) { FP_SET (SCM_I_INUMP (si) ? SCM_I_INUM (si) : scm_to_uint32 (si), FP_REF (ntotal + n + 1)); } else vm_error_kwargs_missing_value (FP_REF (0), FP_REF (ntotal + n)); break; } VM_ASSERT (scm_is_pair (walk) || allow_other_keys, vm_error_kwargs_unrecognized_keyword (FP_REF (0), FP_REF (ntotal + n))); n++; } else VM_ASSERT (has_rest, vm_error_kwargs_invalid_keyword (FP_REF (0), FP_REF (ntotal + n))); if (has_rest) { SCM rest = SCM_EOL; n = nkw; SYNC_IP (); while (n--) rest = scm_inline_cons (thread, FP_REF (ntotal + n), rest); FP_SET (nreq_and_opt, rest); } RESET_FRAME (ntotal); NEXT (4); } /* bind-rest dst:24 * * Collect any arguments at or above DST into a list, and store that * list at DST. */ VM_DEFINE_OP (32, bind_rest, "bind-rest", OP1 (X8_F24) | OP_DST) { scm_t_uint32 dst, nargs; SCM rest = SCM_EOL; UNPACK_24 (op, dst); nargs = FRAME_LOCALS_COUNT (); if (nargs <= dst) { ALLOC_FRAME (dst + 1); while (nargs < dst) FP_SET (nargs++, SCM_UNDEFINED); } else { SYNC_IP (); while (nargs-- > dst) { rest = scm_inline_cons (thread, FP_REF (nargs), rest); FP_SET (nargs, SCM_UNDEFINED); } RESET_FRAME (dst + 1); } FP_SET (dst, rest); NEXT (1); } /* * Branching instructions */ /* br offset:24 * * Add OFFSET, a signed 24-bit number, to the current instruction * pointer. */ VM_DEFINE_OP (33, br, "br", OP1 (X8_L24)) { scm_t_int32 offset = op; offset >>= 8; /* Sign-extending shift. */ NEXT (offset); } /* br-if-true test:24 invert:1 _:7 offset:24 * * If the value in TEST is true for the purposes of Scheme, add * OFFSET, a signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (34, br_if_true, "br-if-true", OP2 (X8_S24, B1_X7_L24)) { BR_UNARY (x, scm_is_true (x)); } /* br-if-null test:24 invert:1 _:7 offset:24 * * If the value in TEST is the end-of-list or Lisp nil, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (35, br_if_null, "br-if-null", OP2 (X8_S24, B1_X7_L24)) { BR_UNARY (x, scm_is_null (x)); } /* br-if-nil test:24 invert:1 _:7 offset:24 * * If the value in TEST is false to Lisp, add OFFSET, a signed 24-bit * number, to the current instruction pointer. */ VM_DEFINE_OP (36, br_if_nil, "br-if-nil", OP2 (X8_S24, B1_X7_L24)) { BR_UNARY (x, scm_is_lisp_false (x)); } /* br-if-pair test:24 invert:1 _:7 offset:24 * * If the value in TEST is a pair, add OFFSET, a signed 24-bit number, * to the current instruction pointer. */ VM_DEFINE_OP (37, br_if_pair, "br-if-pair", OP2 (X8_S24, B1_X7_L24)) { BR_UNARY (x, scm_is_pair (x)); } /* br-if-struct test:24 invert:1 _:7 offset:24 * * If the value in TEST is a struct, add OFFSET, a signed 24-bit * number, to the current instruction pointer. */ VM_DEFINE_OP (38, br_if_struct, "br-if-struct", OP2 (X8_S24, B1_X7_L24)) { BR_UNARY (x, SCM_STRUCTP (x)); } /* br-if-char test:24 invert:1 _:7 offset:24 * * If the value in TEST is a char, add OFFSET, a signed 24-bit number, * to the current instruction pointer. */ VM_DEFINE_OP (39, br_if_char, "br-if-char", OP2 (X8_S24, B1_X7_L24)) { BR_UNARY (x, SCM_CHARP (x)); } /* br-if-tc7 test:24 invert:1 tc7:7 offset:24 * * If the value in TEST has the TC7 given in the second word, add * OFFSET, a signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (40, br_if_tc7, "br-if-tc7", OP2 (X8_S24, B1_C7_L24)) { BR_UNARY (x, SCM_HAS_TYP7 (x, (ip[1] >> 1) & 0x7f)); } /* br-if-eq a:12 b:12 invert:1 _:7 offset:24 * * If the value in A is eq? to the value in B, add OFFSET, a signed * 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (41, br_if_eq, "br-if-eq", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_BINARY (x, y, scm_is_eq (x, y)); } /* br-if-eqv a:12 b:12 invert:1 _:7 offset:24 * * If the value in A is eqv? to the value in B, add OFFSET, a signed * 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (42, br_if_eqv, "br-if-eqv", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_BINARY (x, y, scm_is_eq (x, y) || (SCM_NIMP (x) && SCM_NIMP (y) && scm_is_true (scm_eqv_p (x, y)))); } VM_DEFINE_OP (43, unused_43, NULL, NOP) { abort (); } /* br-if-logtest a:24 _:8 b:24 invert:1 _:7 offset:24 * * If the exact integer in A has any bits in common with the exact * integer in B, add OFFSET, a signed 24-bit number, to the current * instruction pointer. */ VM_DEFINE_OP (44, br_if_logtest, "br-if-logtest", OP3 (X8_S24, X8_S24, B1_X7_L24)) { SYNC_IP (); { BR_BINARY (x, y, ((SCM_I_INUMP (x) && SCM_I_INUMP (y)) ? (SCM_UNPACK (x) & SCM_UNPACK (y) & ~scm_tc2_int) : scm_is_true (scm_logtest (x, y)))); } } /* br-if-= a:12 b:12 invert:1 _:7 offset:24 * * If the value in A is = to the value in B, add OFFSET, a signed * 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (45, br_if_ee, "br-if-=", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_ARITHMETIC (==, scm_num_eq_p); } /* br-if-< a:12 b:12 invert:1 _:7 offset:24 * * If the value in A is < to the value in B, add OFFSET, a signed * 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (46, br_if_lt, "br-if-<", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_ARITHMETIC (<, scm_less_p); } /* br-if-<= a:12 b:12 invert:1 _:7 offset:24 * * If the value in A is <= to the value in B, add OFFSET, a signed * 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (47, br_if_le, "br-if-<=", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_ARITHMETIC (<=, scm_leq_p); } /* * Lexical binding instructions */ /* mov dst:12 src:12 * * Copy a value from one local slot to another. */ VM_DEFINE_OP (48, mov, "mov", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst; scm_t_uint16 src; UNPACK_12_12 (op, dst, src); /* FIXME: The compiler currently emits "mov" for SCM, F64, U64, and S64 variables. However SCM values are the usual case, and on a 32-bit machine it might be cheaper to move a SCM than to move a 64-bit number. */ SP_SET_SLOT (dst, SP_REF_SLOT (src)); NEXT (1); } /* long-mov dst:24 _:8 src:24 * * Copy a value from one local slot to another. */ VM_DEFINE_OP (49, long_mov, "long-mov", OP2 (X8_S24, X8_S24) | OP_DST) { scm_t_uint32 dst; scm_t_uint32 src; UNPACK_24 (op, dst); UNPACK_24 (ip[1], src); /* FIXME: The compiler currently emits "long-mov" for SCM, F64, U64, and S64 variables. However SCM values are the usual case, and on a 32-bit machine it might be cheaper to move a SCM than to move a 64-bit number. */ SP_SET_SLOT (dst, SP_REF_SLOT (src)); NEXT (2); } /* long-fmov dst:24 _:8 src:24 * * Copy a value from one local slot to another. Slot indexes are * relative to the FP. */ VM_DEFINE_OP (50, long_fmov, "long-fmov", OP2 (X8_F24, X8_F24) | OP_DST) { scm_t_uint32 dst; scm_t_uint32 src; UNPACK_24 (op, dst); UNPACK_24 (ip[1], src); FP_SET (dst, FP_REF (src)); NEXT (2); } /* box dst:12 src:12 * * Create a new variable holding SRC, and place it in DST. */ VM_DEFINE_OP (51, box, "box", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET (dst, scm_inline_cell (thread, scm_tc7_variable, SCM_UNPACK (SP_REF (src)))); NEXT (1); } /* box-ref dst:12 src:12 * * Unpack the variable at SRC into DST, asserting that the variable is * actually bound. */ VM_DEFINE_OP (52, box_ref, "box-ref", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; SCM var; UNPACK_12_12 (op, dst, src); var = SP_REF (src); VM_VALIDATE_VARIABLE (var, "variable-ref"); VM_ASSERT (VARIABLE_BOUNDP (var), vm_error_unbound (var)); SP_SET (dst, VARIABLE_REF (var)); NEXT (1); } /* box-set! dst:12 src:12 * * Set the contents of the variable at DST to SET. */ VM_DEFINE_OP (53, box_set, "box-set!", OP1 (X8_S12_S12)) { scm_t_uint16 dst, src; SCM var; UNPACK_12_12 (op, dst, src); var = SP_REF (dst); VM_VALIDATE_VARIABLE (var, "variable-set!"); VARIABLE_SET (var, SP_REF (src)); NEXT (1); } /* make-closure dst:24 offset:32 _:8 nfree:24 * * Make a new closure, and write it to DST. The code for the closure * will be found at OFFSET words from the current IP. OFFSET is a * signed 32-bit integer. Space for NFREE free variables will be * allocated. */ VM_DEFINE_OP (54, make_closure, "make-closure", OP3 (X8_S24, L32, X8_C24) | OP_DST) { scm_t_uint32 dst, nfree, n; scm_t_int32 offset; SCM closure; UNPACK_24 (op, dst); offset = ip[1]; UNPACK_24 (ip[2], nfree); // FIXME: Assert range of nfree? closure = scm_inline_words (thread, scm_tc7_program | (nfree << 16), nfree + 2); SCM_SET_CELL_WORD_1 (closure, ip + offset); // FIXME: Elide these initializations? for (n = 0; n < nfree; n++) SCM_PROGRAM_FREE_VARIABLE_SET (closure, n, SCM_BOOL_F); SP_SET (dst, closure); NEXT (3); } /* free-ref dst:12 src:12 _:8 idx:24 * * Load free variable IDX from the closure SRC into local slot DST. */ VM_DEFINE_OP (55, free_ref, "free-ref", OP2 (X8_S12_S12, X8_C24) | OP_DST) { scm_t_uint16 dst, src; scm_t_uint32 idx; UNPACK_12_12 (op, dst, src); UNPACK_24 (ip[1], idx); /* CHECK_FREE_VARIABLE (src); */ SP_SET (dst, SCM_PROGRAM_FREE_VARIABLE_REF (SP_REF (src), idx)); NEXT (2); } /* free-set! dst:12 src:12 _:8 idx:24 * * Set free variable IDX from the closure DST to SRC. */ VM_DEFINE_OP (56, free_set, "free-set!", OP2 (X8_S12_S12, X8_C24)) { scm_t_uint16 dst, src; scm_t_uint32 idx; UNPACK_12_12 (op, dst, src); UNPACK_24 (ip[1], idx); /* CHECK_FREE_VARIABLE (src); */ SCM_PROGRAM_FREE_VARIABLE_SET (SP_REF (dst), idx, SP_REF (src)); NEXT (2); } /* * Immediates and statically allocated non-immediates */ /* make-short-immediate dst:8 low-bits:16 * * Make an immediate whose low bits are LOW-BITS, and whose top bits are * 0. */ VM_DEFINE_OP (57, make_short_immediate, "make-short-immediate", OP1 (X8_S8_I16) | OP_DST) { scm_t_uint8 dst; scm_t_bits val; UNPACK_8_16 (op, dst, val); SP_SET (dst, SCM_PACK (val)); NEXT (1); } /* make-long-immediate dst:24 low-bits:32 * * Make an immediate whose low bits are LOW-BITS, and whose top bits are * 0. */ VM_DEFINE_OP (58, make_long_immediate, "make-long-immediate", OP2 (X8_S24, I32) | OP_DST) { scm_t_uint32 dst; scm_t_bits val; UNPACK_24 (op, dst); val = ip[1]; SP_SET (dst, SCM_PACK (val)); NEXT (2); } /* make-long-long-immediate dst:24 high-bits:32 low-bits:32 * * Make an immediate with HIGH-BITS and LOW-BITS. */ VM_DEFINE_OP (59, make_long_long_immediate, "make-long-long-immediate", OP3 (X8_S24, A32, B32) | OP_DST) { scm_t_uint32 dst; scm_t_bits val; UNPACK_24 (op, dst); #if SIZEOF_SCM_T_BITS > 4 val = ip[1]; val <<= 32; val |= ip[2]; #else ASSERT (ip[1] == 0); val = ip[2]; #endif SP_SET (dst, SCM_PACK (val)); NEXT (3); } /* make-non-immediate dst:24 offset:32 * * Load a pointer to statically allocated memory into DST. The * object's memory is will be found OFFSET 32-bit words away from the * current instruction pointer. OFFSET is a signed value. The * intention here is that the compiler would produce an object file * containing the words of a non-immediate object, and this * instruction creates a pointer to that memory, effectively * resurrecting that object. * * Whether the object is mutable or immutable depends on where it was * allocated by the compiler, and loaded by the loader. */ VM_DEFINE_OP (60, make_non_immediate, "make-non-immediate", OP2 (X8_S24, N32) | OP_DST) { scm_t_uint32 dst; scm_t_int32 offset; scm_t_uint32* loc; scm_t_bits unpacked; UNPACK_24 (op, dst); offset = ip[1]; loc = ip + offset; unpacked = (scm_t_bits) loc; VM_ASSERT (!(unpacked & 0x7), abort()); SP_SET (dst, SCM_PACK (unpacked)); NEXT (2); } /* static-ref dst:24 offset:32 * * Load a SCM value into DST. The SCM value will be fetched from * memory, OFFSET 32-bit words away from the current instruction * pointer. OFFSET is a signed value. * * The intention is for this instruction to be used to load constants * that the compiler is unable to statically allocate, like symbols. * These values would be initialized when the object file loads. */ VM_DEFINE_OP (61, static_ref, "static-ref", OP2 (X8_S24, R32) | OP_DST) { scm_t_uint32 dst; scm_t_int32 offset; scm_t_uint32* loc; scm_t_uintptr loc_bits; UNPACK_24 (op, dst); offset = ip[1]; loc = ip + offset; loc_bits = (scm_t_uintptr) loc; VM_ASSERT (ALIGNED_P (loc, SCM), abort()); SP_SET (dst, *((SCM *) loc_bits)); NEXT (2); } /* static-set! src:24 offset:32 * * Store a SCM value into memory, OFFSET 32-bit words away from the * current instruction pointer. OFFSET is a signed value. */ VM_DEFINE_OP (62, static_set, "static-set!", OP2 (X8_S24, LO32)) { scm_t_uint32 src; scm_t_int32 offset; scm_t_uint32* loc; UNPACK_24 (op, src); offset = ip[1]; loc = ip + offset; VM_ASSERT (ALIGNED_P (loc, SCM), abort()); *((SCM *) loc) = SP_REF (src); NEXT (2); } /* static-patch! _:24 dst-offset:32 src-offset:32 * * Patch a pointer at DST-OFFSET to point to SRC-OFFSET. Both offsets * are signed 32-bit values, indicating a memory address as a number * of 32-bit words away from the current instruction pointer. */ VM_DEFINE_OP (63, static_patch, "static-patch!", OP3 (X32, LO32, L32)) { scm_t_int32 dst_offset, src_offset; void *src; void** dst_loc; dst_offset = ip[1]; src_offset = ip[2]; dst_loc = (void **) (ip + dst_offset); src = ip + src_offset; VM_ASSERT (ALIGNED_P (dst_loc, void*), abort()); *dst_loc = src; NEXT (3); } /* * Mutable top-level bindings */ /* There are three slightly different ways to resolve toplevel variables. 1. A toplevel reference outside of a function. These need to be looked up when the expression is evaluated -- no later, and no before. They are looked up relative to the module that is current when the expression is evaluated. For example: (if (foo) a b) The "resolve" instruction resolves the variable (box), and then access is via box-ref or box-set!. 2. A toplevel reference inside a function. These are looked up relative to the module that was current when the function was defined. Unlike code at the toplevel, which is usually run only once, these bindings benefit from memoized lookup, in which the variable resulting from the lookup is cached in the function. (lambda () (if (foo) a b)) The toplevel-box instruction is equivalent to "resolve", but caches the resulting variable in statically allocated memory. 3. A reference to an identifier with respect to a particular module. This can happen for primitive references, and references residualized by macro expansions. These can always be cached. Use module-box for these. */ /* current-module dst:24 * * Store the current module in DST. */ VM_DEFINE_OP (64, current_module, "current-module", OP1 (X8_S24) | OP_DST) { scm_t_uint32 dst; UNPACK_24 (op, dst); SYNC_IP (); SP_SET (dst, scm_current_module ()); NEXT (1); } /* resolve dst:24 bound?:1 _:7 sym:24 * * Resolve SYM in the current module, and place the resulting variable * in DST. */ VM_DEFINE_OP (65, resolve, "resolve", OP2 (X8_S24, B1_X7_S24) | OP_DST) { scm_t_uint32 dst; scm_t_uint32 sym; SCM var; UNPACK_24 (op, dst); UNPACK_24 (ip[1], sym); SYNC_IP (); var = scm_lookup (SP_REF (sym)); CACHE_SP (); if (ip[1] & 0x1) VM_ASSERT (VARIABLE_BOUNDP (var), vm_error_unbound (SP_REF (sym))); SP_SET (dst, var); NEXT (2); } /* define! dst:12 sym:12 * * Look up a binding for SYM in the current module, creating it if * necessary. Set its value to VAL. */ VM_DEFINE_OP (66, define, "define!", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, sym; SCM var; UNPACK_12_12 (op, dst, sym); SYNC_IP (); var = scm_module_ensure_local_variable (scm_current_module (), SP_REF (sym)); CACHE_SP (); SP_SET (dst, var); NEXT (1); } /* toplevel-box dst:24 var-offset:32 mod-offset:32 sym-offset:32 bound?:1 _:31 * * Load a SCM value. The SCM value will be fetched from memory, * VAR-OFFSET 32-bit words away from the current instruction pointer. * VAR-OFFSET is a signed value. Up to here, toplevel-box is like * static-ref. * * Then, if the loaded value is a variable, it is placed in DST, and control * flow continues. * * Otherwise, we have to resolve the variable. In that case we load * the module from MOD-OFFSET, just as we loaded the variable. * Usually the module gets set when the closure is created. The name * is an offset to a symbol. * * We use the module and the symbol to resolve the variable, placing it in * DST, and caching the resolved variable so that we will hit the cache next * time. */ VM_DEFINE_OP (67, toplevel_box, "toplevel-box", OP5 (X8_S24, R32, R32, N32, B1_X31) | OP_DST) { scm_t_uint32 dst; scm_t_int32 var_offset; scm_t_uint32* var_loc_u32; SCM *var_loc; SCM var; UNPACK_24 (op, dst); var_offset = ip[1]; var_loc_u32 = ip + var_offset; VM_ASSERT (ALIGNED_P (var_loc_u32, SCM), abort()); var_loc = (SCM *) var_loc_u32; var = *var_loc; if (SCM_UNLIKELY (!SCM_VARIABLEP (var))) { SCM mod, sym; scm_t_int32 mod_offset = ip[2]; /* signed */ scm_t_int32 sym_offset = ip[3]; /* signed */ scm_t_uint32 *mod_loc = ip + mod_offset; scm_t_uint32 *sym_loc = ip + sym_offset; SYNC_IP (); VM_ASSERT (ALIGNED_P (mod_loc, SCM), abort()); VM_ASSERT (ALIGNED_P (sym_loc, SCM), abort()); mod = *((SCM *) mod_loc); sym = *((SCM *) sym_loc); /* If the toplevel scope was captured before modules were booted, use the root module. */ if (scm_is_false (mod)) mod = scm_the_root_module (); var = scm_module_lookup (mod, sym); CACHE_SP (); if (ip[4] & 0x1) VM_ASSERT (VARIABLE_BOUNDP (var), vm_error_unbound (sym)); *var_loc = var; } SP_SET (dst, var); NEXT (5); } /* module-box dst:24 var-offset:32 mod-offset:32 sym-offset:32 bound?:1 _:31 * * Like toplevel-box, except MOD-OFFSET points at the name of a module * instead of the module itself. */ VM_DEFINE_OP (68, module_box, "module-box", OP5 (X8_S24, R32, N32, N32, B1_X31) | OP_DST) { scm_t_uint32 dst; scm_t_int32 var_offset; scm_t_uint32* var_loc_u32; SCM *var_loc; SCM var; UNPACK_24 (op, dst); var_offset = ip[1]; var_loc_u32 = ip + var_offset; VM_ASSERT (ALIGNED_P (var_loc_u32, SCM), abort()); var_loc = (SCM *) var_loc_u32; var = *var_loc; if (SCM_UNLIKELY (!SCM_VARIABLEP (var))) { SCM modname, sym; scm_t_int32 modname_offset = ip[2]; /* signed */ scm_t_int32 sym_offset = ip[3]; /* signed */ scm_t_uint32 *modname_words = ip + modname_offset; scm_t_uint32 *sym_loc = ip + sym_offset; SYNC_IP (); VM_ASSERT (!(((scm_t_uintptr) modname_words) & 0x7), abort()); VM_ASSERT (ALIGNED_P (sym_loc, SCM), abort()); modname = SCM_PACK ((scm_t_bits) modname_words); sym = *((SCM *) sym_loc); if (!scm_module_system_booted_p) { ASSERT (scm_is_true scm_equal_p (modname, scm_list_2 (SCM_BOOL_T, scm_from_utf8_symbol ("guile")))); var = scm_lookup (sym); } else if (scm_is_true (SCM_CAR (modname))) var = scm_public_lookup (SCM_CDR (modname), sym); else var = scm_private_lookup (SCM_CDR (modname), sym); CACHE_SP (); if (ip[4] & 0x1) VM_ASSERT (VARIABLE_BOUNDP (var), vm_error_unbound (sym)); *var_loc = var; } SP_SET (dst, var); NEXT (5); } /* * The dynamic environment */ /* prompt tag:24 escape-only?:1 _:7 proc-slot:24 _:8 handler-offset:24 * * Push a new prompt on the dynamic stack, with a tag from TAG and a * handler at HANDLER-OFFSET words from the current IP. The handler * will expect a multiple-value return as if from a call with the * procedure at PROC-SLOT. */ VM_DEFINE_OP (69, prompt, "prompt", OP3 (X8_S24, B1_X7_F24, X8_L24)) { scm_t_uint32 tag, proc_slot; scm_t_int32 offset; scm_t_uint8 escape_only_p; scm_t_dynstack_prompt_flags flags; UNPACK_24 (op, tag); escape_only_p = ip[1] & 0x1; UNPACK_24 (ip[1], proc_slot); offset = ip[2]; offset >>= 8; /* Sign extension */ /* Push the prompt onto the dynamic stack. */ flags = escape_only_p ? SCM_F_DYNSTACK_PROMPT_ESCAPE_ONLY : 0; SYNC_IP (); scm_dynstack_push_prompt (&thread->dynstack, flags, SP_REF (tag), vp->stack_top - vp->fp, vp->stack_top - FP_SLOT (proc_slot), ip + offset, registers); NEXT (3); } /* wind winder:12 unwinder:12 * * Push wind and unwind procedures onto the dynamic stack. Note that * neither are actually called; the compiler should emit calls to wind * and unwind for the normal dynamic-wind control flow. Also note that * the compiler should have inserted checks that they wind and unwind * procs are thunks, if it could not prove that to be the case. */ VM_DEFINE_OP (70, wind, "wind", OP1 (X8_S12_S12)) { scm_t_uint16 winder, unwinder; UNPACK_12_12 (op, winder, unwinder); SYNC_IP (); scm_dynstack_push_dynwind (&thread->dynstack, SP_REF (winder), SP_REF (unwinder)); NEXT (1); } /* unwind _:24 * * A normal exit from the dynamic extent of an expression. Pop the top * entry off of the dynamic stack. */ VM_DEFINE_OP (71, unwind, "unwind", OP1 (X32)) { scm_dynstack_pop (&thread->dynstack); NEXT (1); } /* push-fluid fluid:12 value:12 * * Dynamically bind VALUE to FLUID. */ VM_DEFINE_OP (72, push_fluid, "push-fluid", OP1 (X8_S12_S12)) { scm_t_uint32 fluid, value; UNPACK_12_12 (op, fluid, value); SYNC_IP (); scm_dynstack_push_fluid (&thread->dynstack, SP_REF (fluid), SP_REF (value), thread->dynamic_state); NEXT (1); } /* pop-fluid _:24 * * Leave the dynamic extent of a with-fluid* expression, restoring the * fluid to its previous value. */ VM_DEFINE_OP (73, pop_fluid, "pop-fluid", OP1 (X32)) { /* This function must not allocate. */ scm_dynstack_unwind_fluid (&thread->dynstack, thread->dynamic_state); NEXT (1); } /* fluid-ref dst:12 src:12 * * Reference the fluid in SRC, and place the value in DST. */ VM_DEFINE_OP (74, fluid_ref, "fluid-ref", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; SCM fluid; struct scm_cache_entry *entry; UNPACK_12_12 (op, dst, src); fluid = SP_REF (src); /* If we find FLUID in the cache, then it is indeed a fluid. */ entry = scm_cache_lookup (&thread->dynamic_state->cache, fluid); if (SCM_LIKELY (scm_is_eq (SCM_PACK (entry->key), fluid) && !SCM_UNBNDP (SCM_PACK (entry->value)))) { SP_SET (dst, SCM_PACK (entry->value)); NEXT (1); } else { SYNC_IP (); SP_SET (dst, scm_fluid_ref (fluid)); NEXT (1); } } /* fluid-set fluid:12 val:12 * * Set the value of the fluid in DST to the value in SRC. */ VM_DEFINE_OP (75, fluid_set, "fluid-set!", OP1 (X8_S12_S12)) { scm_t_uint16 a, b; SCM fluid, value; struct scm_cache_entry *entry; UNPACK_12_12 (op, a, b); fluid = SP_REF (a); value = SP_REF (b); /* If we find FLUID in the cache, then it is indeed a fluid. */ entry = scm_cache_lookup (&thread->dynamic_state->cache, fluid); if (SCM_LIKELY (scm_is_eq (SCM_PACK (entry->key), fluid))) { entry->value = SCM_UNPACK (value); NEXT (1); } else { SYNC_IP (); scm_fluid_set_x (fluid, value); NEXT (1); } } /* * Strings, symbols, and keywords */ /* string-length dst:12 src:12 * * Store the length of the string in SRC in DST. */ VM_DEFINE_OP (76, string_length, "string-length", OP1 (X8_S12_S12) | OP_DST) { ARGS1 (str); VM_VALIDATE_STRING (str, "string-length"); SP_SET_U64 (dst, scm_i_string_length (str)); NEXT (1); } /* string-ref dst:8 src:8 idx:8 * * Fetch the character at position IDX in the string in SRC, and store * it in DST. */ VM_DEFINE_OP (77, string_ref, "string-ref", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, src, idx; SCM str; scm_t_uint64 c_idx; UNPACK_8_8_8 (op, dst, src, idx); str = SP_REF (src); c_idx = SP_REF_U64 (idx); VM_VALIDATE_STRING (str, "string-ref"); VM_VALIDATE_INDEX (c_idx, scm_i_string_length (str), "string-ref"); RETURN (SCM_MAKE_CHAR (scm_i_string_ref (str, c_idx))); } /* string-set! instruction is currently number 192. Probably need to reorder before releasing. */ /* string->number dst:12 src:12 * * Parse a string in SRC to a number, and store in DST. */ VM_DEFINE_OP (78, string_to_number, "string->number", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET (dst, scm_string_to_number (SP_REF (src), SCM_UNDEFINED /* radix = 10 */)); NEXT (1); } /* string->symbol dst:12 src:12 * * Parse a string in SRC to a symbol, and store in DST. */ VM_DEFINE_OP (79, string_to_symbol, "string->symbol", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET (dst, scm_string_to_symbol (SP_REF (src))); NEXT (1); } /* symbol->keyword dst:12 src:12 * * Make a keyword from the symbol in SRC, and store it in DST. */ VM_DEFINE_OP (80, symbol_to_keyword, "symbol->keyword", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET (dst, scm_symbol_to_keyword (SP_REF (src))); NEXT (1); } /* * Pairs */ /* cons dst:8 car:8 cdr:8 * * Cons CAR and CDR, and store the result in DST. */ VM_DEFINE_OP (81, cons, "cons", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); SYNC_IP (); RETURN (scm_inline_cons (thread, x, y)); } /* car dst:12 src:12 * * Place the car of SRC in DST. */ VM_DEFINE_OP (82, car, "car", OP1 (X8_S12_S12) | OP_DST) { ARGS1 (x); VM_VALIDATE_PAIR (x, "car"); RETURN (SCM_CAR (x)); } /* cdr dst:12 src:12 * * Place the cdr of SRC in DST. */ VM_DEFINE_OP (83, cdr, "cdr", OP1 (X8_S12_S12) | OP_DST) { ARGS1 (x); VM_VALIDATE_PAIR (x, "cdr"); RETURN (SCM_CDR (x)); } /* set-car! pair:12 car:12 * * Set the car of DST to SRC. */ VM_DEFINE_OP (84, set_car, "set-car!", OP1 (X8_S12_S12)) { scm_t_uint16 a, b; SCM x, y; UNPACK_12_12 (op, a, b); x = SP_REF (a); y = SP_REF (b); VM_VALIDATE_PAIR (x, "set-car!"); SCM_SETCAR (x, y); NEXT (1); } /* set-cdr! pair:12 cdr:12 * * Set the cdr of DST to SRC. */ VM_DEFINE_OP (85, set_cdr, "set-cdr!", OP1 (X8_S12_S12)) { scm_t_uint16 a, b; SCM x, y; UNPACK_12_12 (op, a, b); x = SP_REF (a); y = SP_REF (b); VM_VALIDATE_PAIR (x, "set-car!"); SCM_SETCDR (x, y); NEXT (1); } /* * Numeric operations */ /* add dst:8 a:8 b:8 * * Add A to B, and place the result in DST. */ VM_DEFINE_OP (86, add, "add", OP1 (X8_S8_S8_S8) | OP_DST) { BINARY_INTEGER_OP (+, scm_sum); } /* add/immediate dst:8 src:8 imm:8 * * Add the unsigned 8-bit value IMM to the value from SRC, and place * the result in DST. */ VM_DEFINE_OP (87, add_immediate, "add/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, src, imm; SCM x; UNPACK_8_8_8 (op, dst, src, imm); x = SP_REF (src); if (SCM_LIKELY (SCM_I_INUMP (x))) { scm_t_signed_bits sum = SCM_I_INUM (x) + (scm_t_signed_bits) imm; if (SCM_LIKELY (SCM_POSFIXABLE (sum))) RETURN (SCM_I_MAKINUM (sum)); } RETURN_EXP (scm_sum (x, SCM_I_MAKINUM (imm))); } /* sub dst:8 a:8 b:8 * * Subtract B from A, and place the result in DST. */ VM_DEFINE_OP (88, sub, "sub", OP1 (X8_S8_S8_S8) | OP_DST) { BINARY_INTEGER_OP (-, scm_difference); } /* sub/immediate dst:8 src:8 imm:8 * * Subtract the unsigned 8-bit value IMM from the value in SRC, and * place the result in DST. */ VM_DEFINE_OP (89, sub_immediate, "sub/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, src, imm; SCM x; UNPACK_8_8_8 (op, dst, src, imm); x = SP_REF (src); if (SCM_LIKELY (SCM_I_INUMP (x))) { scm_t_signed_bits diff = SCM_I_INUM (x) - (scm_t_signed_bits) imm; if (SCM_LIKELY (SCM_NEGFIXABLE (diff))) RETURN (SCM_I_MAKINUM (diff)); } RETURN_EXP (scm_difference (x, SCM_I_MAKINUM (imm))); } /* mul dst:8 a:8 b:8 * * Multiply A and B, and place the result in DST. */ VM_DEFINE_OP (90, mul, "mul", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); RETURN_EXP (scm_product (x, y)); } /* div dst:8 a:8 b:8 * * Divide A by B, and place the result in DST. */ VM_DEFINE_OP (91, div, "div", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); RETURN_EXP (scm_divide (x, y)); } /* quo dst:8 a:8 b:8 * * Divide A by B, and place the quotient in DST. */ VM_DEFINE_OP (92, quo, "quo", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); RETURN_EXP (scm_quotient (x, y)); } /* rem dst:8 a:8 b:8 * * Divide A by B, and place the remainder in DST. */ VM_DEFINE_OP (93, rem, "rem", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); RETURN_EXP (scm_remainder (x, y)); } /* mod dst:8 a:8 b:8 * * Place the modulo of A by B in DST. */ VM_DEFINE_OP (94, mod, "mod", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); RETURN_EXP (scm_modulo (x, y)); } /* ash dst:8 a:8 b:8 * * Shift A arithmetically by B bits, and place the result in DST. */ VM_DEFINE_OP (95, ash, "ash", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); if (SCM_I_INUMP (x) && SCM_I_INUMP (y)) { if (SCM_I_INUM (y) < 0) /* Right shift, will be a fixnum. */ RETURN (SCM_I_MAKINUM (SCM_SRS (SCM_I_INUM (x), (-SCM_I_INUM (y) <= SCM_I_FIXNUM_BIT-1) ? -SCM_I_INUM (y) : SCM_I_FIXNUM_BIT-1))); else /* Left shift. See comments in scm_ash. */ { scm_t_signed_bits nn, bits_to_shift; nn = SCM_I_INUM (x); bits_to_shift = SCM_I_INUM (y); if (bits_to_shift < SCM_I_FIXNUM_BIT-1 && ((scm_t_bits) (SCM_SRS (nn, (SCM_I_FIXNUM_BIT-1 - bits_to_shift)) + 1) <= 1)) RETURN (SCM_I_MAKINUM (nn < 0 ? -(-nn << bits_to_shift) : (nn << bits_to_shift))); /* fall through */ } /* fall through */ } RETURN_EXP (scm_ash (x, y)); } /* logand dst:8 a:8 b:8 * * Place the bitwise AND of A and B into DST. */ VM_DEFINE_OP (96, logand, "logand", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); if (SCM_I_INUMP (x) && SCM_I_INUMP (y)) /* Compute bitwise AND without untagging */ RETURN (SCM_PACK (SCM_UNPACK (x) & SCM_UNPACK (y))); RETURN_EXP (scm_logand (x, y)); } /* logior dst:8 a:8 b:8 * * Place the bitwise inclusive OR of A with B in DST. */ VM_DEFINE_OP (97, logior, "logior", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); if (SCM_I_INUMP (x) && SCM_I_INUMP (y)) /* Compute bitwise OR without untagging */ RETURN (SCM_PACK (SCM_UNPACK (x) | SCM_UNPACK (y))); RETURN_EXP (scm_logior (x, y)); } /* logxor dst:8 a:8 b:8 * * Place the bitwise exclusive OR of A with B in DST. */ VM_DEFINE_OP (98, logxor, "logxor", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); if (SCM_I_INUMP (x) && SCM_I_INUMP (y)) RETURN (SCM_I_MAKINUM (SCM_I_INUM (x) ^ SCM_I_INUM (y))); RETURN_EXP (scm_logxor (x, y)); } /* make-vector dst:8 length:8 init:8 * * Make a vector and write it to DST. The vector will have space for * LENGTH slots. They will be filled with the value in slot INIT. */ VM_DEFINE_OP (99, make_vector, "make-vector", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, length, init; scm_t_uint64 length_val; UNPACK_8_8_8 (op, dst, length, init); length_val = SP_REF_U64 (length); VM_VALIDATE_INDEX (length_val, (size_t) -1, "make-vector"); /* TODO: Inline this allocation. */ SYNC_IP (); SP_SET (dst, scm_c_make_vector (length_val, SP_REF (init))); NEXT (1); } /* make-vector/immediate dst:8 length:8 init:8 * * Make a short vector of known size and write it to DST. The vector * will have space for LENGTH slots, an immediate value. They will be * filled with the value in slot INIT. */ VM_DEFINE_OP (100, make_vector_immediate, "make-vector/immediate", OP1 (X8_S8_C8_S8) | OP_DST) { scm_t_uint8 dst, init; scm_t_int32 length, n; SCM val, vector; UNPACK_8_8_8 (op, dst, length, init); val = SP_REF (init); SYNC_IP (); vector = scm_inline_words (thread, scm_tc7_vector | (length << 8), length + 1); for (n = 0; n < length; n++) SCM_SIMPLE_VECTOR_SET (vector, n, val); SP_SET (dst, vector); NEXT (1); } /* vector-length dst:12 src:12 * * Store the length of the vector in SRC in DST. */ VM_DEFINE_OP (101, vector_length, "vector-length", OP1 (X8_S12_S12) | OP_DST) { ARGS1 (vect); VM_VALIDATE_VECTOR (vect, "vector-length"); SP_SET_U64 (dst, SCM_I_VECTOR_LENGTH (vect)); NEXT (1); } /* vector-ref dst:8 src:8 idx:8 * * Fetch the item at position IDX in the vector in SRC, and store it * in DST. */ VM_DEFINE_OP (102, vector_ref, "vector-ref", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, src, idx; SCM vect; scm_t_uint64 c_idx; UNPACK_8_8_8 (op, dst, src, idx); vect = SP_REF (src); c_idx = SP_REF_U64 (idx); VM_VALIDATE_VECTOR (vect, "vector-ref"); VM_VALIDATE_INDEX (c_idx, SCM_I_VECTOR_LENGTH (vect), "vector-ref"); RETURN (SCM_I_VECTOR_ELTS (vect)[c_idx]); } /* vector-ref/immediate dst:8 src:8 idx:8 * * Fill DST with the item IDX elements into the vector at SRC. Useful * for building data types using vectors. */ VM_DEFINE_OP (103, vector_ref_immediate, "vector-ref/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, src, idx; SCM vect; UNPACK_8_8_8 (op, dst, src, idx); vect = SP_REF (src); VM_VALIDATE_VECTOR (vect, "vector-ref"); VM_VALIDATE_INDEX (idx, SCM_I_VECTOR_LENGTH (vect), "vector-ref"); SP_SET (dst, SCM_I_VECTOR_ELTS (vect)[idx]); NEXT (1); } /* vector-set! dst:8 idx:8 src:8 * * Store SRC into the vector DST at index IDX. */ VM_DEFINE_OP (104, vector_set, "vector-set!", OP1 (X8_S8_S8_S8)) { scm_t_uint8 dst, idx, src; SCM vect, val; scm_t_uint64 c_idx; UNPACK_8_8_8 (op, dst, idx, src); vect = SP_REF (dst); c_idx = SP_REF_U64 (idx); val = SP_REF (src); VM_VALIDATE_VECTOR (vect, "vector-set!"); VM_VALIDATE_INDEX (c_idx, SCM_I_VECTOR_LENGTH (vect), "vector-set!"); SCM_I_VECTOR_WELTS (vect)[c_idx] = val; NEXT (1); } /* vector-set!/immediate dst:8 idx:8 src:8 * * Store SRC into the vector DST at index IDX. Here IDX is an * immediate value. */ VM_DEFINE_OP (105, vector_set_immediate, "vector-set!/immediate", OP1 (X8_S8_C8_S8)) { scm_t_uint8 dst, idx, src; SCM vect, val; UNPACK_8_8_8 (op, dst, idx, src); vect = SP_REF (dst); val = SP_REF (src); VM_VALIDATE_VECTOR (vect, "vector-set!"); VM_VALIDATE_INDEX (idx, SCM_I_VECTOR_LENGTH (vect), "vector-set!"); SCM_I_VECTOR_WELTS (vect)[idx] = val; NEXT (1); } /* * Structs and GOOPS */ /* struct-vtable dst:12 src:12 * * Store the vtable of SRC into DST. */ VM_DEFINE_OP (106, struct_vtable, "struct-vtable", OP1 (X8_S12_S12) | OP_DST) { ARGS1 (obj); VM_VALIDATE_STRUCT (obj, "struct_vtable"); RETURN (SCM_STRUCT_VTABLE (obj)); } /* allocate-struct dst:8 vtable:8 nfields:8 * * Allocate a new struct with VTABLE, and place it in DST. The struct * will be constructed with space for NFIELDS fields, which should * correspond to the field count of the VTABLE. */ VM_DEFINE_OP (107, allocate_struct, "allocate-struct", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, vtable, nfields; SCM ret; UNPACK_8_8_8 (op, dst, vtable, nfields); /* TODO: Specify nfields as untagged value when calling allocate-struct. */ SYNC_IP (); ret = scm_allocate_struct (SP_REF (vtable), scm_from_uint64 (SP_REF_U64 (nfields))); SP_SET (dst, ret); NEXT (1); } /* struct-ref dst:8 src:8 idx:8 * * Fetch the item at slot IDX in the struct in SRC, and store it * in DST. */ VM_DEFINE_OP (108, struct_ref, "struct-ref", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, src, idx; SCM obj; scm_t_uint64 index; UNPACK_8_8_8 (op, dst, src, idx); obj = SP_REF (src); index = SP_REF_U64 (idx); if (SCM_LIKELY (SCM_STRUCTP (obj) && SCM_STRUCT_VTABLE_FLAG_IS_SET (obj, SCM_VTABLE_FLAG_SIMPLE) && index < (SCM_STRUCT_DATA_REF (SCM_STRUCT_VTABLE (obj), scm_vtable_index_size)))) RETURN (SCM_STRUCT_SLOT_REF (obj, index)); SYNC_IP (); RETURN (scm_struct_ref (obj, scm_from_uint64 (index))); } /* struct-set! dst:8 idx:8 src:8 * * Store SRC into the struct DST at slot IDX. */ VM_DEFINE_OP (109, struct_set, "struct-set!", OP1 (X8_S8_S8_S8)) { scm_t_uint8 dst, idx, src; SCM obj, val; scm_t_uint64 index; UNPACK_8_8_8 (op, dst, idx, src); obj = SP_REF (dst); val = SP_REF (src); index = SP_REF_U64 (idx); if (SCM_LIKELY (SCM_STRUCTP (obj) && SCM_STRUCT_VTABLE_FLAG_IS_SET (obj, SCM_VTABLE_FLAG_SIMPLE) && SCM_STRUCT_VTABLE_FLAG_IS_SET (obj, SCM_VTABLE_FLAG_SIMPLE_RW) && index < (SCM_STRUCT_DATA_REF (SCM_STRUCT_VTABLE (obj), scm_vtable_index_size)))) { SCM_STRUCT_SLOT_SET (obj, index, val); NEXT (1); } SYNC_IP (); scm_struct_set_x (obj, scm_from_uint64 (index), val); NEXT (1); } /* allocate-struct/immediate dst:8 vtable:8 nfields:8 * * Allocate a new struct with VTABLE, and place it in DST. The struct * will be constructed with space for NFIELDS fields, which should * correspond to the field count of the VTABLE. */ VM_DEFINE_OP (110, allocate_struct_immediate, "allocate-struct/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, vtable, nfields; SCM ret; UNPACK_8_8_8 (op, dst, vtable, nfields); SYNC_IP (); ret = scm_allocate_struct (SP_REF (vtable), SCM_I_MAKINUM (nfields)); SP_SET (dst, ret); NEXT (1); } /* struct-ref/immediate dst:8 src:8 idx:8 * * Fetch the item at slot IDX in the struct in SRC, and store it * in DST. IDX is an immediate unsigned 8-bit value. */ VM_DEFINE_OP (111, struct_ref_immediate, "struct-ref/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, src, idx; SCM obj; UNPACK_8_8_8 (op, dst, src, idx); obj = SP_REF (src); if (SCM_LIKELY (SCM_STRUCTP (obj) && SCM_STRUCT_VTABLE_FLAG_IS_SET (obj, SCM_VTABLE_FLAG_SIMPLE) && idx < SCM_STRUCT_DATA_REF (SCM_STRUCT_VTABLE (obj), scm_vtable_index_size))) RETURN (SCM_STRUCT_SLOT_REF (obj, idx)); SYNC_IP (); RETURN (scm_struct_ref (obj, SCM_I_MAKINUM (idx))); } /* struct-set!/immediate dst:8 idx:8 src:8 * * Store SRC into the struct DST at slot IDX. IDX is an immediate * unsigned 8-bit value. */ VM_DEFINE_OP (112, struct_set_immediate, "struct-set!/immediate", OP1 (X8_S8_C8_S8)) { scm_t_uint8 dst, idx, src; SCM obj, val; UNPACK_8_8_8 (op, dst, idx, src); obj = SP_REF (dst); val = SP_REF (src); if (SCM_LIKELY (SCM_STRUCTP (obj) && SCM_STRUCT_VTABLE_FLAG_IS_SET (obj, SCM_VTABLE_FLAG_SIMPLE) && SCM_STRUCT_VTABLE_FLAG_IS_SET (obj, SCM_VTABLE_FLAG_SIMPLE_RW) && idx < SCM_STRUCT_DATA_REF (SCM_STRUCT_VTABLE (obj), scm_vtable_index_size))) { SCM_STRUCT_SLOT_SET (obj, idx, val); NEXT (1); } SYNC_IP (); scm_struct_set_x (obj, SCM_I_MAKINUM (idx), val); NEXT (1); } /* class-of dst:12 type:12 * * Store the vtable of SRC into DST. */ VM_DEFINE_OP (113, class_of, "class-of", OP1 (X8_S12_S12) | OP_DST) { ARGS1 (obj); if (SCM_INSTANCEP (obj)) RETURN (SCM_CLASS_OF (obj)); RETURN_EXP (scm_class_of (obj)); } /* * Arrays, packed uniform arrays, and bytevectors. */ /* load-typed-array dst:24 _:8 type:24 _:8 shape:24 offset:32 len:32 * * Load the contiguous typed array located at OFFSET 32-bit words away * from the instruction pointer, and store into DST. LEN is a byte * length. OFFSET is signed. */ VM_DEFINE_OP (114, load_typed_array, "load-typed-array", OP5 (X8_S24, X8_S24, X8_S24, N32, C32) | OP_DST) { scm_t_uint32 dst, type, shape; scm_t_int32 offset; scm_t_uint32 len; UNPACK_24 (op, dst); UNPACK_24 (ip[1], type); UNPACK_24 (ip[2], shape); offset = ip[3]; len = ip[4]; SYNC_IP (); SP_SET (dst, scm_from_contiguous_typed_array (SP_REF (type), SP_REF (shape), ip + offset, len)); NEXT (5); } /* make-array dst:24 _:8 type:24 _:8 fill:24 _:8 bounds:24 * * Make a new array with TYPE, FILL, and BOUNDS, storing it in DST. */ VM_DEFINE_OP (115, make_array, "make-array", OP4 (X8_S24, X8_S24, X8_S24, X8_S24) | OP_DST) { scm_t_uint32 dst, type, fill, bounds; UNPACK_24 (op, dst); UNPACK_24 (ip[1], type); UNPACK_24 (ip[2], fill); UNPACK_24 (ip[3], bounds); SYNC_IP (); SP_SET (dst, scm_make_typed_array (SP_REF (type), SP_REF (fill), SP_REF (bounds))); NEXT (4); } /* bv-u8-ref dst:8 src:8 idx:8 * bv-s8-ref dst:8 src:8 idx:8 * bv-u16-ref dst:8 src:8 idx:8 * bv-s16-ref dst:8 src:8 idx:8 * bv-u32-ref dst:8 src:8 idx:8 * bv-s32-ref dst:8 src:8 idx:8 * bv-u64-ref dst:8 src:8 idx:8 * bv-s64-ref dst:8 src:8 idx:8 * bv-f32-ref dst:8 src:8 idx:8 * bv-f64-ref dst:8 src:8 idx:8 * * Fetch the item at byte offset IDX in the bytevector SRC, and store * it in DST. All accesses use native endianness. */ #define BV_REF(stem, type, size, slot) \ do { \ type result; \ scm_t_uint8 dst, src, idx; \ SCM bv; \ scm_t_uint64 c_idx; \ UNPACK_8_8_8 (op, dst, src, idx); \ bv = SP_REF (src); \ c_idx = SP_REF_U64 (idx); \ \ VM_VALIDATE_BYTEVECTOR (bv, "bv-" #stem "-ref"); \ \ VM_ASSERT (SCM_BYTEVECTOR_LENGTH (bv) >= size \ && SCM_BYTEVECTOR_LENGTH (bv) - size >= c_idx, \ vm_error_out_of_range_uint64 ("bv-" #stem "-ref", c_idx)); \ \ memcpy (&result, SCM_BYTEVECTOR_CONTENTS (bv) + c_idx, size); \ SP_SET_ ## slot (dst, result); \ NEXT (1); \ } while (0) VM_DEFINE_OP (116, bv_u8_ref, "bv-u8-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (u8, scm_t_uint8, 1, U64); VM_DEFINE_OP (117, bv_s8_ref, "bv-s8-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (s8, scm_t_int8, 1, S64); VM_DEFINE_OP (118, bv_u16_ref, "bv-u16-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (u16, scm_t_uint16, 2, U64); VM_DEFINE_OP (119, bv_s16_ref, "bv-s16-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (s16, scm_t_int16, 2, S64); VM_DEFINE_OP (120, bv_u32_ref, "bv-u32-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (u32, scm_t_uint32, 4, U64); VM_DEFINE_OP (121, bv_s32_ref, "bv-s32-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (s32, scm_t_int32, 4, S64); VM_DEFINE_OP (122, bv_u64_ref, "bv-u64-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (u64, scm_t_uint64, 8, U64); VM_DEFINE_OP (123, bv_s64_ref, "bv-s64-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (s64, scm_t_int64, 8, S64); VM_DEFINE_OP (124, bv_f32_ref, "bv-f32-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (f32, float, 4, F64); VM_DEFINE_OP (125, bv_f64_ref, "bv-f64-ref", OP1 (X8_S8_S8_S8) | OP_DST) BV_REF (f64, double, 8, F64); /* bv-u8-set! dst:8 idx:8 src:8 * bv-s8-set! dst:8 idx:8 src:8 * bv-u16-set! dst:8 idx:8 src:8 * bv-s16-set! dst:8 idx:8 src:8 * bv-u32-set! dst:8 idx:8 src:8 * bv-s32-set! dst:8 idx:8 src:8 * bv-u64-set! dst:8 idx:8 src:8 * bv-s64-set! dst:8 idx:8 src:8 * bv-f32-set! dst:8 idx:8 src:8 * bv-f64-set! dst:8 idx:8 src:8 * * Store SRC into the bytevector DST at byte offset IDX. Multibyte * values are written using native endianness. */ #define BV_BOUNDED_SET(stem, type, min, max, size, slot_type, slot) \ do { \ scm_t_ ## slot_type slot_val; \ type val; \ scm_t_uint8 dst, idx, src; \ SCM bv; \ scm_t_uint64 c_idx; \ UNPACK_8_8_8 (op, dst, idx, src); \ bv = SP_REF (dst); \ c_idx = SP_REF_U64 (idx); \ slot_val = SP_REF_ ## slot (src); \ \ VM_VALIDATE_BYTEVECTOR (bv, "bv-" #stem "-set!"); \ \ VM_ASSERT (SCM_BYTEVECTOR_LENGTH (bv) >= size \ && SCM_BYTEVECTOR_LENGTH (bv) - size >= c_idx, \ vm_error_out_of_range_uint64 ("bv-" #stem "-set!", c_idx)); \ \ VM_ASSERT (slot_val >= min && slot_val <= max, \ vm_error_out_of_range_ ## slot_type ("bv-" #stem "-set!", \ slot_val)); \ \ val = slot_val; \ memcpy (SCM_BYTEVECTOR_CONTENTS (bv) + c_idx, &val, size); \ NEXT (1); \ } while (0) #define BV_SET(stem, type, size, slot) \ do { \ type val; \ scm_t_uint8 dst, idx, src; \ SCM bv; \ scm_t_uint64 c_idx; \ UNPACK_8_8_8 (op, dst, idx, src); \ bv = SP_REF (dst); \ c_idx = SP_REF_U64 (idx); \ val = SP_REF_ ## slot (src); \ \ VM_VALIDATE_BYTEVECTOR (bv, "bv-" #stem "-set!"); \ \ VM_ASSERT (SCM_BYTEVECTOR_LENGTH (bv) >= size \ && SCM_BYTEVECTOR_LENGTH (bv) - size >= c_idx, \ vm_error_out_of_range_uint64 ("bv-" #stem "-set!", c_idx)); \ \ memcpy (SCM_BYTEVECTOR_CONTENTS (bv) + c_idx, &val, size); \ NEXT (1); \ } while (0) VM_DEFINE_OP (126, bv_u8_set, "bv-u8-set!", OP1 (X8_S8_S8_S8)) BV_BOUNDED_SET (u8, scm_t_uint8, 0, SCM_T_UINT8_MAX, 1, uint64, U64); VM_DEFINE_OP (127, bv_s8_set, "bv-s8-set!", OP1 (X8_S8_S8_S8)) BV_BOUNDED_SET (s8, scm_t_int8, SCM_T_INT8_MIN, SCM_T_INT8_MAX, 1, int64, S64); VM_DEFINE_OP (128, bv_u16_set, "bv-u16-set!", OP1 (X8_S8_S8_S8)) BV_BOUNDED_SET (u16, scm_t_uint16, 0, SCM_T_UINT16_MAX, 2, uint64, U64); VM_DEFINE_OP (129, bv_s16_set, "bv-s16-set!", OP1 (X8_S8_S8_S8)) BV_BOUNDED_SET (s16, scm_t_int16, SCM_T_INT16_MIN, SCM_T_INT16_MAX, 2, int64, S64); VM_DEFINE_OP (130, bv_u32_set, "bv-u32-set!", OP1 (X8_S8_S8_S8)) BV_BOUNDED_SET (u32, scm_t_uint32, 0, SCM_T_UINT32_MAX, 4, uint64, U64); VM_DEFINE_OP (131, bv_s32_set, "bv-s32-set!", OP1 (X8_S8_S8_S8)) BV_BOUNDED_SET (s32, scm_t_int32, SCM_T_INT32_MIN, SCM_T_INT32_MAX, 4, int64, S64); VM_DEFINE_OP (132, bv_u64_set, "bv-u64-set!", OP1 (X8_S8_S8_S8)) BV_SET (u64, scm_t_uint64, 8, U64); VM_DEFINE_OP (133, bv_s64_set, "bv-s64-set!", OP1 (X8_S8_S8_S8)) BV_SET (s64, scm_t_int64, 8, S64); VM_DEFINE_OP (134, bv_f32_set, "bv-f32-set!", OP1 (X8_S8_S8_S8)) BV_SET (f32, float, 4, F64); VM_DEFINE_OP (135, bv_f64_set, "bv-f64-set!", OP1 (X8_S8_S8_S8)) BV_SET (f6, double, 8, F64); /* scm->f64 dst:12 src:12 * * Unpack a raw double-precision floating-point value from SRC and * place it in DST. Note that SRC can be any value on which * scm_to_double can operate. */ VM_DEFINE_OP (136, scm_to_f64, "scm->f64", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET_F64 (dst, scm_to_double (SP_REF (src))); NEXT (1); } /* f64->scm dst:12 src:12 * * Pack a raw double-precision floating point value into an inexact * number allocated on the heap. */ VM_DEFINE_OP (137, f64_to_scm, "f64->scm", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET (dst, scm_from_double (SP_REF_F64 (src))); NEXT (1); } /* fadd dst:8 a:8 b:8 * * Add A to B, and place the result in DST. The operands and the * result are unboxed double-precision floating-point numbers. */ VM_DEFINE_OP (138, fadd, "fadd", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_F64 (dst, SP_REF_F64 (a) + SP_REF_F64 (b)); NEXT (1); } /* fsub dst:8 a:8 b:8 * * Subtract B from A, and place the result in DST. The operands and * the result are unboxed double-precision floating-point numbers. */ VM_DEFINE_OP (139, fsub, "fsub", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_F64 (dst, SP_REF_F64 (a) - SP_REF_F64 (b)); NEXT (1); } /* fmul dst:8 a:8 b:8 * * Multiply A and B, and place the result in DST. The operands and * the result are unboxed double-precision floating-point numbers. */ VM_DEFINE_OP (140, fmul, "fmul", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_F64 (dst, SP_REF_F64 (a) * SP_REF_F64 (b)); NEXT (1); } /* fdiv dst:8 a:8 b:8 * * Divide A by B, and place the result in DST. The operands and the * result are unboxed double-precision floating-point numbers. */ VM_DEFINE_OP (141, fdiv, "fdiv", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_F64 (dst, SP_REF_F64 (a) / SP_REF_F64 (b)); NEXT (1); } /* apply-non-program _:24 * * Used by the VM as a trampoline to apply non-programs. */ VM_DEFINE_OP (142, apply_non_program, "apply-non-program", OP1 (X32)) { SCM proc = FP_REF (0); while (!SCM_PROGRAM_P (proc)) { if (SCM_STRUCTP (proc) && SCM_STRUCT_APPLICABLE_P (proc)) { proc = SCM_STRUCT_PROCEDURE (proc); FP_SET (0, proc); continue; } if (SCM_HAS_TYP7 (proc, scm_tc7_smob) && SCM_SMOB_APPLICABLE_P (proc)) { scm_t_uint32 n = FRAME_LOCALS_COUNT(); /* Shuffle args up. (FIXME: no real need to shuffle; just set IP and go. ) */ ALLOC_FRAME (n + 1); while (n--) FP_SET (n + 1, FP_REF (n)); proc = SCM_SMOB_DESCRIPTOR (proc).apply_trampoline; FP_SET (0, proc); continue; } SYNC_IP(); vm_error_wrong_type_apply (proc); } ip = SCM_PROGRAM_CODE (proc); NEXT (0); } /* scm->u64 dst:12 src:12 * * Unpack an unsigned 64-bit integer from SRC and place it in DST. */ VM_DEFINE_OP (143, scm_to_u64, "scm->u64", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET_U64 (dst, scm_to_uint64 (SP_REF (src))); NEXT (1); } /* u64->scm dst:12 src:12 * * Pack an unsigned 64-bit integer into a SCM value. */ VM_DEFINE_OP (144, u64_to_scm, "u64->scm", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET (dst, scm_from_uint64 (SP_REF_U64 (src))); NEXT (1); } /* bv-length dst:12 src:12 * * Store the length of the bytevector in SRC in DST, as an untagged * 64-bit integer. */ VM_DEFINE_OP (145, bv_length, "bv-length", OP1 (X8_S12_S12) | OP_DST) { ARGS1 (bv); VM_VALIDATE_BYTEVECTOR (bv, "bytevector-length"); SP_SET_U64 (dst, SCM_BYTEVECTOR_LENGTH (bv)); NEXT (1); } /* br-if-= a:12 b:12 invert:1 _:7 offset:24 * * If the value in A is = to the value in B, add OFFSET, a signed * 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (146, br_if_u64_ee, "br-if-u64-=", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_U64_ARITHMETIC (==); } /* br-if-< a:12 b:12 invert:1 _:7 offset:24 * * If the value in A is < to the value in B, add OFFSET, a signed * 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (147, br_if_u64_lt, "br-if-u64-<", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_U64_ARITHMETIC (<); } VM_DEFINE_OP (148, br_if_u64_le, "br-if-u64-<=", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_U64_ARITHMETIC (<=); } /* uadd dst:8 a:8 b:8 * * Add A to B, and place the result in DST. The operands and the * result are unboxed unsigned 64-bit integers. Overflow will wrap * around. */ VM_DEFINE_OP (149, uadd, "uadd", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) + SP_REF_U64 (b)); NEXT (1); } /* usub dst:8 a:8 b:8 * * Subtract B from A, and place the result in DST. The operands and * the result are unboxed unsigned 64-bit integers. Overflow will * wrap around. */ VM_DEFINE_OP (150, usub, "usub", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) - SP_REF_U64 (b)); NEXT (1); } /* umul dst:8 a:8 b:8 * * Multiply A and B, and place the result in DST. The operands and * the result are unboxed unsigned 64-bit integers. Overflow will * wrap around. */ VM_DEFINE_OP (151, umul, "umul", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) * SP_REF_U64 (b)); NEXT (1); } /* uadd/immediate dst:8 src:8 imm:8 * * Add the unsigned 64-bit value from SRC with the unsigned 8-bit * value IMM and place the raw unsigned 64-bit result in DST. * Overflow will wrap around. */ VM_DEFINE_OP (152, uadd_immediate, "uadd/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, src, imm; scm_t_uint64 x; UNPACK_8_8_8 (op, dst, src, imm); x = SP_REF_U64 (src); SP_SET_U64 (dst, x + (scm_t_uint64) imm); NEXT (1); } /* usub/immediate dst:8 src:8 imm:8 * * Subtract the unsigned 8-bit value IMM from the unsigned 64-bit * value in SRC and place the raw unsigned 64-bit result in DST. * Overflow will wrap around. */ VM_DEFINE_OP (153, usub_immediate, "usub/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, src, imm; scm_t_uint64 x; UNPACK_8_8_8 (op, dst, src, imm); x = SP_REF_U64 (src); SP_SET_U64 (dst, x - (scm_t_uint64) imm); NEXT (1); } /* umul/immediate dst:8 src:8 imm:8 * * Multiply the unsigned 64-bit value from SRC by the unsigned 8-bit * value IMM and place the raw unsigned 64-bit result in DST. * Overflow will wrap around. */ VM_DEFINE_OP (154, umul_immediate, "umul/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, src, imm; scm_t_uint64 x; UNPACK_8_8_8 (op, dst, src, imm); x = SP_REF_U64 (src); SP_SET_U64 (dst, x * (scm_t_uint64) imm); NEXT (1); } /* load-f64 dst:24 high-bits:32 low-bits:32 * * Make a double-precision floating-point value with HIGH-BITS and * LOW-BITS. */ VM_DEFINE_OP (155, load_f64, "load-f64", OP3 (X8_S24, AF32, BF32) | OP_DST) { scm_t_uint32 dst; scm_t_uint64 val; UNPACK_24 (op, dst); val = ip[1]; val <<= 32; val |= ip[2]; SP_SET_U64 (dst, val); NEXT (3); } /* load-u64 dst:24 high-bits:32 low-bits:32 * * Make an unsigned 64-bit integer with HIGH-BITS and LOW-BITS. */ VM_DEFINE_OP (156, load_u64, "load-u64", OP3 (X8_S24, AU32, BU32) | OP_DST) { scm_t_uint32 dst; scm_t_uint64 val; UNPACK_24 (op, dst); val = ip[1]; val <<= 32; val |= ip[2]; SP_SET_U64 (dst, val); NEXT (3); } /* scm->s64 dst:12 src:12 * * Unpack a signed 64-bit integer from SRC and place it in DST. */ VM_DEFINE_OP (157, scm_to_s64, "scm->s64", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET_S64 (dst, scm_to_int64 (SP_REF (src))); NEXT (1); } /* s64->scm dst:12 src:12 * * Pack an signed 64-bit integer into a SCM value. */ VM_DEFINE_OP (158, s64_to_scm, "s64->scm", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); SP_SET (dst, scm_from_int64 (SP_REF_S64 (src))); NEXT (1); } /* load-s64 dst:24 high-bits:32 low-bits:32 * * Make an unsigned 64-bit integer with HIGH-BITS and LOW-BITS. */ VM_DEFINE_OP (159, load_s64, "load-s64", OP3 (X8_S24, AS32, BS32) | OP_DST) { scm_t_uint32 dst; scm_t_uint64 val; UNPACK_24 (op, dst); val = ip[1]; val <<= 32; val |= ip[2]; SP_SET_U64 (dst, val); NEXT (3); } /* current-thread dst:24 * * Write the current thread into DST. */ VM_DEFINE_OP (160, current_thread, "current-thread", OP1 (X8_S24) | OP_DST) { scm_t_uint32 dst; UNPACK_24 (op, dst); SP_SET (dst, thread->handle); NEXT (1); } /* logsub dst:8 a:8 b:8 * * Place the bitwise AND of A and the bitwise NOT of B into DST. */ VM_DEFINE_OP (161, logsub, "logsub", OP1 (X8_S8_S8_S8) | OP_DST) { ARGS2 (x, y); if (SCM_I_INUMP (x) && SCM_I_INUMP (y)) { scm_t_signed_bits a, b; a = SCM_I_INUM (x); b = SCM_I_INUM (y); RETURN (SCM_I_MAKINUM (a & ~b)); } RETURN_EXP (scm_logand (x, scm_lognot (y))); } /* ulogand dst:8 a:8 b:8 * * Place the bitwise AND of the u64 values in A and B into DST. */ VM_DEFINE_OP (162, ulogand, "ulogand", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) & SP_REF_U64 (b)); NEXT (1); } /* ulogior dst:8 a:8 b:8 * * Place the bitwise inclusive OR of the u64 values in A and B into * DST. */ VM_DEFINE_OP (163, ulogior, "ulogior", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) | SP_REF_U64 (b)); NEXT (1); } /* ulogsub dst:8 a:8 b:8 * * Place the (A & ~B) of the u64 values A and B into DST. */ VM_DEFINE_OP (164, ulogsub, "ulogsub", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) & ~SP_REF_U64 (b)); NEXT (1); } /* ursh dst:8 a:8 b:8 * * Shift the u64 value in A right by B bits, and place the result in * DST. Only the lower 6 bits of B are used. */ VM_DEFINE_OP (165, ursh, "ursh", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) >> (SP_REF_U64 (b) & 63)); NEXT (1); } /* ulsh dst:8 a:8 b:8 * * Shift the u64 value in A left by B bits, and place the result in * DST. Only the lower 6 bits of B are used. */ VM_DEFINE_OP (166, ulsh, "ulsh", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) << (SP_REF_U64 (b) & 63)); NEXT (1); } /* scm->u64/truncate dst:12 src:12 * * Unpack an exact integer from SRC and place it in the unsigned * 64-bit register DST, truncating any high bits. If the number in * SRC is negative, all the high bits will be set. */ VM_DEFINE_OP (167, scm_to_u64_truncate, "scm->u64/truncate", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; SCM x; UNPACK_12_12 (op, dst, src); x = SP_REF (src); if (SCM_I_INUMP (x)) SP_SET_U64 (dst, (scm_t_uint64) SCM_I_INUM (x)); else { SYNC_IP (); SP_SET_U64 (dst, scm_to_uint64 (scm_logand (x, scm_from_uint64 ((scm_t_uint64) -1)))); } NEXT (1); } /* ursh/immediate dst:8 a:8 b:8 * * Shift the u64 value in A right by the immediate B bits, and place * the result in DST. Only the lower 6 bits of B are used. */ VM_DEFINE_OP (168, ursh_immediate, "ursh/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) >> (b & 63)); NEXT (1); } /* ulsh/immediate dst:8 a:8 b:8 * * Shift the u64 value in A left by the immediate B bits, and place * the result in DST. Only the lower 6 bits of B are used. */ VM_DEFINE_OP (169, ulsh_immediate, "ulsh/immediate", OP1 (X8_S8_S8_C8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) << (b & 63)); NEXT (1); } #define BR_U64_SCM_COMPARISON(x, y, unboxed, boxed) \ do { \ scm_t_uint32 a, b; \ scm_t_uint64 x; \ SCM y_scm; \ \ UNPACK_24 (op, a); \ UNPACK_24 (ip[1], b); \ x = SP_REF_U64 (a); \ y_scm = SP_REF (b); \ \ if (SCM_I_INUMP (y_scm)) \ { \ scm_t_signed_bits y = SCM_I_INUM (y_scm); \ \ if ((ip[2] & 0x1) ? !(unboxed) : (unboxed)) \ { \ scm_t_int32 offset = ip[2]; \ offset >>= 8; /* Sign-extending shift. */ \ NEXT (offset); \ } \ NEXT (3); \ } \ else \ { \ SCM res; \ SYNC_IP (); \ res = boxed (scm_from_uint64 (x), y_scm); \ CACHE_SP (); \ if ((ip[2] & 0x1) ? scm_is_false (res) : scm_is_true (res)) \ { \ scm_t_int32 offset = ip[2]; \ offset >>= 8; /* Sign-extending shift. */ \ NEXT (offset); \ } \ NEXT (3); \ } \ } while (0) /* br-if-u64-=-scm a:24 _:8 b:24 invert:1 _:7 offset:24 * * If the U64 value in A is = to the SCM value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (170, br_if_u64_ee_scm, "br-if-u64-=-scm", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_U64_SCM_COMPARISON(x, y, y >= 0 && (scm_t_uint64) y == x, scm_num_eq_p); } /* br-if-u64-<-scm a:24 _:8 b:24 invert:1 _:7 offset:24 * * If the U64 value in A is < than the SCM value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (171, br_if_u64_lt_scm, "br-if-u64-<-scm", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_U64_SCM_COMPARISON(x, y, y > 0 && (scm_t_uint64) y > x, scm_less_p); } /* br-if-u64-=-scm a:24 _:8 b:24 invert:1 _:7 offset:24 * * If the U64 value in A is <= than the SCM value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (172, br_if_u64_le_scm, "br-if-u64-<=-scm", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_U64_SCM_COMPARISON(x, y, y >= 0 && (scm_t_uint64) y >= x, scm_leq_p); } /* br-if-u64->-scm a:24 _:8 b:24 invert:1 _:7 offset:24 * * If the U64 value in A is > than the SCM value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (173, br_if_u64_gt_scm, "br-if-u64->-scm", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_U64_SCM_COMPARISON(x, y, y < 0 || (scm_t_uint64) y < x, scm_gr_p); } /* br-if-u64->=-scm a:24 _:8 b:24 invert:1 _:7 offset:24 * * If the U64 value in A is >= than the SCM value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (174, br_if_u64_ge_scm, "br-if-u64->=-scm", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_U64_SCM_COMPARISON(x, y, y <= 0 || (scm_t_uint64) y <= x, scm_geq_p); } /* integer->char a:12 b:12 * * Convert the U64 value in B to a Scheme character, and return it in * A. */ VM_DEFINE_OP (175, integer_to_char, "integer->char", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; scm_t_uint64 x; UNPACK_12_12 (op, dst, src); x = SP_REF_U64 (src); VM_ASSERT (x <= (scm_t_uint64) SCM_CODEPOINT_MAX, vm_error_out_of_range_uint64 ("integer->char", x)); SP_SET (dst, SCM_MAKE_ITAG8 ((scm_t_bits) (scm_t_wchar) x, scm_tc8_char)); NEXT (1); } /* char->integer a:12 b:12 * * Untag the character in B to U64, and return it in A. */ VM_DEFINE_OP (176, char_to_integer, "char->integer", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; SCM x; UNPACK_12_12 (op, dst, src); x = SP_REF (src); VM_VALIDATE_CHAR (x, "char->integer"); SP_SET_U64 (dst, SCM_CHAR (x)); NEXT (1); } /* ulogxor dst:8 a:8 b:8 * * Place the bitwise exclusive OR of the u64 values in A and B into * DST. */ VM_DEFINE_OP (177, ulogxor, "ulogxor", OP1 (X8_S8_S8_S8) | OP_DST) { scm_t_uint8 dst, a, b; UNPACK_8_8_8 (op, dst, a, b); SP_SET_U64 (dst, SP_REF_U64 (a) ^ SP_REF_U64 (b)); NEXT (1); } /* make-atomic-box dst:12 src:12 * * Create a new atomic box initialized to SRC, and place it in DST. */ VM_DEFINE_OP (178, make_atomic_box, "make-atomic-box", OP1 (X8_S12_S12) | OP_DST) { SCM box; scm_t_uint16 dst, src; UNPACK_12_12 (op, dst, src); SYNC_IP (); box = scm_inline_cell (thread, scm_tc7_atomic_box, SCM_UNPACK (SCM_UNSPECIFIED)); scm_atomic_set_scm (scm_atomic_box_loc (box), SP_REF (src)); SP_SET (dst, box); NEXT (1); } /* atomic-box-ref dst:12 src:12 * * Fetch the value of the atomic box at SRC into DST. */ VM_DEFINE_OP (179, atomic_box_ref, "atomic-box-ref", OP1 (X8_S12_S12) | OP_DST) { scm_t_uint16 dst, src; SCM box; UNPACK_12_12 (op, dst, src); box = SP_REF (src); VM_VALIDATE_ATOMIC_BOX (box, "atomic-box-ref"); SP_SET (dst, scm_atomic_ref_scm (scm_atomic_box_loc (box))); NEXT (1); } /* atomic-box-set! dst:12 src:12 * * Set the contents of the atomic box at DST to SRC. */ VM_DEFINE_OP (180, atomic_box_set, "atomic-box-set!", OP1 (X8_S12_S12)) { scm_t_uint16 dst, src; SCM box; UNPACK_12_12 (op, dst, src); box = SP_REF (dst); VM_VALIDATE_ATOMIC_BOX (box, "atomic-box-set!"); scm_atomic_set_scm (scm_atomic_box_loc (box), SP_REF (src)); NEXT (1); } /* atomic-box-swap! dst:12 box:12 _:8 val:24 * * Replace the contents of the atomic box at BOX to VAL and store the * previous value at DST. */ VM_DEFINE_OP (181, atomic_box_swap, "atomic-box-swap!", OP2 (X8_S12_S12, X8_S24) | OP_DST) { scm_t_uint16 dst, box; scm_t_uint32 val; SCM scm_box; UNPACK_12_12 (op, dst, box); UNPACK_24 (ip[1], val); scm_box = SP_REF (box); VM_VALIDATE_ATOMIC_BOX (scm_box, "atomic-box-swap!"); SP_SET (dst, scm_atomic_swap_scm (scm_atomic_box_loc (scm_box), SP_REF (val))); NEXT (2); } /* atomic-box-compare-and-swap! dst:12 box:12 _:8 expected:24 _:8 desired:24 * * Set the contents of the atomic box at DST to SET. */ VM_DEFINE_OP (182, atomic_box_compare_and_swap, "atomic-box-compare-and-swap!", OP3 (X8_S12_S12, X8_S24, X8_S24) | OP_DST) { scm_t_uint16 dst, box; scm_t_uint32 expected, desired; SCM scm_box, scm_expected; UNPACK_12_12 (op, dst, box); UNPACK_24 (ip[1], expected); UNPACK_24 (ip[2], desired); scm_box = SP_REF (box); VM_VALIDATE_ATOMIC_BOX (scm_box, "atomic-box-compare-and-swap!"); scm_expected = SP_REF (expected); scm_atomic_compare_and_swap_scm (scm_atomic_box_loc (scm_box), &scm_expected, SP_REF (desired)); SP_SET (dst, scm_expected); NEXT (3); } /* handle-interrupts _:24 * * Handle pending interrupts. */ VM_DEFINE_OP (183, handle_interrupts, "handle-interrupts", OP1 (X32)) { if (SCM_LIKELY (scm_is_null (scm_atomic_ref_scm (&thread->pending_asyncs)))) NEXT (1); if (thread->block_asyncs > 0) NEXT (1); { union scm_vm_stack_element *old_fp; size_t old_frame_size = FRAME_LOCALS_COUNT (); SCM proc = scm_i_async_pop (thread); /* No PUSH_CONTINUATION_HOOK, as we can't usefully POP_CONTINUATION_HOOK because there are no return values. */ /* Three slots: two for RA and dynamic link, one for proc. */ ALLOC_FRAME (old_frame_size + 3); /* Set up a frame that will return right back to this handle-interrupts opcode to handle any additional interrupts. */ old_fp = vp->fp; vp->fp = SCM_FRAME_SLOT (old_fp, old_frame_size + 1); SCM_FRAME_SET_DYNAMIC_LINK (vp->fp, old_fp); SCM_FRAME_SET_RETURN_ADDRESS (vp->fp, ip); SP_SET (0, proc); ip = (scm_t_uint32 *) vm_handle_interrupt_code; APPLY_HOOK (); NEXT (0); } } /* return-from-interrupt _:24 * * Return from handling an interrupt, discarding any return values and * stripping away the interrupt frame. */ VM_DEFINE_OP (184, return_from_interrupt, "return-from-interrupt", OP1 (X32)) { vp->sp = sp = SCM_FRAME_PREVIOUS_SP (vp->fp); ip = SCM_FRAME_RETURN_ADDRESS (vp->fp); vp->fp = SCM_FRAME_DYNAMIC_LINK (vp->fp); NEXT (0); } /* push-dynamic-state state:24 * * Save the current fluid bindings on the dynamic stack, and use STATE * instead. */ VM_DEFINE_OP (185, push_dynamic_state, "push-dynamic-state", OP1 (X8_S24)) { scm_t_uint32 state; UNPACK_24 (op, state); SYNC_IP (); scm_dynstack_push_dynamic_state (&thread->dynstack, SP_REF (state), thread->dynamic_state); NEXT (1); } /* pop-dynamic-state _:24 * * Restore the saved fluid bindings from the dynamic stack. */ VM_DEFINE_OP (186, pop_dynamic_state, "pop-dynamic-state", OP1 (X32)) { SYNC_IP (); scm_dynstack_unwind_dynamic_state (&thread->dynstack, thread->dynamic_state); NEXT (1); } /* br-if-f64-= a:12 b:12 invert:1 _:7 offset:24 * * If the F64 value in A is = to the F64 value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (187, br_if_f64_ee, "br-if-f64-=", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_F64_ARITHMETIC (==); } /* br-if-f64-< a:12 b:12 invert:1 _:7 offset:24 * * If the F64 value in A is < to the F64 value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (188, br_if_f64_lt, "br-if-f64-<", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_F64_ARITHMETIC (<); } /* br-if-f64-<= a:24 _:8 b:24 invert:1 _:7 offset:24 * * If the F64 value in A is <= than the F64 value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (189, br_if_f64_le, "br-if-f64-<=", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_F64_ARITHMETIC (<=); } /* br-if-f64-> a:24 _:8 b:24 invert:1 _:7 offset:24 * * If the F64 value in A is > than the F64 value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (190, br_if_f64_gt, "br-if-f64->", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_F64_ARITHMETIC (>); } /* br-if-uf4->= a:24 _:8 b:24 invert:1 _:7 offset:24 * * If the F64 value in A is >= than the F64 value in B, add OFFSET, a * signed 24-bit number, to the current instruction pointer. */ VM_DEFINE_OP (191, br_if_f64_ge, "br-if-f64->=", OP3 (X8_S24, X8_S24, B1_X7_L24)) { BR_F64_ARITHMETIC (>=); } /* string-set! dst:8 idx:8 src:8 * * Store the character SRC into the string DST at index IDX. */ VM_DEFINE_OP (192, string_set, "string-set!", OP1 (X8_S8_S8_S8)) { scm_t_uint8 dst, idx, src; SCM str, chr; scm_t_uint64 c_idx; UNPACK_8_8_8 (op, dst, idx, src); str = SP_REF (dst); c_idx = SP_REF_U64 (idx); chr = SP_REF (src); VM_VALIDATE_STRING (str, "string-ref"); VM_VALIDATE_INDEX (c_idx, scm_i_string_length (str), "string-ref"); /* If needed we can speed this up and only SYNC_IP + scm_i_string_writing if the string isn't already a non-shared stringbuf. */ SYNC_IP (); scm_i_string_start_writing (str); scm_i_string_set_x (str, c_idx, SCM_CHAR (chr)); scm_i_string_stop_writing (); NEXT (1); } VM_DEFINE_OP (193, unused_193, NULL, NOP) VM_DEFINE_OP (194, unused_194, NULL, NOP) VM_DEFINE_OP (195, unused_195, NULL, NOP) VM_DEFINE_OP (196, unused_196, NULL, NOP) VM_DEFINE_OP (197, unused_197, NULL, NOP) VM_DEFINE_OP (198, unused_198, NULL, NOP) VM_DEFINE_OP (199, unused_199, NULL, NOP) VM_DEFINE_OP (200, unused_200, NULL, NOP) VM_DEFINE_OP (201, unused_201, NULL, NOP) VM_DEFINE_OP (202, unused_202, NULL, NOP) VM_DEFINE_OP (203, unused_203, NULL, NOP) VM_DEFINE_OP (204, unused_204, NULL, NOP) VM_DEFINE_OP (205, unused_205, NULL, NOP) VM_DEFINE_OP (206, unused_206, NULL, NOP) VM_DEFINE_OP (207, unused_207, NULL, NOP) VM_DEFINE_OP (208, unused_208, NULL, NOP) VM_DEFINE_OP (209, unused_209, NULL, NOP) VM_DEFINE_OP (210, unused_210, NULL, NOP) VM_DEFINE_OP (211, unused_211, NULL, NOP) VM_DEFINE_OP (212, unused_212, NULL, NOP) VM_DEFINE_OP (213, unused_213, NULL, NOP) VM_DEFINE_OP (214, unused_214, NULL, NOP) VM_DEFINE_OP (215, unused_215, NULL, NOP) VM_DEFINE_OP (216, unused_216, NULL, NOP) VM_DEFINE_OP (217, unused_217, NULL, NOP) VM_DEFINE_OP (218, unused_218, NULL, NOP) VM_DEFINE_OP (219, unused_219, NULL, NOP) VM_DEFINE_OP (220, unused_220, NULL, NOP) VM_DEFINE_OP (221, unused_221, NULL, NOP) VM_DEFINE_OP (222, unused_222, NULL, NOP) VM_DEFINE_OP (223, unused_223, NULL, NOP) VM_DEFINE_OP (224, unused_224, NULL, NOP) VM_DEFINE_OP (225, unused_225, NULL, NOP) VM_DEFINE_OP (226, unused_226, NULL, NOP) VM_DEFINE_OP (227, unused_227, NULL, NOP) VM_DEFINE_OP (228, unused_228, NULL, NOP) VM_DEFINE_OP (229, unused_229, NULL, NOP) VM_DEFINE_OP (230, unused_230, NULL, NOP) VM_DEFINE_OP (231, unused_231, NULL, NOP) VM_DEFINE_OP (232, unused_232, NULL, NOP) VM_DEFINE_OP (233, unused_233, NULL, NOP) VM_DEFINE_OP (234, unused_234, NULL, NOP) VM_DEFINE_OP (235, unused_235, NULL, NOP) VM_DEFINE_OP (236, unused_236, NULL, NOP) VM_DEFINE_OP (237, unused_237, NULL, NOP) VM_DEFINE_OP (238, unused_238, NULL, NOP) VM_DEFINE_OP (239, unused_239, NULL, NOP) VM_DEFINE_OP (240, unused_240, NULL, NOP) VM_DEFINE_OP (241, unused_241, NULL, NOP) VM_DEFINE_OP (242, unused_242, NULL, NOP) VM_DEFINE_OP (243, unused_243, NULL, NOP) VM_DEFINE_OP (244, unused_244, NULL, NOP) VM_DEFINE_OP (245, unused_245, NULL, NOP) VM_DEFINE_OP (246, unused_246, NULL, NOP) VM_DEFINE_OP (247, unused_247, NULL, NOP) VM_DEFINE_OP (248, unused_248, NULL, NOP) VM_DEFINE_OP (249, unused_249, NULL, NOP) VM_DEFINE_OP (250, unused_250, NULL, NOP) VM_DEFINE_OP (251, unused_251, NULL, NOP) VM_DEFINE_OP (252, unused_252, NULL, NOP) VM_DEFINE_OP (253, unused_253, NULL, NOP) VM_DEFINE_OP (254, unused_254, NULL, NOP) VM_DEFINE_OP (255, unused_255, NULL, NOP) { vm_error_bad_instruction (op); abort (); /* never reached */ } END_DISPATCH_SWITCH; } #undef ABORT_CONTINUATION_HOOK #undef ALIGNED_P #undef APPLY_HOOK #undef ARGS1 #undef ARGS2 #undef BEGIN_DISPATCH_SWITCH #undef BINARY_INTEGER_OP #undef BR_ARITHMETIC #undef BR_BINARY #undef BR_NARGS #undef BR_UNARY #undef BV_FIXABLE_INT_REF #undef BV_FIXABLE_INT_SET #undef BV_FLOAT_REF #undef BV_FLOAT_SET #undef BV_INT_REF #undef BV_INT_SET #undef CACHE_REGISTER #undef END_DISPATCH_SWITCH #undef FREE_VARIABLE_REF #undef INIT #undef INUM_MAX #undef INUM_MIN #undef FP_REF #undef FP_SET #undef FP_SLOT #undef SP_REF #undef SP_SET #undef NEXT #undef NEXT_HOOK #undef NEXT_JUMP #undef POP_CONTINUATION_HOOK #undef PUSH_CONTINUATION_HOOK #undef RETURN #undef RUN_HOOK #undef RUN_HOOK0 #undef RUN_HOOK1 #undef SYNC_IP #undef UNPACK_8_8_8 #undef UNPACK_8_16 #undef UNPACK_16_8 #undef UNPACK_12_12 #undef UNPACK_24 #undef VARIABLE_BOUNDP #undef VARIABLE_REF #undef VARIABLE_SET #undef VM_CHECK_FREE_VARIABLE #undef VM_CHECK_OBJECT #undef VM_CHECK_UNDERFLOW #undef VM_DEFINE_OP #undef VM_INSTRUCTION_TO_LABEL #undef VM_USE_HOOKS #undef VM_VALIDATE_ATOMIC_BOX #undef VM_VALIDATE_BYTEVECTOR #undef VM_VALIDATE_PAIR #undef VM_VALIDATE_STRUCT /* (defun renumber-ops () "start from top of buffer and renumber 'VM_DEFINE_FOO (\n' sequences" (interactive "") (save-excursion (let ((counter -1)) (goto-char (point-min)) (while (re-search-forward "^ *VM_DEFINE_[^ ]+ (\\([^,]+\\)," (point-max) t) (replace-match (number-to-string (setq counter (1+ counter))) t t nil 1))))) (renumber-ops) */ /* Local Variables: c-file-style: "gnu" End: */