mirror of
https://git.code.sf.net/p/quake/quakeforge
synced 2024-12-19 17:12:33 +00:00
d9d0a80752
The extend instruction is for loading narrower data types into wider data types, eg, single element into 2, 3, or 4 element types, with a small set of extension schemes: 0, 1, -1, copy (for 1->any and 2 -> 4). Possibly most importantly, it works with unaligned data. Progress towards #30
2906 lines
78 KiB
C
2906 lines
78 KiB
C
/*
|
|
pr_exec.c
|
|
|
|
(description)
|
|
|
|
Copyright (C) 1996-1997 Id Software, Inc.
|
|
|
|
This program is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU General Public License
|
|
as published by the Free Software Foundation; either version 2
|
|
of the License, or (at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
|
|
|
|
See the GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to:
|
|
|
|
Free Software Foundation, Inc.
|
|
59 Temple Place - Suite 330
|
|
Boston, MA 02111-1307, USA
|
|
|
|
*/
|
|
#ifdef HAVE_CONFIG_H
|
|
# include "config.h"
|
|
#endif
|
|
|
|
#ifdef HAVE_STRING_H
|
|
# include <string.h>
|
|
#endif
|
|
#ifdef HAVE_STRINGS_H
|
|
# include <strings.h>
|
|
#endif
|
|
#include <signal.h>
|
|
#include <stdarg.h>
|
|
|
|
#include "QF/cvar.h"
|
|
#include "QF/dstring.h"
|
|
#include "QF/mathlib.h"
|
|
#include "QF/progs.h"
|
|
#include "QF/sys.h"
|
|
#include "QF/zone.h"
|
|
|
|
#include "QF/simd/vec2d.h"
|
|
#include "QF/simd/vec2f.h"
|
|
#include "QF/simd/vec2i.h"
|
|
#include "QF/simd/vec4d.h"
|
|
#include "QF/simd/vec4f.h"
|
|
#include "QF/simd/vec4i.h"
|
|
#include "compat.h"
|
|
|
|
const char *prdebug_names[] = {
|
|
[prd_none] = "none",
|
|
[prd_trace] = "trace",
|
|
[prd_breakpoint] = "breakpoint",
|
|
[prd_watchpoint] = "watchpoint",
|
|
[prd_subenter] = "subenter",
|
|
[prd_subexit] = "subexit",
|
|
[prd_begin] = "begin",
|
|
[prd_terminate] = "terminate",
|
|
[prd_runerror] = "runerror",
|
|
[prd_error] = "error",
|
|
};
|
|
|
|
/*
|
|
PR_RunError
|
|
|
|
Aborts the currently executing function
|
|
*/
|
|
VISIBLE void
|
|
PR_RunError (progs_t * pr, const char *error, ...)
|
|
{
|
|
dstring_t *string = dstring_new ();//FIXME leaks when debugging
|
|
va_list argptr;
|
|
|
|
va_start (argptr, error);
|
|
dvsprintf (string, error, argptr);
|
|
va_end (argptr);
|
|
|
|
if (pr->debug_handler) {
|
|
pr->debug_handler (prd_runerror, string->str, pr->debug_data);
|
|
// not expected to return, but if so, behave as if there was no handler
|
|
}
|
|
|
|
Sys_Printf ("%s\n", string->str);
|
|
|
|
PR_DumpState (pr);
|
|
|
|
// dump the stack so PR_Error can shutdown functions
|
|
pr->pr_depth = 0;
|
|
pr->localstack_used = 0;
|
|
|
|
PR_Error (pr, "Program error: %s", string->str);
|
|
}
|
|
|
|
VISIBLE pr_stashed_params_t *
|
|
_PR_SaveParams (progs_t *pr, pr_stashed_params_t *params)
|
|
{
|
|
int i;
|
|
int size = pr->pr_param_size * sizeof (pr_type_t);
|
|
|
|
params->param_ptrs[0] = pr->pr_params[0];
|
|
params->param_ptrs[1] = pr->pr_params[1];
|
|
pr->pr_params[0] = pr->pr_real_params[0];
|
|
pr->pr_params[1] = pr->pr_real_params[1];
|
|
for (i = 0; i < pr->pr_argc; i++) {
|
|
memcpy (params->params + i * pr->pr_param_size,
|
|
pr->pr_real_params[i], size);
|
|
if (i < 2) { //XXX FIXME what the what?!?
|
|
memcpy (pr->pr_real_params[i], params->param_ptrs[0], size);
|
|
}
|
|
}
|
|
params->argc = pr->pr_argc;
|
|
return params;
|
|
}
|
|
|
|
VISIBLE void
|
|
PR_RestoreParams (progs_t *pr, pr_stashed_params_t *params)
|
|
{
|
|
int i;
|
|
int size = pr->pr_param_size * sizeof (pr_type_t);
|
|
|
|
pr->pr_params[0] = params->param_ptrs[0];
|
|
pr->pr_params[1] = params->param_ptrs[1];
|
|
pr->pr_argc = params->argc;
|
|
for (i = 0; i < pr->pr_argc; i++) {
|
|
memcpy (pr->pr_real_params[i],
|
|
params->params + i * pr->pr_param_size, size);
|
|
}
|
|
}
|
|
|
|
VISIBLE inline void
|
|
PR_PushFrame (progs_t *pr)
|
|
{
|
|
prstack_t *frame;
|
|
|
|
if (pr->pr_depth == PR_MAX_STACK_DEPTH)
|
|
PR_RunError (pr, "stack overflow");
|
|
|
|
frame = pr->pr_stack + pr->pr_depth++;
|
|
|
|
frame->staddr = pr->pr_xstatement;
|
|
if (pr->globals.stack) {
|
|
frame->stack_ptr = *pr->globals.stack;
|
|
}
|
|
frame->bases = pr->pr_bases;
|
|
frame->func = pr->pr_xfunction;
|
|
frame->tstr = pr->pr_xtstr;
|
|
frame->return_ptr = pr->pr_return;
|
|
|
|
pr->pr_xtstr = pr->pr_pushtstr;
|
|
pr->pr_pushtstr = 0;
|
|
pr->pr_xfunction = 0;
|
|
}
|
|
|
|
VISIBLE inline void
|
|
PR_PopFrame (progs_t *pr)
|
|
{
|
|
prstack_t *frame;
|
|
|
|
if (pr->pr_depth <= 0)
|
|
PR_Error (pr, "prog stack underflow");
|
|
|
|
if (pr->pr_xtstr)
|
|
PR_FreeTempStrings (pr);
|
|
// normally, this won't happen, but if a builtin pushed a temp string
|
|
// when calling a function and the callee was another builtin that
|
|
// did not call a progs function, then the push strings will still be
|
|
// valid because PR_EnterFunction was never called
|
|
// however, not if a temp string survived: better to hold on to the push
|
|
// strings a little longer than lose one erroneously
|
|
if (!pr->pr_xtstr && pr->pr_pushtstr) {
|
|
pr->pr_xtstr = pr->pr_pushtstr;
|
|
pr->pr_pushtstr = 0;
|
|
PR_FreeTempStrings (pr);
|
|
}
|
|
|
|
// up stack
|
|
frame = pr->pr_stack + --pr->pr_depth;
|
|
|
|
pr->pr_return = frame->return_ptr;
|
|
pr->pr_xfunction = frame->func;
|
|
pr->pr_xstatement = frame->staddr;
|
|
pr->pr_xtstr = frame->tstr;
|
|
pr->pr_bases = frame->bases;
|
|
// restore data stack (discard any locals)
|
|
if (pr->globals.stack) {
|
|
*pr->globals.stack = frame->stack_ptr;
|
|
}
|
|
}
|
|
|
|
static __attribute__((pure)) long
|
|
align_offset (long offset, dparmsize_t paramsize)
|
|
{
|
|
int mask = (1 << paramsize.alignment) - 1;
|
|
return (offset + mask) & ~mask;
|
|
}
|
|
|
|
static void
|
|
copy_param (pr_type_t *dst, pr_type_t *src, size_t size)
|
|
{
|
|
while (size--) {
|
|
memcpy (dst++, src++, sizeof (pr_type_t));
|
|
}
|
|
}
|
|
|
|
/** Setup the stackframe prior to calling a progs function. Saves all local
|
|
data the called function will trample on and copies the parameters used
|
|
by the function into the function's local data space.
|
|
\param pr pointer to progs_t VM struct
|
|
\param f pointer to the descriptor for the called function
|
|
\note Passing a descriptor for a builtin function will result in
|
|
undefined behavior.
|
|
*/
|
|
static void
|
|
PR_EnterFunction (progs_t *pr, bfunction_t *f)
|
|
{
|
|
pr_int_t i;
|
|
pr_type_t *dstParams[PR_MAX_PARAMS];
|
|
pr_ptr_t paramofs = 0;
|
|
|
|
if (pr->pr_trace && !pr->debug_handler) {
|
|
Sys_Printf ("Entering function %s\n",
|
|
PR_GetString (pr, f->descriptor->name));
|
|
}
|
|
|
|
PR_PushFrame (pr);
|
|
|
|
//Sys_Printf("%s:\n", PR_GetString(pr,f->name));
|
|
pr->pr_xfunction = f;
|
|
pr->pr_xstatement = f->first_statement - 1; // offset the st++
|
|
|
|
if (pr->progs->version == PROG_VERSION) {
|
|
return;
|
|
}
|
|
|
|
if (f->numparams > 0) {
|
|
paramofs = f->params_start;
|
|
for (i = 0; i < f->numparams; i++) {
|
|
paramofs = align_offset (paramofs, f->param_size[i]);
|
|
dstParams[i] = pr->pr_globals + paramofs;
|
|
paramofs += f->param_size[i].size;
|
|
if (pr->pr_params[i] != pr->pr_real_params[i]) {
|
|
copy_param (pr->pr_real_params[i], pr->pr_params[i],
|
|
f->param_size[i].size);
|
|
pr->pr_params[i] = pr->pr_real_params[i];
|
|
}
|
|
}
|
|
} else if (f->numparams < 0) {
|
|
paramofs = f->params_start + 2; // argc and argv
|
|
for (i = 0; i < -f->numparams - 1; i++) {
|
|
paramofs = align_offset (paramofs, f->param_size[i]);
|
|
dstParams[i] = pr->pr_globals + paramofs;
|
|
paramofs += f->param_size[i].size;
|
|
if (pr->pr_params[i] != pr->pr_real_params[i]) {
|
|
copy_param (pr->pr_real_params[i], pr->pr_params[i],
|
|
f->param_size[i].size);
|
|
pr->pr_params[i] = pr->pr_real_params[i];
|
|
}
|
|
}
|
|
dparmsize_t paramsize = { pr->pr_param_size, pr->pr_param_alignment };
|
|
paramofs = align_offset (paramofs, paramsize );
|
|
if (i < PR_MAX_PARAMS) {
|
|
dstParams[i] = pr->pr_globals + paramofs;
|
|
}
|
|
for (; i < pr->pr_argc; i++) {
|
|
if (pr->pr_params[i] != pr->pr_real_params[i]) {
|
|
copy_param (pr->pr_real_params[i], pr->pr_params[i],
|
|
paramsize.size);
|
|
pr->pr_params[i] = pr->pr_real_params[i];
|
|
}
|
|
}
|
|
}
|
|
|
|
// save off any locals that the new function steps on
|
|
if (pr->localstack_used + f->locals > PR_LOCAL_STACK_SIZE)
|
|
PR_RunError (pr, "PR_EnterFunction: locals stack overflow");
|
|
|
|
memcpy (&pr->localstack[pr->localstack_used],
|
|
&pr->pr_globals[f->params_start],
|
|
sizeof (pr_type_t) * f->locals);
|
|
pr->localstack_used += f->locals;
|
|
|
|
if (pr_deadbeef_locals) {
|
|
for (pr_uint_t i = f->params_start;
|
|
i < f->params_start + f->locals; i++) {
|
|
pr->pr_globals[i].value = 0xdeadbeef;
|
|
}
|
|
}
|
|
|
|
// copy parameters
|
|
if (f->numparams >= 0) {
|
|
for (i = 0; i < f->numparams; i++) {
|
|
copy_param (dstParams[i], pr->pr_params[i], f->param_size[i].size);
|
|
}
|
|
} else {
|
|
int copy_args;
|
|
pr_type_t *argc = &pr->pr_globals[f->params_start + 0];
|
|
pr_type_t *argv = &pr->pr_globals[f->params_start + 1];
|
|
for (i = 0; i < -f->numparams - 1; i++) {
|
|
copy_param (dstParams[i], pr->pr_params[i], f->param_size[i].size);
|
|
}
|
|
copy_args = pr->pr_argc - i;
|
|
PR_PTR (int, argc) = copy_args;
|
|
PR_PTR (ptr, argv) = PR_SetPointer (pr, dstParams[i]);
|
|
if (i < PR_MAX_PARAMS) {
|
|
memcpy (dstParams[i], pr->pr_params[i],
|
|
(copy_args * pr->pr_param_size) * sizeof (pr_type_t));
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
PR_LeaveFunction (progs_t *pr, int to_engine)
|
|
{
|
|
bfunction_t *f = pr->pr_xfunction;
|
|
|
|
PR_PopFrame (pr);
|
|
|
|
if (pr->pr_trace && !pr->debug_handler) {
|
|
Sys_Printf ("Leaving function %s\n",
|
|
PR_GetString (pr, f->descriptor->name));
|
|
if (to_engine) {
|
|
Sys_Printf ("Returning to engine\n");
|
|
} else {
|
|
bfunction_t *rf = pr->pr_xfunction;
|
|
if (rf) {
|
|
Sys_Printf ("Returning to function %s\n",
|
|
PR_GetString (pr, rf->descriptor->name));
|
|
}
|
|
}
|
|
}
|
|
|
|
if (pr->progs->version == PROG_VERSION) {
|
|
return;
|
|
}
|
|
|
|
// restore locals from the stack
|
|
pr->localstack_used -= f->locals;
|
|
if (pr->localstack_used < 0)
|
|
PR_RunError (pr, "PR_LeaveFunction: locals stack underflow");
|
|
|
|
memcpy (&pr->pr_globals[f->params_start],
|
|
&pr->localstack[pr->localstack_used],
|
|
sizeof (pr_type_t) * f->locals);
|
|
}
|
|
|
|
VISIBLE void
|
|
PR_BoundsCheckSize (progs_t *pr, pr_ptr_t addr, unsigned size)
|
|
{
|
|
if (addr < pr->null_size)
|
|
PR_RunError (pr, "null pointer access");
|
|
if (addr >= pr->globals_size
|
|
|| size > (unsigned) (pr->globals_size - addr))
|
|
PR_RunError (pr, "invalid memory access: %d (0 to %d-%d)", addr,
|
|
pr->globals_size, size);
|
|
if (pr_boundscheck >= 2
|
|
&& PR_GetPointer (pr, addr + size) > (pr_type_t *) pr->zone) {
|
|
void *mem = (void *) PR_GetPointer (pr, addr);
|
|
Z_CheckPointer (pr->zone, mem, size * sizeof (pr_type_t));
|
|
}
|
|
}
|
|
|
|
VISIBLE void
|
|
PR_BoundsCheck (progs_t *pr, int addr, etype_t type)
|
|
{
|
|
PR_BoundsCheckSize (pr, addr, pr_type_size[type]);
|
|
}
|
|
|
|
#define OPA(type) (*((pr_##type##_t *) (op_a)))
|
|
#define OPB(type) (*((pr_##type##_t *) (op_b)))
|
|
#define OPC(type) (*((pr_##type##_t *) (op_c)))
|
|
|
|
/*
|
|
This gets around the problem of needing to test for -0.0 but denormals
|
|
causing exceptions (or wrong results for what we need) on the alpha.
|
|
*/
|
|
#define FNZ(x) ((x) & ~0x80000000u)
|
|
|
|
static int
|
|
signal_hook (int sig, void *data)
|
|
{
|
|
progs_t *pr = (progs_t *) data;
|
|
|
|
if (sig == SIGFPE && pr_faultchecks) {
|
|
dstatement_t *st;
|
|
pr_type_t *op_a, *op_b, *op_c;
|
|
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
op_a = pr->pr_globals + st->a;
|
|
op_b = pr->pr_globals + st->b;
|
|
op_c = pr->pr_globals + st->c;
|
|
|
|
switch (st->op) {
|
|
case OP_DIV_F_v6p:
|
|
if ((OPA(int) & 0x80000000)
|
|
^ (OPB(int) & 0x80000000))
|
|
OPC(int) = 0xff7fffff;
|
|
else
|
|
OPC(int) = 0x7f7fffff;
|
|
return 1;
|
|
case OP_DIV_I_v6p:
|
|
if (OPA(int) & 0x80000000)
|
|
OPC(int) = -0x80000000;
|
|
else
|
|
OPC(int) = 0x7fffffff;
|
|
return 1;
|
|
case OP_MOD_I_v6p:
|
|
case OP_MOD_F_v6p:
|
|
case OP_REM_I_v6p:
|
|
case OP_REM_F_v6p:
|
|
OPC(int) = 0x00000000;
|
|
return 1;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
PR_DumpState (pr);
|
|
fflush (stdout);
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
error_handler (void *data)
|
|
{
|
|
progs_t *pr = (progs_t *) data;
|
|
PR_DumpState (pr);
|
|
fflush (stdout);
|
|
}
|
|
|
|
VISIBLE int
|
|
PR_CallFunction (progs_t *pr, pr_func_t fnum, pr_type_t *return_ptr)
|
|
{
|
|
bfunction_t *f;
|
|
|
|
if (!fnum)
|
|
PR_RunError (pr, "NULL function");
|
|
if (!return_ptr || return_ptr == pr->pr_globals) {
|
|
return_ptr = pr->pr_return_buffer;
|
|
}
|
|
f = pr->function_table + fnum;
|
|
if (f->first_statement < 0) {
|
|
// negative statements are built in functions
|
|
if (pr->progs->version == PROG_VERSION) {
|
|
PR_SetupParams (pr, 0, 0);
|
|
}
|
|
if (pr->pr_trace && !pr->debug_handler) {
|
|
Sys_Printf ("Calling builtin %s @ %p\n",
|
|
PR_GetString (pr, f->descriptor->name), f->func);
|
|
}
|
|
pr_type_t *saved_return = pr->pr_return;
|
|
int builtin_depth = pr->pr_depth;
|
|
pr->pr_return = return_ptr;
|
|
f->func (pr, f->data);
|
|
f->profile++; // to show number times the builtin is called
|
|
if (builtin_depth == pr->pr_depth) {
|
|
pr->pr_return = saved_return;
|
|
} else if (builtin_depth < pr->pr_depth) {
|
|
pr->pr_stack[builtin_depth].return_ptr = saved_return;
|
|
}
|
|
return 0;
|
|
} else {
|
|
PR_EnterFunction (pr, f);
|
|
pr->pr_return = return_ptr;
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
static void
|
|
check_stack_pointer (progs_t *pr, pr_ptr_t stack, int size)
|
|
{
|
|
if (stack & 3) {
|
|
PR_RunError (pr, "Progs stack not aligned");
|
|
}
|
|
if (stack < pr->stack_bottom) {
|
|
PR_RunError (pr, "Progs stack overflow");
|
|
}
|
|
if (stack > pr->globals_size - size) {
|
|
PR_RunError (pr, "Progs stack underflow");
|
|
}
|
|
}
|
|
|
|
VISIBLE pr_type_t *
|
|
PR_SetupParams (progs_t *pr, int num_params, int min_alignment)
|
|
{
|
|
if (pr->progs->version < PROG_VERSION) {
|
|
if (num_params > PR_MAX_PARAMS) {
|
|
PR_Error (pr, "attempt to settup more than %d params",
|
|
PR_MAX_PARAMS);
|
|
}
|
|
pr->pr_params[0] = pr->pr_real_params[0];
|
|
pr->pr_params[1] = pr->pr_real_params[1];
|
|
return pr->pr_real_params[0];
|
|
}
|
|
int offset = num_params * 4;
|
|
if (min_alignment < 4) {
|
|
min_alignment = 4;
|
|
}
|
|
pr_ptr_t mask = ~(min_alignment - 1);
|
|
pr_ptr_t stack = (*pr->globals.stack - offset) & mask;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 0);
|
|
}
|
|
*pr->globals.stack = stack;
|
|
pr->pr_params[0] = pr->pr_globals + stack;
|
|
num_params = max (num_params, PR_MAX_PARAMS);
|
|
for (int i = 1; i < num_params; i++) {
|
|
pr->pr_params[i] = pr->pr_params[0] + i * 4;
|
|
}
|
|
return pr->pr_params[0];
|
|
}
|
|
|
|
static inline void
|
|
pr_memset (pr_type_t *dst, int val, pr_uint_t count)
|
|
{
|
|
while (count-- > 0) {
|
|
(*dst++).value = val;
|
|
}
|
|
}
|
|
|
|
static void
|
|
pr_exec_quakec (progs_t *pr, int exitdepth)
|
|
{
|
|
int profile, startprofile;
|
|
int fldofs;
|
|
pr_uint_t pointer;
|
|
dstatement_t *st;
|
|
pr_type_t *ptr;
|
|
pr_type_t old_val = {0};
|
|
|
|
// make a stack frame
|
|
startprofile = profile = 0;
|
|
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
|
|
if (pr->watch) {
|
|
old_val = *pr->watch;
|
|
}
|
|
while (1) {
|
|
pr_type_t *op_a, *op_b, *op_c;
|
|
|
|
st++;
|
|
++pr->pr_xstatement;
|
|
if (pr->pr_xstatement != st - pr->pr_statements)
|
|
PR_RunError (pr, "internal error");
|
|
if (++profile > 1000000 && !pr->no_exec_limit) {
|
|
PR_RunError (pr, "runaway loop error");
|
|
}
|
|
|
|
op_a = pr->pr_globals + st->a;
|
|
op_b = pr->pr_globals + st->b;
|
|
op_c = pr->pr_globals + st->c;
|
|
|
|
if (pr->pr_trace) {
|
|
if (pr->debug_handler) {
|
|
pr->debug_handler (prd_trace, 0, pr->debug_data);
|
|
} else {
|
|
PR_PrintStatement (pr, st, 1);
|
|
}
|
|
}
|
|
|
|
if (st->op & OP_BREAK) {
|
|
if (pr->debug_handler) {
|
|
pr->debug_handler (prd_breakpoint, 0, pr->debug_data);
|
|
} else {
|
|
PR_RunError (pr, "breakpoint hit");
|
|
}
|
|
}
|
|
|
|
pr_opcode_v6p_e op = st->op & ~OP_BREAK;
|
|
switch (op) {
|
|
case OP_ADD_D_v6p:
|
|
OPC(double) = OPA(double) + OPB(double);
|
|
break;
|
|
case OP_ADD_F_v6p:
|
|
OPC(float) = OPA(float) + OPB(float);
|
|
break;
|
|
case OP_ADD_V_v6p:
|
|
VectorAdd (&OPA(float), &OPB(float), &OPC(float));
|
|
break;
|
|
case OP_ADD_Q_v6p:
|
|
QuatAdd (&OPA(float), &OPB(float), &OPC(float));
|
|
break;
|
|
case OP_ADD_S_v6p:
|
|
OPC(string) = PR_CatStrings (pr,
|
|
PR_GetString (pr,
|
|
OPA(string)),
|
|
PR_GetString (pr,
|
|
OPB(string)));
|
|
break;
|
|
case OP_SUB_D_v6p:
|
|
OPC(double) = OPA(double) - OPB(double);
|
|
break;
|
|
case OP_SUB_F_v6p:
|
|
OPC(float) = OPA(float) - OPB(float);
|
|
break;
|
|
case OP_SUB_V_v6p:
|
|
VectorSubtract (&OPA(float), &OPB(float),
|
|
&OPC(float));
|
|
break;
|
|
case OP_SUB_Q_v6p:
|
|
QuatSubtract (&OPA(float), &OPB(float), &OPC(float));
|
|
break;
|
|
case OP_MUL_D_v6p:
|
|
OPC(double) = OPA(double) * OPB(double);
|
|
break;
|
|
case OP_MUL_F_v6p:
|
|
OPC(float) = OPA(float) * OPB(float);
|
|
break;
|
|
case OP_MUL_V_v6p:
|
|
OPC(float) = DotProduct (&OPA(float), &OPB(float));
|
|
break;
|
|
case OP_MUL_DV_v6p:
|
|
{
|
|
// avoid issues with the likes of x = x.x * x;
|
|
// makes for faster code, too
|
|
double scale = OPA(double);
|
|
VectorScale (&OPB(float), scale, &OPC(float));
|
|
}
|
|
break;
|
|
case OP_MUL_VD_v6p:
|
|
{
|
|
// avoid issues with the likes of x = x * x.x;
|
|
// makes for faster code, too
|
|
double scale = OPB(double);
|
|
VectorScale (&OPA(float), scale, &OPC(float));
|
|
}
|
|
break;
|
|
case OP_MUL_FV_v6p:
|
|
{
|
|
// avoid issues with the likes of x = x.x * x;
|
|
// makes for faster code, too
|
|
float scale = OPA(float);
|
|
VectorScale (&OPB(float), scale, &OPC(float));
|
|
}
|
|
break;
|
|
case OP_MUL_VF_v6p:
|
|
{
|
|
// avoid issues with the likes of x = x * x.x;
|
|
// makes for faster code, too
|
|
float scale = OPB(float);
|
|
VectorScale (&OPA(float), scale, &OPC(float));
|
|
}
|
|
break;
|
|
case OP_MUL_Q_v6p:
|
|
QuatMult (&OPA(float), &OPB(float), &OPC(float));
|
|
break;
|
|
case OP_MUL_QV_v6p:
|
|
QuatMultVec (&OPA(float), &OPB(float), &OPC(float));
|
|
break;
|
|
case OP_MUL_DQ_v6p:
|
|
{
|
|
// avoid issues with the likes of x = x.s * x;
|
|
// makes for faster code, too
|
|
double scale = OPA(double);
|
|
QuatScale (&OPB(float), scale, &OPC(float));
|
|
}
|
|
break;
|
|
case OP_MUL_QD_v6p:
|
|
{
|
|
// avoid issues with the likes of x = x * x.s;
|
|
// makes for faster code, too
|
|
double scale = OPB(double);
|
|
QuatScale (&OPA(float), scale, &OPC(float));
|
|
}
|
|
break;
|
|
case OP_MUL_FQ_v6p:
|
|
{
|
|
// avoid issues with the likes of x = x.s * x;
|
|
// makes for faster code, too
|
|
float scale = OPA(float);
|
|
QuatScale (&OPB(float), scale, &OPC(float));
|
|
}
|
|
break;
|
|
case OP_MUL_QF_v6p:
|
|
{
|
|
// avoid issues with the likes of x = x * x.s;
|
|
// makes for faster code, too
|
|
float scale = OPB(float);
|
|
QuatScale (&OPA(float), scale, &OPC(float));
|
|
}
|
|
break;
|
|
case OP_CONJ_Q_v6p:
|
|
QuatConj (&OPA(float), &OPC(float));
|
|
break;
|
|
case OP_DIV_D_v6p:
|
|
OPC(double) = OPA(double) / OPB(double);
|
|
break;
|
|
case OP_DIV_F_v6p:
|
|
OPC(float) = OPA(float) / OPB(float);
|
|
break;
|
|
case OP_BITAND_v6p:
|
|
OPC(float) = (int) OPA(float) & (int) OPB(float);
|
|
break;
|
|
case OP_BITOR_v6p:
|
|
OPC(float) = (int) OPA(float) | (int) OPB(float);
|
|
break;
|
|
case OP_BITXOR_F_v6p:
|
|
OPC(float) = (int) OPA(float) ^ (int) OPB(float);
|
|
break;
|
|
case OP_BITNOT_F_v6p:
|
|
OPC(float) = ~ (int) OPA(float);
|
|
break;
|
|
case OP_SHL_F_v6p:
|
|
OPC(float) = (int) OPA(float) << (int) OPB(float);
|
|
break;
|
|
case OP_SHR_F_v6p:
|
|
OPC(float) = (int) OPA(float) >> (int) OPB(float);
|
|
break;
|
|
case OP_SHL_I_v6p:
|
|
OPC(int) = OPA(int) << OPB(int);
|
|
break;
|
|
case OP_SHR_I_v6p:
|
|
OPC(int) = OPA(int) >> OPB(int);
|
|
break;
|
|
case OP_SHR_U_v6p:
|
|
OPC(uint) = OPA(uint) >> OPB(int);
|
|
break;
|
|
case OP_GE_F_v6p:
|
|
OPC(float) = OPA(float) >= OPB(float);
|
|
break;
|
|
case OP_LE_F_v6p:
|
|
OPC(float) = OPA(float) <= OPB(float);
|
|
break;
|
|
case OP_GT_F_v6p:
|
|
OPC(float) = OPA(float) > OPB(float);
|
|
break;
|
|
case OP_LT_F_v6p:
|
|
OPC(float) = OPA(float) < OPB(float);
|
|
break;
|
|
case OP_AND_v6p: // OPA and OPB have to be float for -0.0
|
|
OPC(int) = FNZ (OPA(uint)) && FNZ (OPB(uint));
|
|
break;
|
|
case OP_OR_v6p: // OPA and OPB have to be float for -0.0
|
|
OPC(int) = FNZ (OPA(uint)) || FNZ (OPB(uint));
|
|
break;
|
|
case OP_NOT_F_v6p:
|
|
OPC(int) = !FNZ (OPA(uint));
|
|
break;
|
|
case OP_NOT_V_v6p:
|
|
OPC(int) = VectorIsZero (&OPA(float));
|
|
break;
|
|
case OP_NOT_Q_v6p:
|
|
OPC(int) = QuatIsZero (&OPA(float));
|
|
break;
|
|
case OP_NOT_S_v6p:
|
|
OPC(int) = !OPA(string) || !*PR_GetString (pr, OPA(string));
|
|
break;
|
|
case OP_NOT_FN_v6p:
|
|
OPC(int) = !OPA(func);
|
|
break;
|
|
case OP_NOT_ENT_v6p:
|
|
OPC(int) = !OPA(entity);
|
|
break;
|
|
case OP_EQ_F_v6p:
|
|
OPC(int) = OPA(float) == OPB(float);
|
|
break;
|
|
case OP_EQ_V_v6p:
|
|
OPC(int) = VectorCompare (&OPA(float), &OPB(float));
|
|
break;
|
|
case OP_EQ_Q_v6p:
|
|
OPC(int) = QuatCompare (&OPA(float), &OPB(float));
|
|
break;
|
|
case OP_EQ_E_v6p:
|
|
OPC(int) = OPA(field) == OPB(field);
|
|
break;
|
|
case OP_EQ_FN_v6p:
|
|
OPC(int) = OPA(func) == OPB(func);
|
|
break;
|
|
case OP_NE_F_v6p:
|
|
OPC(int) = OPA(float) != OPB(float);
|
|
break;
|
|
case OP_NE_V_v6p:
|
|
OPC(int) = !VectorCompare (&OPA(float), &OPB(float));
|
|
break;
|
|
case OP_NE_Q_v6p:
|
|
OPC(int) = !QuatCompare (&OPA(float), &OPB(float));
|
|
break;
|
|
case OP_LE_S_v6p:
|
|
case OP_GE_S_v6p:
|
|
case OP_LT_S_v6p:
|
|
case OP_GT_S_v6p:
|
|
case OP_NE_S_v6p:
|
|
case OP_EQ_S_v6p:
|
|
{
|
|
int cmp = strcmp (PR_GetString (pr, OPA(string)),
|
|
PR_GetString (pr, OPB(string)));
|
|
switch (st->op) {
|
|
case OP_LE_S_v6p: cmp = (cmp <= 0); break;
|
|
case OP_GE_S_v6p: cmp = (cmp >= 0); break;
|
|
case OP_LT_S_v6p: cmp = (cmp < 0); break;
|
|
case OP_GT_S_v6p: cmp = (cmp > 0); break;
|
|
case OP_NE_S_v6p: break;
|
|
case OP_EQ_S_v6p: cmp = !cmp; break;
|
|
default: break;
|
|
}
|
|
OPC(int) = cmp;
|
|
}
|
|
break;
|
|
case OP_NE_E_v6p:
|
|
OPC(int) = OPA(entity) != OPB(entity);
|
|
break;
|
|
case OP_NE_FN_v6p:
|
|
OPC(int) = OPA(func) != OPB(func);
|
|
break;
|
|
|
|
// ==================
|
|
case OP_STORE_F_v6p:
|
|
case OP_STORE_ENT_v6p:
|
|
case OP_STORE_FLD_v6p: // integers
|
|
case OP_STORE_S_v6p:
|
|
case OP_STORE_FN_v6p: // pointers
|
|
case OP_STORE_I_v6p:
|
|
case OP_STORE_P_v6p:
|
|
OPB(int) = OPA(int);
|
|
break;
|
|
case OP_STORE_V_v6p:
|
|
VectorCopy (&OPA(float), &OPB(float));
|
|
break;
|
|
case OP_STORE_Q_v6p:
|
|
QuatCopy (&OPA(float), &OPB(float));
|
|
break;
|
|
case OP_STORE_D_v6p:
|
|
OPB(double) = OPA(double);
|
|
break;
|
|
|
|
case OP_STOREP_F_v6p:
|
|
case OP_STOREP_ENT_v6p:
|
|
case OP_STOREP_FLD_v6p: // integers
|
|
case OP_STOREP_S_v6p:
|
|
case OP_STOREP_FN_v6p: // pointers
|
|
case OP_STOREP_I_v6p:
|
|
case OP_STOREP_P_v6p:
|
|
pointer = OPB(ptr);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
ptr->value = OPA(int);
|
|
break;
|
|
case OP_STOREP_V_v6p:
|
|
pointer = OPB(ptr);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_vector);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
VectorCopy (&OPA(float), &PR_PTR (float, ptr));
|
|
break;
|
|
case OP_STOREP_Q_v6p:
|
|
pointer = OPB(ptr);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
QuatCopy (&OPA(float), &PR_PTR (float, ptr));
|
|
break;
|
|
case OP_STOREP_D_v6p:
|
|
pointer = OPB(ptr);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_double);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
*(double *) ptr = OPA(double);
|
|
break;
|
|
|
|
case OP_ADDRESS_v6p:
|
|
if (pr_boundscheck) {
|
|
if (OPA(entity) >= pr->pr_edict_area_size)
|
|
PR_RunError (pr, "Progs attempted to address an out "
|
|
"of bounds edict");
|
|
if (OPA(entity) == 0 && pr->null_bad)
|
|
PR_RunError (pr, "assignment to world entity");
|
|
if (OPB(field) >= pr->progs->entityfields)
|
|
PR_RunError (pr, "Progs attempted to address an "
|
|
"invalid field in an edict");
|
|
}
|
|
fldofs = OPA(entity) + OPB(field);
|
|
OPC(ptr) = &pr->pr_edict_area[fldofs] - pr->pr_globals;
|
|
break;
|
|
case OP_ADDRESS_VOID_v6p:
|
|
case OP_ADDRESS_F_v6p:
|
|
case OP_ADDRESS_V_v6p:
|
|
case OP_ADDRESS_Q_v6p:
|
|
case OP_ADDRESS_S_v6p:
|
|
case OP_ADDRESS_ENT_v6p:
|
|
case OP_ADDRESS_FLD_v6p:
|
|
case OP_ADDRESS_FN_v6p:
|
|
case OP_ADDRESS_I_v6p:
|
|
case OP_ADDRESS_P_v6p:
|
|
case OP_ADDRESS_D_v6p:
|
|
OPC(int) = st->a;
|
|
break;
|
|
|
|
case OP_LOAD_F_v6p:
|
|
case OP_LOAD_FLD_v6p:
|
|
case OP_LOAD_ENT_v6p:
|
|
case OP_LOAD_S_v6p:
|
|
case OP_LOAD_FN_v6p:
|
|
case OP_LOAD_I_v6p:
|
|
case OP_LOAD_P_v6p:
|
|
if (pr_boundscheck) {
|
|
if (OPA(entity) >= pr->pr_edict_area_size)
|
|
PR_RunError (pr, "Progs attempted to read an out of "
|
|
"bounds edict number");
|
|
if (OPB(field) >= pr->progs->entityfields)
|
|
PR_RunError (pr, "Progs attempted to read an invalid "
|
|
"field in an edict");
|
|
}
|
|
fldofs = OPA(entity) + OPB(field);
|
|
OPC(int) = PR_PTR (int, &pr->pr_edict_area[fldofs]);
|
|
break;
|
|
case OP_LOAD_V_v6p:
|
|
if (pr_boundscheck) {
|
|
if (OPA(entity) >= pr->pr_edict_area_size)
|
|
PR_RunError (pr, "Progs attempted to read an out of "
|
|
"bounds edict number");
|
|
if (OPB(field) + 2 >= pr->progs->entityfields)
|
|
PR_RunError (pr, "Progs attempted to read an invalid "
|
|
"field in an edict");
|
|
}
|
|
fldofs = OPA(entity) + OPB(field);
|
|
memcpy (op_c, &pr->pr_edict_area[fldofs], 3 * sizeof (*op_c));
|
|
break;
|
|
case OP_LOAD_Q_v6p:
|
|
if (pr_boundscheck) {
|
|
if (OPA(entity) >= pr->pr_edict_area_size)
|
|
PR_RunError (pr, "Progs attempted to read an out of "
|
|
"bounds edict number");
|
|
if (OPB(field) + 3 >= pr->progs->entityfields)
|
|
PR_RunError (pr, "Progs attempted to read an invalid "
|
|
"field in an edict");
|
|
}
|
|
fldofs = OPA(entity) + OPB(field);
|
|
memcpy (op_c, &pr->pr_edict_area[fldofs], 4 * sizeof (*op_c));
|
|
break;
|
|
case OP_LOAD_D_v6p:
|
|
if (pr_boundscheck) {
|
|
if (OPA(entity) >= pr->pr_edict_area_size)
|
|
PR_RunError (pr, "Progs attempted to read an out of "
|
|
"bounds edict number");
|
|
if (OPB(field) + 1 >= pr->progs->entityfields)
|
|
PR_RunError (pr, "Progs attempted to read an invalid "
|
|
"field in an edict");
|
|
}
|
|
fldofs = OPA(entity) + OPB(field);
|
|
memcpy (op_c, &pr->pr_edict_area[fldofs], sizeof (double));
|
|
break;
|
|
|
|
case OP_LOADB_F_v6p:
|
|
case OP_LOADB_S_v6p:
|
|
case OP_LOADB_ENT_v6p:
|
|
case OP_LOADB_FLD_v6p:
|
|
case OP_LOADB_FN_v6p:
|
|
case OP_LOADB_I_v6p:
|
|
case OP_LOADB_P_v6p:
|
|
pointer = OPA(entity) + OPB(field);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
OPC(int) = ptr->value;
|
|
break;
|
|
case OP_LOADB_V_v6p:
|
|
pointer = OPA(entity) + OPB(field);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_vector);
|
|
}
|
|
VectorCopy (G_VECTOR (pr, pointer), &OPC(float));
|
|
break;
|
|
case OP_LOADB_Q_v6p:
|
|
pointer = OPA(entity) + OPB(field);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
QuatCopy (G_QUAT (pr, pointer), &OPC(float));
|
|
break;
|
|
case OP_LOADB_D_v6p:
|
|
pointer = OPA(entity) + OPB(field);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_double);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
OPC(double) = *(double *) ptr;
|
|
break;
|
|
|
|
case OP_LOADBI_F_v6p:
|
|
case OP_LOADBI_S_v6p:
|
|
case OP_LOADBI_ENT_v6p:
|
|
case OP_LOADBI_FLD_v6p:
|
|
case OP_LOADBI_FN_v6p:
|
|
case OP_LOADBI_I_v6p:
|
|
case OP_LOADBI_P_v6p:
|
|
pointer = OPA(ptr) + (short) st->b;
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
OPC(int) = ptr->value;
|
|
break;
|
|
case OP_LOADBI_V_v6p:
|
|
pointer = OPA(ptr) + (short) st->b;
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_vector);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
VectorCopy (G_VECTOR (pr, pointer), &OPC(float));
|
|
break;
|
|
case OP_LOADBI_Q_v6p:
|
|
pointer = OPA(ptr) + (short) st->b;
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
QuatCopy (G_QUAT (pr, pointer), &OPC(float));
|
|
break;
|
|
case OP_LOADBI_D_v6p:
|
|
pointer = OPA(ptr) + (short) st->b;
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
OPC(double) = *(double *) ptr;
|
|
break;
|
|
|
|
case OP_LEA_v6p:
|
|
pointer = OPA(ptr) + OPB(int);
|
|
OPC(ptr) = pointer;
|
|
break;
|
|
|
|
case OP_LEAI_v6p:
|
|
pointer = OPA(ptr) + (short) st->b;
|
|
OPC(ptr) = pointer;
|
|
break;
|
|
|
|
case OP_STOREB_F_v6p:
|
|
case OP_STOREB_S_v6p:
|
|
case OP_STOREB_ENT_v6p:
|
|
case OP_STOREB_FLD_v6p:
|
|
case OP_STOREB_FN_v6p:
|
|
case OP_STOREB_I_v6p:
|
|
case OP_STOREB_P_v6p:
|
|
pointer = OPB(ptr) + OPC(int);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
ptr->value = OPA(int);
|
|
break;
|
|
case OP_STOREB_V_v6p:
|
|
pointer = OPB(ptr) + OPC(int);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_vector);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
VectorCopy (&OPA(float), G_VECTOR (pr, pointer));
|
|
break;
|
|
case OP_STOREB_Q_v6p:
|
|
pointer = OPB(ptr) + OPC(int);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
QuatCopy (&OPA(float), G_QUAT (pr, pointer));
|
|
break;
|
|
case OP_STOREB_D_v6p:
|
|
pointer = OPB(ptr) + OPC(int);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
*(double *) ptr = OPA(double);
|
|
break;
|
|
|
|
case OP_STOREBI_F_v6p:
|
|
case OP_STOREBI_S_v6p:
|
|
case OP_STOREBI_ENT_v6p:
|
|
case OP_STOREBI_FLD_v6p:
|
|
case OP_STOREBI_FN_v6p:
|
|
case OP_STOREBI_I_v6p:
|
|
case OP_STOREBI_P_v6p:
|
|
pointer = OPB(ptr) + (short) st->c;
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
ptr->value = OPA(int);
|
|
break;
|
|
case OP_STOREBI_V_v6p:
|
|
pointer = OPB(ptr) + (short) st->c;
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_vector);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
VectorCopy (&OPA(float), G_VECTOR (pr, pointer));
|
|
break;
|
|
case OP_STOREBI_Q_v6p:
|
|
pointer = OPB(ptr) + (short) st->c;
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
QuatCopy (&OPA(float), G_QUAT (pr, pointer));
|
|
break;
|
|
case OP_STOREBI_D_v6p:
|
|
pointer = OPB(ptr) + (short) st->c;
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
ptr = pr->pr_globals + pointer;
|
|
*(double *) ptr = OPA(double);
|
|
break;
|
|
|
|
case OP_PUSH_F_v6p:
|
|
case OP_PUSH_FLD_v6p:
|
|
case OP_PUSH_ENT_v6p:
|
|
case OP_PUSH_S_v6p:
|
|
case OP_PUSH_FN_v6p:
|
|
case OP_PUSH_I_v6p:
|
|
case OP_PUSH_P_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack - 1;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 1);
|
|
}
|
|
stk->value = OPA(int);
|
|
*pr->globals.stack = stack;
|
|
}
|
|
break;
|
|
case OP_PUSH_V_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack - 3;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 3);
|
|
}
|
|
memcpy (stk, op_a, 3 * sizeof (*op_c));
|
|
*pr->globals.stack = stack;
|
|
}
|
|
break;
|
|
case OP_PUSH_Q_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack - 4;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 4);
|
|
}
|
|
memcpy (stk, op_a, 4 * sizeof (*op_c));
|
|
*pr->globals.stack = stack;
|
|
}
|
|
break;
|
|
|
|
case OP_PUSHB_F_v6p:
|
|
case OP_PUSHB_S_v6p:
|
|
case OP_PUSHB_ENT_v6p:
|
|
case OP_PUSHB_FLD_v6p:
|
|
case OP_PUSHB_FN_v6p:
|
|
case OP_PUSHB_I_v6p:
|
|
case OP_PUSHB_P_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack - 1;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
|
|
pointer = OPA(ptr) + OPB(int);
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 1);
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
|
|
stk->value = ptr->value;
|
|
*pr->globals.stack = stack;
|
|
}
|
|
break;
|
|
case OP_PUSHB_V_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack - 3;
|
|
|
|
pointer = OPA(ptr) + OPB(int);
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 3);
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
|
|
VectorCopy (G_VECTOR (pr, pointer), G_VECTOR (pr, stack));
|
|
*pr->globals.stack = stack;
|
|
}
|
|
break;
|
|
case OP_PUSHB_Q_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack - 4;
|
|
|
|
pointer = OPA(ptr) + OPB(int);
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 4);
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
|
|
QuatCopy (G_QUAT (pr, pointer), G_QUAT (pr, stack));
|
|
*pr->globals.stack = stack;
|
|
}
|
|
break;
|
|
|
|
case OP_PUSHBI_F_v6p:
|
|
case OP_PUSHBI_S_v6p:
|
|
case OP_PUSHBI_ENT_v6p:
|
|
case OP_PUSHBI_FLD_v6p:
|
|
case OP_PUSHBI_FN_v6p:
|
|
case OP_PUSHBI_I_v6p:
|
|
case OP_PUSHBI_P_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack - 1;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
|
|
pointer = OPA(ptr) + st->b;
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 1);
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
|
|
stk->value = ptr->value;
|
|
*pr->globals.stack = stack;
|
|
}
|
|
break;
|
|
case OP_PUSHBI_V_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack - 3;
|
|
|
|
pointer = OPA(ptr) + st->b;
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 3);
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
|
|
VectorCopy (G_VECTOR (pr, pointer), G_VECTOR (pr, stack));
|
|
*pr->globals.stack = stack;
|
|
}
|
|
break;
|
|
case OP_PUSHBI_Q_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack - 4;
|
|
|
|
pointer = OPA(ptr) + st->b;
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 4);
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
|
|
QuatCopy (G_QUAT (pr, pointer), G_QUAT (pr, stack));
|
|
*pr->globals.stack = stack;
|
|
}
|
|
break;
|
|
|
|
case OP_POP_F_v6p:
|
|
case OP_POP_FLD_v6p:
|
|
case OP_POP_ENT_v6p:
|
|
case OP_POP_S_v6p:
|
|
case OP_POP_FN_v6p:
|
|
case OP_POP_I_v6p:
|
|
case OP_POP_P_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 1);
|
|
}
|
|
OPA(int) = stk->value;
|
|
*pr->globals.stack = stack + 1;
|
|
}
|
|
break;
|
|
case OP_POP_V_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 3);
|
|
}
|
|
memcpy (op_a, stk, 3 * sizeof (*op_c));
|
|
*pr->globals.stack = stack + 3;
|
|
}
|
|
break;
|
|
case OP_POP_Q_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 4);
|
|
}
|
|
memcpy (op_a, stk, 4 * sizeof (*op_c));
|
|
*pr->globals.stack = stack + 4;
|
|
}
|
|
break;
|
|
|
|
case OP_POPB_F_v6p:
|
|
case OP_POPB_S_v6p:
|
|
case OP_POPB_ENT_v6p:
|
|
case OP_POPB_FLD_v6p:
|
|
case OP_POPB_FN_v6p:
|
|
case OP_POPB_I_v6p:
|
|
case OP_POPB_P_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
|
|
pointer = OPA(ptr) + OPB(int);
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 1);
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
|
|
ptr->value = stk->value;
|
|
*pr->globals.stack = stack + 1;
|
|
}
|
|
break;
|
|
case OP_POPB_V_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
|
|
pointer = OPA(ptr) + OPB(int);
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 3);
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
|
|
VectorCopy (G_VECTOR (pr, stack), G_VECTOR (pr, pointer));
|
|
*pr->globals.stack = stack + 3;
|
|
}
|
|
break;
|
|
case OP_POPB_Q_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
|
|
pointer = OPA(ptr) + OPB(int);
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 4);
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
|
|
QuatCopy (G_QUAT (pr, stack), G_QUAT (pr, pointer));
|
|
*pr->globals.stack = stack + 4;
|
|
}
|
|
break;
|
|
|
|
case OP_POPBI_F_v6p:
|
|
case OP_POPBI_S_v6p:
|
|
case OP_POPBI_ENT_v6p:
|
|
case OP_POPBI_FLD_v6p:
|
|
case OP_POPBI_FN_v6p:
|
|
case OP_POPBI_I_v6p:
|
|
case OP_POPBI_P_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
|
|
pointer = OPA(ptr) + st->b;
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 1);
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
|
|
ptr->value = stk->value;
|
|
*pr->globals.stack = stack + 1;
|
|
}
|
|
break;
|
|
case OP_POPBI_V_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
|
|
pointer = OPA(ptr) + st->b;
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 3);
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
|
|
VectorCopy (G_VECTOR (pr, stack), G_VECTOR (pr, pointer));
|
|
*pr->globals.stack = stack + 3;
|
|
}
|
|
break;
|
|
case OP_POPBI_Q_v6p:
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
|
|
pointer = OPA(ptr) + st->b;
|
|
ptr = pr->pr_globals + pointer;
|
|
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 4);
|
|
PR_BoundsCheck (pr, pointer, ev_quaternion);
|
|
}
|
|
|
|
QuatCopy (G_QUAT (pr, stack), G_QUAT (pr, pointer));
|
|
*pr->globals.stack = stack + 4;
|
|
}
|
|
break;
|
|
|
|
// ==================
|
|
case OP_IFNOT_v6p:
|
|
if (!OPA(int)) {
|
|
pr->pr_xstatement += (short)st->b - 1; // offset the st++
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_IF_v6p:
|
|
if (OPA(int)) {
|
|
pr->pr_xstatement += (short)st->b - 1; // offset the st++
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_IFBE_v6p:
|
|
if (OPA(int) <= 0) {
|
|
pr->pr_xstatement += (short)st->b - 1; // offset the st++
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_IFB_v6p:
|
|
if (OPA(int) < 0) {
|
|
pr->pr_xstatement += (short)st->b - 1; // offset the st++
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_IFAE_v6p:
|
|
if (OPA(int) >= 0) {
|
|
pr->pr_xstatement += (short)st->b - 1; // offset the st++
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_IFA_v6p:
|
|
if (OPA(int) > 0) {
|
|
pr->pr_xstatement += (short)st->b - 1; // offset the st++
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_GOTO_v6p:
|
|
pr->pr_xstatement += (short)st->a - 1; // offset the st++
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
break;
|
|
case OP_JUMP_v6p:
|
|
if (pr_boundscheck
|
|
&& (OPA(uint) >= pr->progs->statements.count)) {
|
|
PR_RunError (pr, "Invalid jump destination");
|
|
}
|
|
pr->pr_xstatement = OPA(uint) - 1; // offset the st++
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
break;
|
|
case OP_JUMPB_v6p:
|
|
pointer = st->a + OPB(int);
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheck (pr, pointer, ev_int);
|
|
}
|
|
pointer = G_POINTER (pr, pointer);
|
|
if (pr_boundscheck
|
|
&& (pointer >= pr->progs->statements.count)) {
|
|
PR_RunError (pr, "Invalid jump destination");
|
|
}
|
|
pr->pr_xstatement = pointer - 1; // offset the st++
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
break;
|
|
|
|
case OP_RCALL2_v6p:
|
|
case OP_RCALL3_v6p:
|
|
case OP_RCALL4_v6p:
|
|
case OP_RCALL5_v6p:
|
|
case OP_RCALL6_v6p:
|
|
case OP_RCALL7_v6p:
|
|
case OP_RCALL8_v6p:
|
|
pr->pr_params[1] = op_c;
|
|
goto op_rcall;
|
|
case OP_RCALL1_v6p:
|
|
pr->pr_params[1] = pr->pr_real_params[1];
|
|
op_rcall:
|
|
pr->pr_params[0] = op_b;
|
|
pr->pr_argc = st->op - OP_RCALL1_v6p + 1;
|
|
goto op_call;
|
|
case OP_CALL0_v6p:
|
|
case OP_CALL1_v6p:
|
|
case OP_CALL2_v6p:
|
|
case OP_CALL3_v6p:
|
|
case OP_CALL4_v6p:
|
|
case OP_CALL5_v6p:
|
|
case OP_CALL6_v6p:
|
|
case OP_CALL7_v6p:
|
|
case OP_CALL8_v6p:
|
|
PR_RESET_PARAMS (pr);
|
|
pr->pr_argc = st->op - OP_CALL0_v6p;
|
|
op_call:
|
|
pr->pr_xfunction->profile += profile - startprofile;
|
|
startprofile = profile;
|
|
PR_CallFunction (pr, OPA(func), pr->pr_return);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
break;
|
|
case OP_DONE_v6p:
|
|
case OP_RETURN_v6p:
|
|
if (!st->a)
|
|
memset (&R_INT (pr), 0,
|
|
pr->pr_param_size * sizeof (*op_a));
|
|
else if (&R_INT (pr) != &OPA(int))
|
|
memcpy (&R_INT (pr), op_a,
|
|
pr->pr_param_size * sizeof (*op_a));
|
|
// fallthrough
|
|
case OP_RETURN_V_v6p:
|
|
pr->pr_xfunction->profile += profile - startprofile;
|
|
startprofile = profile;
|
|
PR_LeaveFunction (pr, pr->pr_depth == exitdepth);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
if (pr->pr_depth == exitdepth) {
|
|
if (pr->pr_trace && pr->pr_depth <= pr->pr_trace_depth)
|
|
pr->pr_trace = false;
|
|
// all done
|
|
goto exit_program;
|
|
}
|
|
break;
|
|
case OP_STATE_v6p:
|
|
{
|
|
int self = *pr->globals.self;
|
|
int nextthink = pr->fields.nextthink + self;
|
|
int frame = pr->fields.frame + self;
|
|
int think = pr->fields.think + self;
|
|
float time = *pr->globals.ftime + 0.1;
|
|
PR_PTR (float, &pr->pr_edict_area[nextthink]) = time;
|
|
PR_PTR (float, &pr->pr_edict_area[frame]) = OPA(float);
|
|
PR_PTR (func, &pr->pr_edict_area[think]) = OPB(func);
|
|
}
|
|
break;
|
|
case OP_STATE_F_v6p:
|
|
{
|
|
int self = *pr->globals.self;
|
|
int nextthink = pr->fields.nextthink + self;
|
|
int frame = pr->fields.frame + self;
|
|
int think = pr->fields.think + self;
|
|
float time = *pr->globals.ftime + OPC(float);
|
|
PR_PTR (float, &pr->pr_edict_area[nextthink]) = time;
|
|
PR_PTR (float, &pr->pr_edict_area[frame]) = OPA(float);
|
|
PR_PTR (func, &pr->pr_edict_area[think]) = OPB(func);
|
|
}
|
|
break;
|
|
case OP_ADD_I_v6p:
|
|
OPC(int) = OPA(int) + OPB(int);
|
|
break;
|
|
case OP_SUB_I_v6p:
|
|
OPC(int) = OPA(int) - OPB(int);
|
|
break;
|
|
case OP_MUL_I_v6p:
|
|
OPC(int) = OPA(int) * OPB(int);
|
|
break;
|
|
case OP_DIV_I_v6p:
|
|
OPC(int) = OPA(int) / OPB(int);
|
|
break;
|
|
case OP_MOD_I_v6p:
|
|
{
|
|
// implement true modulo for integers:
|
|
// 5 mod 3 = 2
|
|
// -5 mod 3 = 1
|
|
// 5 mod -3 = -1
|
|
// -5 mod -3 = -2
|
|
int a = OPA(int);
|
|
int b = OPB(int);
|
|
int c = a % b;
|
|
// % is really remainder and so has the same sign rules
|
|
// as division: -5 % 3 = -2, so need to add b (3 here)
|
|
// if c's sign is incorrect, but only if c is non-zero
|
|
int mask = (a ^ b) >> 31;
|
|
mask &= ~(!!c + 0) + 1; // +0 to convert bool to int (gcc)
|
|
OPC(int) = c + (mask & b);
|
|
}
|
|
break;
|
|
case OP_REM_I_v6p:
|
|
OPC(int) = OPA(int) % OPB(int);
|
|
break;
|
|
case OP_MOD_D_v6p:
|
|
{
|
|
double a = OPA(double);
|
|
double b = OPB(double);
|
|
// floating point modulo is so much easier :P
|
|
OPC(double) = a - b * floor (a / b);
|
|
}
|
|
break;
|
|
case OP_REM_D_v6p:
|
|
{
|
|
double a = OPA(double);
|
|
double b = OPB(double);
|
|
OPC(double) = a - b * trunc (a / b);
|
|
}
|
|
break;
|
|
case OP_MOD_F_v6p:
|
|
{
|
|
float a = OPA(float);
|
|
float b = OPB(float);
|
|
OPC(float) = a - b * floorf (a / b);
|
|
}
|
|
break;
|
|
case OP_REM_F_v6p:
|
|
{
|
|
float a = OPA(float);
|
|
float b = OPB(float);
|
|
OPC(float) = a - b * truncf (a / b);
|
|
}
|
|
break;
|
|
case OP_CONV_IF_v6p:
|
|
OPC(float) = OPA(int);
|
|
break;
|
|
case OP_CONV_FI_v6p:
|
|
OPC(int) = OPA(float);
|
|
break;
|
|
case OP_BITAND_I_v6p:
|
|
OPC(int) = OPA(int) & OPB(int);
|
|
break;
|
|
case OP_BITOR_I_v6p:
|
|
OPC(int) = OPA(int) | OPB(int);
|
|
break;
|
|
case OP_BITXOR_I_v6p:
|
|
OPC(int) = OPA(int) ^ OPB(int);
|
|
break;
|
|
case OP_BITNOT_I_v6p:
|
|
OPC(int) = ~OPA(int);
|
|
break;
|
|
|
|
case OP_GE_I_v6p:
|
|
case OP_GE_P_v6p:
|
|
OPC(int) = OPA(int) >= OPB(int);
|
|
break;
|
|
case OP_GE_U_v6p:
|
|
OPC(int) = OPA(uint) >= OPB(uint);
|
|
break;
|
|
case OP_LE_I_v6p:
|
|
case OP_LE_P_v6p:
|
|
OPC(int) = OPA(int) <= OPB(int);
|
|
break;
|
|
case OP_LE_U_v6p:
|
|
OPC(int) = OPA(uint) <= OPB(uint);
|
|
break;
|
|
case OP_GT_I_v6p:
|
|
case OP_GT_P_v6p:
|
|
OPC(int) = OPA(int) > OPB(int);
|
|
break;
|
|
case OP_GT_U_v6p:
|
|
OPC(int) = OPA(uint) > OPB(uint);
|
|
break;
|
|
case OP_LT_I_v6p:
|
|
case OP_LT_P_v6p:
|
|
OPC(int) = OPA(int) < OPB(int);
|
|
break;
|
|
case OP_LT_U_v6p:
|
|
OPC(int) = OPA(uint) < OPB(uint);
|
|
break;
|
|
|
|
case OP_AND_I_v6p:
|
|
OPC(int) = OPA(int) && OPB(int);
|
|
break;
|
|
case OP_OR_I_v6p:
|
|
OPC(int) = OPA(int) || OPB(int);
|
|
break;
|
|
case OP_NOT_I_v6p:
|
|
case OP_NOT_P_v6p:
|
|
OPC(int) = !OPA(int);
|
|
break;
|
|
case OP_EQ_I_v6p:
|
|
case OP_EQ_P_v6p:
|
|
OPC(int) = OPA(int) == OPB(int);
|
|
break;
|
|
case OP_NE_I_v6p:
|
|
case OP_NE_P_v6p:
|
|
OPC(int) = OPA(int) != OPB(int);
|
|
break;
|
|
|
|
case OP_MOVEI_v6p:
|
|
memmove (op_c, op_a, st->b * 4);
|
|
break;
|
|
case OP_MOVEP_v6p:
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheckSize (pr, OPC(ptr), OPB(uint));
|
|
PR_BoundsCheckSize (pr, OPA(ptr), OPB(uint));
|
|
}
|
|
memmove (pr->pr_globals + OPC(ptr),
|
|
pr->pr_globals + OPA(ptr),
|
|
OPB(uint) * 4);
|
|
break;
|
|
case OP_MOVEPI_v6p:
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheckSize (pr, OPC(ptr), st->b);
|
|
PR_BoundsCheckSize (pr, OPA(ptr), st->b);
|
|
}
|
|
memmove (pr->pr_globals + OPC(ptr),
|
|
pr->pr_globals + OPA(ptr),
|
|
st->b * 4);
|
|
break;
|
|
case OP_MEMSETI_v6p:
|
|
pr_memset (op_c, OPA(ptr), st->b);
|
|
break;
|
|
case OP_MEMSETP_v6p:
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheckSize (pr, OPC(ptr), OPB(uint));
|
|
}
|
|
pr_memset (pr->pr_globals + OPC(ptr), OPA(int),
|
|
OPB(uint));
|
|
break;
|
|
case OP_MEMSETPI_v6p:
|
|
if (pr_boundscheck) {
|
|
PR_BoundsCheckSize (pr, OPC(ptr), st->b);
|
|
}
|
|
pr_memset (pr->pr_globals + OPC(ptr), OPA(int),
|
|
st->b);
|
|
break;
|
|
case OP_GE_D_v6p:
|
|
OPC(float) = OPA(double) >= OPB(double);
|
|
break;
|
|
case OP_LE_D_v6p:
|
|
OPC(float) = OPA(double) <= OPB(double);
|
|
break;
|
|
case OP_GT_D_v6p:
|
|
OPC(float) = OPA(double) > OPB(double);
|
|
break;
|
|
case OP_LT_D_v6p:
|
|
OPC(float) = OPA(double) < OPB(double);
|
|
break;
|
|
case OP_NOT_D_v6p:
|
|
OPC(int) = (op_a[0].value || (op_a[1].value & ~0x80000000u));
|
|
break;
|
|
case OP_EQ_D_v6p:
|
|
OPC(int) = OPA(double) == OPB(double);
|
|
break;
|
|
case OP_NE_D_v6p:
|
|
OPC(int) = OPA(double) != OPB(double);
|
|
break;
|
|
case OP_CONV_ID_v6p:
|
|
OPC(double) = OPA(int);
|
|
break;
|
|
case OP_CONV_DI_v6p:
|
|
OPC(int) = OPA(double);
|
|
break;
|
|
case OP_CONV_FD_v6p:
|
|
OPC(double) = OPA(float);
|
|
break;
|
|
case OP_CONV_DF_v6p:
|
|
OPC(float) = OPA(double);
|
|
break;
|
|
|
|
// LordHavoc: to be enabled when Progs version 7 (or whatever it will be numbered) is finalized
|
|
/*
|
|
case OP_BOUNDCHECK_v6p:
|
|
if (OPA(ptr) >= st->b) {
|
|
PR_RunError (pr, "Progs boundcheck failed at line number "
|
|
"%d, value is < 0 or >= %d", st->b, st->c);
|
|
}
|
|
break;
|
|
|
|
*/
|
|
default:
|
|
PR_RunError (pr, "Bad opcode %i", st->op & ~OP_BREAK);
|
|
}
|
|
if (pr->watch && pr->watch->value != old_val.value) {
|
|
if (!pr->wp_conditional
|
|
|| pr->watch->value == pr->wp_val.value) {
|
|
if (pr->debug_handler) {
|
|
pr->debug_handler (prd_watchpoint, 0, pr->debug_data);
|
|
} else {
|
|
PR_RunError (pr, "watchpoint hit: %d -> %d",
|
|
old_val.value, pr->watch->value);
|
|
}
|
|
}
|
|
old_val.value = pr->watch->value;
|
|
}
|
|
}
|
|
exit_program:
|
|
(void)0;//FIXME for clang
|
|
}
|
|
|
|
#define MM(type) (*((pr_##type##_t *) (mm)))
|
|
#define STK(type) (*((pr_##type##_t *) (stk)))
|
|
|
|
static pr_type_t *
|
|
pr_address_mode (progs_t *pr, const dstatement_t *st, int mm_ind)
|
|
{
|
|
pr_type_t *op_a = pr->pr_globals + st->a + PR_BASE (pr, st, A);
|
|
pr_type_t *op_b = pr->pr_globals + st->b + PR_BASE (pr, st, B);
|
|
pr_ptr_t mm_offs = 0;
|
|
|
|
switch (mm_ind) {
|
|
case 0:
|
|
// regular global access
|
|
mm_offs = op_a - pr->pr_globals;
|
|
break;
|
|
case 1:
|
|
// entity.field (equivalent to OP_LOAD_t_v6p)
|
|
(void)0;//FIXME for clang
|
|
pr_ptr_t edict_area = pr->pr_edict_area - pr->pr_globals;
|
|
mm_offs = edict_area + OPA(entity) + OPB(field);
|
|
break;
|
|
case 2:
|
|
// constant indexed pointer: *a + b (supports -ve offset)
|
|
mm_offs = OPA(ptr) + (short) st->b;
|
|
break;
|
|
case 3:
|
|
// variable indexed pointer: *a + *b (supports -ve offset)
|
|
mm_offs = OPA(ptr) + OPB(int);
|
|
break;
|
|
}
|
|
return pr->pr_globals + mm_offs;
|
|
}
|
|
|
|
static pr_type_t *
|
|
pr_call_mode (progs_t *pr, const dstatement_t *st, int mm_ind)
|
|
{
|
|
pr_type_t *op_a = pr->pr_globals + st->a + PR_BASE (pr, st, A);
|
|
pr_type_t *op_b = pr->pr_globals + st->b + PR_BASE (pr, st, B);
|
|
pr_ptr_t mm_offs = 0;
|
|
|
|
switch (mm_ind) {
|
|
case 1:
|
|
// regular global access
|
|
mm_offs = op_a - pr->pr_globals;
|
|
break;
|
|
case 2:
|
|
// constant indexed pointer: *a + b (supports -ve offset)
|
|
mm_offs = OPA(ptr) + (short) st->b;
|
|
break;
|
|
case 3:
|
|
// variable indexed pointer: *a + *b (supports -ve offset)
|
|
mm_offs = OPA(ptr) + OPB(int);
|
|
break;
|
|
case 4:
|
|
// entity.field (equivalent to OP_LOAD_t_v6p)
|
|
(void)0;//FIXME for clang
|
|
pr_ptr_t edict_area = pr->pr_edict_area - pr->pr_globals;
|
|
mm_offs = edict_area + OPA(entity) + OPB(field);
|
|
break;
|
|
}
|
|
return pr->pr_globals + mm_offs;
|
|
}
|
|
|
|
static pr_ptr_t __attribute__((pure))
|
|
pr_jump_mode (progs_t *pr, const dstatement_t *st, int jump_ind)
|
|
{
|
|
pr_type_t *op_a = pr->pr_globals + st->a + PR_BASE (pr, st, A);
|
|
pr_type_t *op_b = pr->pr_globals + st->b + PR_BASE (pr, st, B);
|
|
pr_ptr_t jump_offs = pr->pr_xstatement;
|
|
|
|
switch (jump_ind) {
|
|
case 0:
|
|
// instruction relative offset
|
|
jump_offs = jump_offs + (short) st->a;
|
|
break;
|
|
case 1:
|
|
// variable indexed array: a + *b (only +ve)
|
|
jump_offs = PR_PTR (uint, op_a + OPB(uint));
|
|
break;
|
|
case 2:
|
|
// constant indexed pointer: *a + b (supports -ve offset)
|
|
jump_offs = OPA(ptr) + (short) st->b;
|
|
break;
|
|
case 3:
|
|
// variable indexed pointer: *a + *b (supports -ve offset)
|
|
jump_offs = OPA(ptr) + OPB(int);
|
|
break;
|
|
}
|
|
if (pr_boundscheck && jump_offs >= pr->progs->statements.count) {
|
|
PR_RunError (pr, "out of bounds: %x", jump_offs);
|
|
}
|
|
return jump_offs - 1; // for st++
|
|
}
|
|
|
|
static pr_type_t *
|
|
pr_stack_push (progs_t *pr)
|
|
{
|
|
// keep the stack 16-byte aligned
|
|
pr_ptr_t stack = *pr->globals.stack - 4;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 4);
|
|
}
|
|
*pr->globals.stack = stack;
|
|
return stk;
|
|
}
|
|
|
|
static pr_type_t *
|
|
pr_stack_pop (progs_t *pr)
|
|
{
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
pr_type_t *stk = pr->pr_globals + stack;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack, 4);
|
|
}
|
|
// keep the stack 16-byte aligned
|
|
*pr->globals.stack = stack + 4;
|
|
return stk;
|
|
}
|
|
|
|
static void
|
|
pr_stack_adjust (progs_t *pr, int mode, int offset)
|
|
{
|
|
// keep the stack 16-byte aligned
|
|
if (mode || (offset & 3)) {
|
|
PR_RunError (pr, "invalid stack adjustment: %d, %d", mode, offset);
|
|
}
|
|
|
|
pr_ptr_t stack = *pr->globals.stack;
|
|
if (pr_boundscheck) {
|
|
check_stack_pointer (pr, stack + offset, 0);
|
|
}
|
|
*pr->globals.stack = stack + offset;
|
|
}
|
|
|
|
static void
|
|
pr_with (progs_t *pr, const dstatement_t *st)
|
|
{
|
|
pr_ptr_t edict_area = pr->pr_edict_area - pr->pr_globals;
|
|
pr_type_t *op_b = pr->pr_globals + PR_BASE (pr, st, B) + st->b;
|
|
pr_type_t *stk;
|
|
|
|
switch (st->a) {
|
|
// fixed offset
|
|
case 0:
|
|
// hard-0 base
|
|
pr->pr_bases[st->c & 3] = st->b;
|
|
return;
|
|
case 1:
|
|
// relative to current base (-ve offset)
|
|
pr->pr_bases[st->c & 3] = PR_BASE (pr, st, B) + (pr_short_t) st->b;
|
|
return;
|
|
case 2:
|
|
// relative to stack (-ve offset)
|
|
pr->pr_bases[st->c & 3] = *pr->globals.stack + (pr_short_t) st->b;
|
|
return;
|
|
case 3:
|
|
// relative to edict_area (only +ve)
|
|
pr->pr_bases[st->c & 3] = edict_area + st->b;
|
|
return;
|
|
|
|
case 4:
|
|
// hard-0 base
|
|
pr->pr_bases[st->c & 3] = G_POINTER (pr, st->b);;
|
|
return;
|
|
case 5:
|
|
pr->pr_bases[st->c & 3] = OPB(ptr);
|
|
return;
|
|
case 6:
|
|
// relative to stack (-ve offset)
|
|
pr->pr_bases[st->c & 3] = *pr->globals.stack + OPB(int);
|
|
return;
|
|
case 7:
|
|
// relative to edict_area (only +ve)
|
|
pr->pr_bases[st->c & 3] = edict_area + OPB(field);
|
|
return;
|
|
|
|
case 8:
|
|
// pushregs
|
|
stk = pr_stack_push (pr);
|
|
STK(uivec4) = pr->pr_bases;
|
|
return;
|
|
case 9:
|
|
// popregs
|
|
stk = pr_stack_pop (pr);
|
|
pr->pr_bases = STK(uivec4);
|
|
return;
|
|
case 10:
|
|
// reset
|
|
pr->pr_bases = (pr_uivec4_t) {};
|
|
return;
|
|
case 11:
|
|
// return pointer
|
|
pr->pr_bases[st->c & 3] = pr->pr_return - pr->pr_globals;
|
|
return;
|
|
}
|
|
PR_RunError (pr, "Invalid with index: %u", st->a);
|
|
}
|
|
|
|
static pr_ivec4_t
|
|
pr_swizzle_f (pr_ivec4_t vec, pr_ushort_t swiz)
|
|
{
|
|
goto do_swizzle;
|
|
#include "libs/gamecode/pr_swizzle32.cinc"
|
|
static const pr_ivec4_t neg[16] = {
|
|
{ 0, 0, 0, 0 },
|
|
{ 1<<31, 0, 0, 0 },
|
|
{ 0, 1<<31, 0, 0 },
|
|
{ 1<<31, 1<<31, 0, 0 },
|
|
{ 0, 0, 1<<31, 0 },
|
|
{ 1<<31, 0, 1<<31, 0 },
|
|
{ 0, 1<<31, 1<<31, 0 },
|
|
{ 1<<31, 1<<31, 1<<31, 0 },
|
|
{ 0, 0, 0, 1<<31 },
|
|
{ 1<<31, 0, 0, 1<<31 },
|
|
{ 0, 1<<31, 0, 1<<31 },
|
|
{ 1<<31, 1<<31, 0, 1<<31 },
|
|
{ 0, 0, 1<<31, 1<<31 },
|
|
{ 1<<31, 0, 1<<31, 1<<31 },
|
|
{ 0, 1<<31, 1<<31, 1<<31 },
|
|
{ 1<<31, 1<<31, 1<<31, 1<<31 },
|
|
};
|
|
static const pr_ivec4_t zero[16] = {
|
|
{ ~0, ~0, ~0, ~0 },
|
|
{ 0, ~0, ~0, ~0 },
|
|
{ ~0, 0, ~0, ~0 },
|
|
{ 0, 0, ~0, ~0 },
|
|
{ ~0, ~0, 0, ~0 },
|
|
{ 0, ~0, 0, ~0 },
|
|
{ ~0, 0, 0, ~0 },
|
|
{ 0, 0, 0, ~0 },
|
|
{ ~0, ~0, ~0, 0 },
|
|
{ 0, ~0, ~0, 0 },
|
|
{ ~0, 0, ~0, 0 },
|
|
{ 0, 0, ~0, 0 },
|
|
{ ~0, ~0, 0, 0 },
|
|
{ 0, ~0, 0, 0 },
|
|
{ ~0, 0, 0, 0 },
|
|
{ 0, 0, 0, 0 },
|
|
};
|
|
|
|
do_swizzle:
|
|
goto *swizzle_table[swiz & 0xff];
|
|
negate:
|
|
vec ^= neg[(swiz >> 8) & 0xf];
|
|
vec &= zero[(swiz >> 12) & 0xf];
|
|
return vec;
|
|
}
|
|
|
|
static pr_lvec4_t
|
|
#ifdef _WIN64
|
|
//force gcc to use registers for the parameters to avoid alignment issues
|
|
//on the stack (gcc bug as of 11.2)
|
|
__attribute__((sysv_abi))
|
|
#endif
|
|
pr_swizzle_d (pr_lvec4_t vec, pr_ushort_t swiz)
|
|
{
|
|
goto do_swizzle;
|
|
#include "libs/gamecode/pr_swizzle64.cinc"
|
|
#define L(x) UINT64_C(x)
|
|
static const pr_lvec4_t neg[16] = {
|
|
{ INT64_C(0), INT64_C(0), INT64_C(0), INT64_C(0) },
|
|
{ INT64_C(1)<<63, INT64_C(0), INT64_C(0), INT64_C(0) },
|
|
{ INT64_C(0), INT64_C(1)<<63, INT64_C(0), INT64_C(0) },
|
|
{ INT64_C(1)<<63, INT64_C(1)<<63, INT64_C(0), INT64_C(0) },
|
|
{ INT64_C(0), INT64_C(0), INT64_C(1)<<63, INT64_C(0) },
|
|
{ INT64_C(1)<<63, INT64_C(0), INT64_C(1)<<63, INT64_C(0) },
|
|
{ INT64_C(0), INT64_C(1)<<63, INT64_C(1)<<63, INT64_C(0) },
|
|
{ INT64_C(1)<<63, INT64_C(1)<<63, INT64_C(1)<<63, INT64_C(0) },
|
|
{ INT64_C(0), INT64_C(0), INT64_C(0), INT64_C(1)<<63 },
|
|
{ INT64_C(1)<<63, INT64_C(0), INT64_C(0), INT64_C(1)<<63 },
|
|
{ INT64_C(0), INT64_C(1)<<63, INT64_C(0), INT64_C(1)<<63 },
|
|
{ INT64_C(1)<<63, INT64_C(1)<<63, INT64_C(0), INT64_C(1)<<63 },
|
|
{ INT64_C(0), INT64_C(0), INT64_C(1)<<63, INT64_C(1)<<63 },
|
|
{ INT64_C(1)<<63, INT64_C(0), INT64_C(1)<<63, INT64_C(1)<<63 },
|
|
{ INT64_C(0), INT64_C(1)<<63, INT64_C(1)<<63, INT64_C(1)<<63 },
|
|
{ INT64_C(1)<<63, INT64_C(1)<<63, INT64_C(1)<<63, INT64_C(1)<<63 },
|
|
};
|
|
static const pr_lvec4_t zero[16] = {
|
|
{ ~INT64_C(0), ~INT64_C(0), ~INT64_C(0), ~INT64_C(0) },
|
|
{ INT64_C(0), ~INT64_C(0), ~INT64_C(0), ~INT64_C(0) },
|
|
{ ~INT64_C(0), INT64_C(0), ~INT64_C(0), ~INT64_C(0) },
|
|
{ INT64_C(0), INT64_C(0), ~INT64_C(0), ~INT64_C(0) },
|
|
{ ~INT64_C(0), ~INT64_C(0), INT64_C(0), ~INT64_C(0) },
|
|
{ INT64_C(0), ~INT64_C(0), INT64_C(0), ~INT64_C(0) },
|
|
{ ~INT64_C(0), INT64_C(0), INT64_C(0), ~INT64_C(0) },
|
|
{ INT64_C(0), INT64_C(0), INT64_C(0), ~INT64_C(0) },
|
|
{ ~INT64_C(0), ~INT64_C(0), ~INT64_C(0), INT64_C(0) },
|
|
{ INT64_C(0), ~INT64_C(0), ~INT64_C(0), INT64_C(0) },
|
|
{ ~INT64_C(0), INT64_C(0), ~INT64_C(0), INT64_C(0) },
|
|
{ INT64_C(0), INT64_C(0), ~INT64_C(0), INT64_C(0) },
|
|
{ ~INT64_C(0), ~INT64_C(0), INT64_C(0), INT64_C(0) },
|
|
{ INT64_C(0), ~INT64_C(0), INT64_C(0), INT64_C(0) },
|
|
{ ~INT64_C(0), INT64_C(0), INT64_C(0), INT64_C(0) },
|
|
{ INT64_C(0), INT64_C(0), INT64_C(0), INT64_C(0) },
|
|
};
|
|
|
|
do_swizzle:
|
|
goto *swizzle_table[swiz & 0xff];
|
|
negate:
|
|
vec ^= neg[(swiz >> 8) & 0xf];
|
|
vec &= zero[(swiz >> 12) & 0xf];
|
|
return vec;
|
|
}
|
|
|
|
static void
|
|
pr_exec_ruamoko (progs_t *pr, int exitdepth)
|
|
{
|
|
int profile, startprofile;
|
|
dstatement_t *st;
|
|
pr_type_t old_val = {0};
|
|
|
|
// make a stack frame
|
|
startprofile = profile = 0;
|
|
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
|
|
if (pr->watch) {
|
|
old_val = *pr->watch;
|
|
}
|
|
|
|
while (1) {
|
|
st++;
|
|
++pr->pr_xstatement;
|
|
if (pr->pr_xstatement != st - pr->pr_statements)
|
|
PR_RunError (pr, "internal error");
|
|
if (++profile > 1000000 && !pr->no_exec_limit) {
|
|
PR_RunError (pr, "runaway loop error");
|
|
}
|
|
|
|
if (pr->pr_trace) {
|
|
if (pr->debug_handler) {
|
|
pr->debug_handler (prd_trace, 0, pr->debug_data);
|
|
} else {
|
|
PR_PrintStatement (pr, st, 1);
|
|
}
|
|
}
|
|
|
|
if (st->op & OP_BREAK) {
|
|
if (pr->debug_handler) {
|
|
pr->debug_handler (prd_breakpoint, 0, pr->debug_data);
|
|
} else {
|
|
PR_RunError (pr, "breakpoint hit");
|
|
}
|
|
}
|
|
|
|
pr_ptr_t st_a = st->a + PR_BASE (pr, st, A);
|
|
pr_ptr_t st_b = st->b + PR_BASE (pr, st, B);
|
|
pr_ptr_t st_c = st->c + PR_BASE (pr, st, C);
|
|
|
|
|
|
pr_type_t *op_a = pr->pr_globals + st_a;
|
|
pr_type_t *op_b = pr->pr_globals + st_b;
|
|
pr_type_t *op_c = pr->pr_globals + st_c;
|
|
|
|
pr_type_t *stk;
|
|
pr_type_t *mm;
|
|
pr_func_t function;
|
|
pr_opcode_e st_op = st->op & OP_MASK;
|
|
switch (st_op) {
|
|
// 0 0000
|
|
case OP_NOP:
|
|
break;
|
|
case OP_ADJSTK:
|
|
pr_stack_adjust (pr, st->a, (short) st->b);
|
|
break;
|
|
case OP_LDCONST:
|
|
PR_RunError (pr, "OP_LDCONST not implemented");
|
|
break;
|
|
case OP_LOAD_B_1:
|
|
case OP_LOAD_C_1:
|
|
case OP_LOAD_D_1:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_LOAD_B_1 + 4) >> 2);
|
|
OPC(int) = MM(int);
|
|
break;
|
|
case OP_LOAD_B_2:
|
|
case OP_LOAD_C_2:
|
|
case OP_LOAD_D_2:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_LOAD_B_2 + 4) >> 2);
|
|
OPC(ivec2) = MM(ivec2);
|
|
break;
|
|
case OP_LOAD_B_3:
|
|
case OP_LOAD_C_3:
|
|
case OP_LOAD_D_3:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_LOAD_B_3 + 4) >> 2);
|
|
VectorCopy (&MM(int), &OPC(int));
|
|
break;
|
|
case OP_LOAD_B_4:
|
|
case OP_LOAD_C_4:
|
|
case OP_LOAD_D_4:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_LOAD_B_4 + 4) >> 2);
|
|
OPC(ivec4) = MM(ivec4);
|
|
break;
|
|
// 0 0001
|
|
case OP_STORE_A_1:
|
|
case OP_STORE_B_1:
|
|
case OP_STORE_C_1:
|
|
case OP_STORE_D_1:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_STORE_A_1) >> 2);
|
|
MM(int) = OPC(int);
|
|
break;
|
|
case OP_STORE_A_2:
|
|
case OP_STORE_B_2:
|
|
case OP_STORE_C_2:
|
|
case OP_STORE_D_2:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_STORE_A_2) >> 2);
|
|
MM(ivec2) = OPC(ivec2);
|
|
break;
|
|
case OP_STORE_A_3:
|
|
case OP_STORE_B_3:
|
|
case OP_STORE_C_3:
|
|
case OP_STORE_D_3:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_STORE_A_3) >> 2);
|
|
VectorCopy (&OPC(int), &MM(int));
|
|
break;
|
|
case OP_STORE_A_4:
|
|
case OP_STORE_B_4:
|
|
case OP_STORE_C_4:
|
|
case OP_STORE_D_4:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_STORE_A_4) >> 2);
|
|
MM(ivec4) = OPC(ivec4);
|
|
break;
|
|
// 0 0010
|
|
case OP_PUSH_A_1:
|
|
case OP_PUSH_B_1:
|
|
case OP_PUSH_C_1:
|
|
case OP_PUSH_D_1:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_PUSH_A_1) >> 2);
|
|
stk = pr_stack_push (pr);
|
|
STK(int) = MM(int);
|
|
break;
|
|
case OP_PUSH_A_2:
|
|
case OP_PUSH_B_2:
|
|
case OP_PUSH_C_2:
|
|
case OP_PUSH_D_2:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_PUSH_A_2) >> 2);
|
|
stk = pr_stack_push (pr);
|
|
STK(ivec2) = MM(ivec2);
|
|
break;
|
|
case OP_PUSH_A_3:
|
|
case OP_PUSH_B_3:
|
|
case OP_PUSH_C_3:
|
|
case OP_PUSH_D_3:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_PUSH_A_3) >> 2);
|
|
stk = pr_stack_push (pr);
|
|
VectorCopy (&MM(int), &STK(int));
|
|
break;
|
|
case OP_PUSH_A_4:
|
|
case OP_PUSH_B_4:
|
|
case OP_PUSH_C_4:
|
|
case OP_PUSH_D_4:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_PUSH_A_4) >> 2);
|
|
stk = pr_stack_push (pr);
|
|
STK(ivec4) = MM(ivec4);
|
|
break;
|
|
// 0 0011
|
|
case OP_POP_A_1:
|
|
case OP_POP_B_1:
|
|
case OP_POP_C_1:
|
|
case OP_POP_D_1:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_POP_A_1) >> 2);
|
|
stk = pr_stack_pop (pr);
|
|
MM(int) = STK(int);
|
|
break;
|
|
case OP_POP_A_2:
|
|
case OP_POP_B_2:
|
|
case OP_POP_C_2:
|
|
case OP_POP_D_2:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_POP_A_2) >> 2);
|
|
stk = pr_stack_pop (pr);
|
|
MM(ivec2) = STK(ivec2);
|
|
break;
|
|
case OP_POP_A_3:
|
|
case OP_POP_B_3:
|
|
case OP_POP_C_3:
|
|
case OP_POP_D_3:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_POP_A_3) >> 2);
|
|
stk = pr_stack_pop (pr);
|
|
VectorCopy (&STK(int), &MM(int));
|
|
break;
|
|
case OP_POP_A_4:
|
|
case OP_POP_B_4:
|
|
case OP_POP_C_4:
|
|
case OP_POP_D_4:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_POP_A_4) >> 2);
|
|
stk = pr_stack_pop (pr);
|
|
MM(ivec4) = STK(ivec4);
|
|
break;
|
|
// 0 0100
|
|
// spare
|
|
// 0 0101
|
|
// spare
|
|
// 0 0110
|
|
// spare
|
|
// 0 0111
|
|
// spare
|
|
|
|
#define OP_cmp_1(OP, T, rt, cmp, ct) \
|
|
case OP_##OP##_##T##_1: \
|
|
OPC(rt) = -(OPA(ct) cmp OPB(ct)); \
|
|
break
|
|
#define OP_cmp_2(OP, T, rt, cmp, ct) \
|
|
case OP_##OP##_##T##_2: \
|
|
OPC(rt) = (OPA(ct) cmp OPB(ct)); \
|
|
break
|
|
#define OP_cmp_3(OP, T, rt, cmp, ct) \
|
|
case OP_##OP##_##T##_3: \
|
|
VectorCompCompare (&OPC(rt), -, &OPA(ct), cmp, &OPB(ct)); \
|
|
break;
|
|
#define OP_cmp_4(OP, T, rt, cmp, ct) \
|
|
case OP_##OP##_##T##_4: \
|
|
OPC(rt) = (OPA(ct) cmp OPB(ct)); \
|
|
break
|
|
#define OP_cmp_T(OP, T, rt1, rt2, rt4, cmp, ct1, ct2, ct4) \
|
|
OP_cmp_1 (OP, T, rt1, cmp, ct1); \
|
|
OP_cmp_2 (OP, T, rt2, cmp, ct2); \
|
|
OP_cmp_3 (OP, T, rt1, cmp, ct1); \
|
|
OP_cmp_4 (OP, T, rt4, cmp, ct4)
|
|
#define OP_cmp(OP, cmp) \
|
|
OP_cmp_T (OP, I, int, ivec2, ivec4, cmp, int, ivec2, ivec4); \
|
|
OP_cmp_T (OP, F, int, ivec2, ivec4, cmp, float, vec2, vec4); \
|
|
OP_cmp_T (OP, L, long, lvec2, lvec4, cmp, long, lvec2, lvec4); \
|
|
OP_cmp_T (OP, D, long, lvec2, lvec4, cmp, double, dvec2, dvec4)
|
|
|
|
// 0 1000
|
|
OP_cmp(EQ, ==);
|
|
// 0 1001
|
|
OP_cmp(LT, <);
|
|
// 0 1010
|
|
OP_cmp(GT, >);
|
|
|
|
#define OP_op_1(OP, T, t, op) \
|
|
case OP_##OP##_##T##_1: \
|
|
OPC(t) = (OPA(t) op OPB(t)); \
|
|
break
|
|
#define OP_op_2(OP, T, t, op) \
|
|
case OP_##OP##_##T##_2: \
|
|
OPC(t) = (OPA(t) op OPB(t)); \
|
|
break
|
|
#define OP_op_3(OP, T, t, op) \
|
|
case OP_##OP##_##T##_3: \
|
|
VectorCompOp (&OPC(t), &OPA(t), op, &OPB(t)); \
|
|
break;
|
|
#define OP_op_4(OP, T, t, op) \
|
|
case OP_##OP##_##T##_4: \
|
|
OPC(t) = (OPA(t) op OPB(t)); \
|
|
break
|
|
#define OP_op_T(OP, T, t1, t2, t4, op) \
|
|
OP_op_1 (OP, T, t1, op); \
|
|
OP_op_2 (OP, T, t2, op); \
|
|
OP_op_3 (OP, T, t1, op); \
|
|
OP_op_4 (OP, T, t4, op)
|
|
// 0 1011
|
|
OP_op_T (DIV, u, uint, uivec2, uivec4, /);
|
|
OP_op_T (DIV, U, ulong, ulvec2, ulvec4, /);
|
|
OP_op_T (REM, u, uint, uivec2, uivec4, %);
|
|
OP_op_T (REM, U, ulong, ulvec2, ulvec4, %);
|
|
// 0 1100
|
|
OP_cmp(NE, !=);
|
|
// 0 1101
|
|
OP_cmp(GE, >=);
|
|
// 0 1110
|
|
OP_cmp(LE, <=);
|
|
// 0 1111
|
|
case OP_LOAD64_B_3:
|
|
case OP_LOAD64_C_3:
|
|
case OP_LOAD64_D_3:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_LOAD64_B_3 + 1));
|
|
VectorCopy (&MM(long), &OPC(long));
|
|
break;
|
|
case OP_STORE64_A_3:
|
|
case OP_STORE64_B_3:
|
|
case OP_STORE64_C_3:
|
|
case OP_STORE64_D_3:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_STORE64_A_3));
|
|
VectorCopy (&OPC(long), &MM(long));
|
|
break;
|
|
case OP_LOAD64_B_4:
|
|
case OP_LOAD64_C_4:
|
|
case OP_LOAD64_D_4:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_LOAD64_B_4 + 1));
|
|
OPC(lvec4) = MM(lvec4);
|
|
break;
|
|
case OP_STORE64_A_4:
|
|
case OP_STORE64_B_4:
|
|
case OP_STORE64_C_4:
|
|
case OP_STORE64_D_4:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_STORE64_A_4));
|
|
MM(lvec4) = OPC(lvec4);
|
|
break;
|
|
// spare
|
|
|
|
#define OP_op(OP, op) \
|
|
OP_op_T (OP, I, int, ivec2, ivec4, op); \
|
|
OP_op_T (OP, F, float, vec2, vec4, op); \
|
|
OP_op_T (OP, L, long, lvec2, lvec4, op); \
|
|
OP_op_T (OP, D, double, dvec2, dvec4, op)
|
|
#define OP_uop_1(OP, T, t, op) \
|
|
case OP_##OP##_##T##_1: \
|
|
OPC(t) = op (OPA(t)); \
|
|
break
|
|
#define OP_uop_2(OP, T, t, op) \
|
|
case OP_##OP##_##T##_2: \
|
|
OPC(t) = op (OPA(t)); \
|
|
break
|
|
#define OP_uop_3(OP, T, t, op) \
|
|
case OP_##OP##_##T##_3: \
|
|
VectorCompUop (&OPC(t), op, &OPA(t)); \
|
|
break;
|
|
#define OP_uop_4(OP, T, t, op) \
|
|
case OP_##OP##_##T##_4: \
|
|
OPC(t) = op (OPA(t)); \
|
|
break
|
|
#define OP_uop_T(OP, T, t1, t2, t4, op) \
|
|
OP_uop_1 (OP, T, t1, op); \
|
|
OP_uop_2 (OP, T, t2, op); \
|
|
OP_uop_3 (OP, T, t1, op); \
|
|
OP_uop_4 (OP, T, t4, op)
|
|
|
|
// 1 0000
|
|
OP_op(MUL, *);
|
|
// 1 0001
|
|
OP_op(DIV, /);
|
|
|
|
// implement remainder (c %) for integers:
|
|
// 5 rem 3 = 2
|
|
// -5 rem 3 = -2
|
|
// 5 rem -3 = 2
|
|
// -5 rem -3 = -2
|
|
#define OP_store(d, s) *(d) = s
|
|
#define OP_remmod_T(OP, T, n, t, l, f, s) \
|
|
case OP_##OP##_##T##_##n: \
|
|
{ \
|
|
__auto_type a = l (&OPA(t)); \
|
|
__auto_type b = l (&OPB(t)); \
|
|
s (&OPC(t), a - b * f(a / b)); \
|
|
} \
|
|
break
|
|
#define OP_rem_T(T, n, t, l, f, s) \
|
|
OP_remmod_T(REM, T, n, t, l, f, s)
|
|
|
|
// 1 0010
|
|
OP_op_T (REM, I, int, ivec2, ivec4, %);
|
|
OP_rem_T (F, 1, float, *, truncf, OP_store);
|
|
OP_rem_T (F, 2, vec2, *, vtrunc2f, OP_store);
|
|
OP_rem_T (F, 3, float, loadvec3f, vtrunc4f, storevec3f);
|
|
OP_rem_T (F, 4, vec4, *, vtrunc4f, OP_store);
|
|
OP_op_T (REM, L, long, lvec2, lvec4, %);
|
|
OP_rem_T (D, 1, double, *, trunc, OP_store);
|
|
OP_rem_T (D, 2, dvec2, *, vtrunc2d, OP_store);
|
|
OP_rem_T (D, 3, double, loadvec3d, vtrunc4d, storevec3d);
|
|
OP_rem_T (D, 4, dvec4, *, vtrunc4d, OP_store);
|
|
|
|
// implement true modulo (python %) for integers:
|
|
// 5 mod 3 = 2
|
|
// -5 mod 3 = 1
|
|
// 5 mod -3 = -1
|
|
// -5 mod -3 = -2
|
|
#define OP_mod_Ti(T, n, t, l, m, s) \
|
|
case OP_MOD_##T##_##n: \
|
|
{ \
|
|
__auto_type a = l(&OPA(t)); \
|
|
__auto_type b = l(&OPB(t)); \
|
|
__auto_type c = a % b; \
|
|
/* % is really remainder and so has the same sign rules */\
|
|
/* as division: -5 % 3 = -2, so need to add b (3 here) */\
|
|
/* if c's sign is incorrect, but only if c is non-zero */\
|
|
__auto_type mask = m((a ^ b) < 0); \
|
|
mask &= m(c != 0); \
|
|
s(&OPC(t), c + (mask & b)); \
|
|
} \
|
|
break
|
|
// floating point modulo is so much easier :P (just use floor instead of trunc)
|
|
#define OP_mod_Tf(T, n, t, l, f, s) \
|
|
OP_remmod_T(MOD, T, n, t, l, f, s)
|
|
|
|
// 1 0011
|
|
OP_mod_Ti (I, 1, int, *, -, OP_store);
|
|
OP_mod_Ti (I, 2, ivec2, *, +, OP_store);
|
|
OP_mod_Ti (I, 3, int, loadvec3i1, +, storevec3i);
|
|
OP_mod_Ti (I, 4, ivec4, *, +, OP_store);
|
|
OP_mod_Tf (F, 1, float, *, floorf, OP_store);
|
|
OP_mod_Tf (F, 2, vec2, *, vfloor2f, OP_store);
|
|
OP_mod_Tf (F, 3, float, loadvec3f, vfloor4f, storevec3f);
|
|
OP_mod_Tf (F, 4, vec4, *, vfloor4f, OP_store);
|
|
OP_mod_Ti (L, 1, long, *, -, OP_store);
|
|
OP_mod_Ti (L, 2, lvec2, *, +, OP_store);
|
|
OP_mod_Ti (L, 3, long, loadvec3l1, +, storevec3l);
|
|
OP_mod_Ti (L, 4, lvec4, *, +, OP_store);
|
|
OP_mod_Tf (D, 1, double, *, floor, OP_store);
|
|
OP_mod_Tf (D, 2, dvec2, *, vfloor2d, OP_store);
|
|
OP_mod_Tf (D, 3, double, loadvec3d, vfloor4d, storevec3d);
|
|
OP_mod_Tf (D, 4, dvec4, *, vfloor4d, OP_store);
|
|
|
|
// 1 0100
|
|
OP_op(ADD, +);
|
|
// 1 0101
|
|
OP_op(SUB, -);
|
|
// 1 0110
|
|
OP_op_T (SHL, I, int, ivec2, ivec4, <<);
|
|
OP_op_T (SHL, L, long, lvec2, lvec4, <<);
|
|
case OP_EQ_S:
|
|
case OP_LT_S:
|
|
case OP_GT_S:
|
|
case OP_CMP_S:
|
|
case OP_GE_S:
|
|
case OP_LE_S:
|
|
{
|
|
int cmp = strcmp (PR_GetString (pr, OPA(string)),
|
|
PR_GetString (pr, OPB(string)));
|
|
switch (st_op) {
|
|
case OP_EQ_S: cmp = -(cmp == 0); break;
|
|
case OP_LT_S: cmp = -(cmp < 0); break;
|
|
case OP_GT_S: cmp = -(cmp > 0); break;
|
|
case OP_GE_S: cmp = -(cmp >= 0); break;
|
|
case OP_LE_S: cmp = -(cmp <= 0); break;
|
|
case OP_CMP_S: break;
|
|
default: break;
|
|
}
|
|
OPC(int) = cmp;
|
|
}
|
|
break;
|
|
case OP_ADD_S:
|
|
OPC(string) = PR_CatStrings(pr, PR_GetString (pr, OPA(string)),
|
|
PR_GetString (pr, OPB(string)));
|
|
break;
|
|
case OP_NOT_S:
|
|
OPC(int) = -(!OPA(string) || !*PR_GetString (pr, OPA(string)));
|
|
break;
|
|
// 1 0111
|
|
OP_op_T (ASR, I, int, ivec2, ivec4, >>);
|
|
OP_op_T (SHR, u, uint, uivec2, uivec4, >>);
|
|
OP_op_T (ASR, L, long, lvec2, lvec4, >>);
|
|
OP_op_T (SHR, U, ulong, ulvec2, ulvec4, >>);
|
|
// 1 1000
|
|
OP_op_T (BITAND, I, int, ivec2, ivec4, &);
|
|
OP_op_T (BITOR, I, int, ivec2, ivec4, |);
|
|
OP_op_T (BITXOR, I, int, ivec2, ivec4, ^);
|
|
OP_uop_T (BITNOT, I, int, ivec2, ivec4, ~);
|
|
// 1 1001
|
|
OP_cmp_T (LT, u, int, ivec2, ivec4, <, uint, uivec2, uivec4);
|
|
case OP_JUMP_A:
|
|
case OP_JUMP_B:
|
|
case OP_JUMP_C:
|
|
case OP_JUMP_D:
|
|
pr->pr_xstatement = pr_jump_mode (pr, st, st_op - OP_JUMP_A);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
break;
|
|
OP_cmp_T (LT, U, long, lvec2, lvec4, <, ulong, ulvec2, ulvec4);
|
|
case OP_RETURN:
|
|
(void)0;//FIXME for clang
|
|
int ret_size = (st->c & 0x1f) + 1; // up to 32 words
|
|
if (st->c != 0xffff) {
|
|
mm = pr_address_mode (pr, st, st->c >> 5);
|
|
memcpy (&R_INT (pr), mm, ret_size * sizeof (*op_a));
|
|
}
|
|
pr->pr_xfunction->profile += profile - startprofile;
|
|
startprofile = profile;
|
|
PR_LeaveFunction (pr, pr->pr_depth == exitdepth);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
if (pr->pr_depth== exitdepth) {
|
|
if (pr->pr_trace && pr->pr_depth <= pr->pr_trace_depth) {
|
|
pr->pr_trace = false;
|
|
}
|
|
goto exit_program;
|
|
}
|
|
break;
|
|
case OP_CALL_B:
|
|
case OP_CALL_C:
|
|
case OP_CALL_D:
|
|
mm = pr_call_mode (pr, st, st_op - OP_CALL_B + 1);
|
|
function = PR_PTR (func, mm);
|
|
pr->pr_argc = 0;
|
|
// op_c specifies the location for the return value if any
|
|
pr->pr_xfunction->profile += profile - startprofile;
|
|
startprofile = profile;
|
|
PR_CallFunction (pr, function, op_c);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
break;
|
|
// 1 1010
|
|
OP_cmp_T (GT, u, int, ivec2, ivec4, >, uint, uivec2, uivec4);
|
|
case OP_SWIZZLE_F:
|
|
OPC(ivec4) = pr_swizzle_f (OPA(ivec4), st->b);
|
|
break;
|
|
case OP_SCALE_F_2:
|
|
OPC(vec2) = OPA(vec2) * OPB(float);
|
|
break;
|
|
case OP_SCALE_F_3:
|
|
VectorScale (&OPA(float), OPB(float), &OPC(float));
|
|
break;
|
|
case OP_SCALE_F_4:
|
|
OPC(vec4) = OPA(vec4) * OPB(float);
|
|
break;
|
|
OP_cmp_T (GT, U, long, lvec2, lvec4, >, ulong, ulvec2, ulvec4);
|
|
case OP_SWIZZLE_D:
|
|
OPC(lvec4) = pr_swizzle_d (OPA(lvec4), st->b);
|
|
break;
|
|
case OP_SCALE_D_2:
|
|
OPC(dvec2) = OPA(dvec2) * OPB(double);
|
|
break;
|
|
case OP_SCALE_D_3:
|
|
VectorScale (&OPA(double), OPB(double), &OPC(double));
|
|
break;
|
|
case OP_SCALE_D_4:
|
|
OPC(dvec4) = OPA(dvec4) * OPB(double);
|
|
break;
|
|
// 1 1011
|
|
case OP_CROSS_F:
|
|
{
|
|
pr_vec4_t a = loadvec3f (&OPA(float));
|
|
pr_vec4_t b = loadvec3f (&OPB(float));
|
|
pr_vec4_t c = crossf (a, b);
|
|
storevec3f (&OPC(float), c);
|
|
}
|
|
break;
|
|
case OP_CDOT_F:
|
|
OPC(vec2) = dot2f (OPA(vec2), OPB(vec2));
|
|
break;
|
|
case OP_VDOT_F:
|
|
{
|
|
vec_t d = DotProduct (&OPA(float),
|
|
&OPB(float));
|
|
VectorSet (d, d, d, &OPC(float));
|
|
}
|
|
break;
|
|
case OP_QDOT_F:
|
|
OPC(vec4) = dotf (OPA(vec4), OPB(vec4));
|
|
break;
|
|
case OP_CMUL_F:
|
|
OPC(vec2) = cmulf (OPA(vec2), OPB(vec2));
|
|
break;
|
|
case OP_QVMUL_F:
|
|
{
|
|
pr_vec4_t v = loadvec3f (&OPB(float));
|
|
v = qvmulf (OPA(vec4), v);
|
|
storevec3f (&OPC(float), v);
|
|
}
|
|
break;
|
|
case OP_VQMUL_F:
|
|
{
|
|
pr_vec4_t v = loadvec3f (&OPA(float));
|
|
v = vqmulf (v, OPB(vec4));
|
|
storevec3f (&OPC(float), v);
|
|
}
|
|
break;
|
|
case OP_QMUL_F:
|
|
OPC(vec4) = qmulf (OPA(vec4), OPB(vec4));
|
|
break;
|
|
case OP_CROSS_D:
|
|
{
|
|
pr_dvec4_t a = loadvec3d (&OPA(double));
|
|
pr_dvec4_t b = loadvec3d (&OPB(double));
|
|
pr_dvec4_t c = crossd (a, b);
|
|
storevec3d (&OPC(double), c);
|
|
}
|
|
break;
|
|
case OP_CDOT_D:
|
|
OPC(dvec2) = dot2d (OPA(dvec2), OPB(dvec2));
|
|
break;
|
|
case OP_VDOT_D:
|
|
{
|
|
double d = DotProduct (&OPA(double),
|
|
&OPB(double));
|
|
VectorSet (d, d, d, &OPC(double));
|
|
}
|
|
break;
|
|
case OP_QDOT_D:
|
|
OPC(dvec4) = dotd (OPA(dvec4), OPB(dvec4));
|
|
break;
|
|
case OP_CMUL_D:
|
|
OPC(dvec2) = cmuld (OPA(dvec2), OPB(dvec2));
|
|
break;
|
|
case OP_QVMUL_D:
|
|
{
|
|
pr_dvec4_t v = loadvec3d (&OPB(double));
|
|
v = qvmuld (OPA(dvec4), v);
|
|
storevec3d (&OPC(double), v);
|
|
}
|
|
break;
|
|
case OP_VQMUL_D:
|
|
{
|
|
pr_dvec4_t v = loadvec3d (&OPA(double));
|
|
v = vqmuld (v, OPB(dvec4));
|
|
storevec3d (&OPC(double), v);
|
|
}
|
|
break;
|
|
case OP_QMUL_D:
|
|
OPC(dvec4) = qmuld (OPA(dvec4), OPB(dvec4));
|
|
break;
|
|
// 1 1100
|
|
OP_op_T (BITAND, L, long, lvec2, lvec4, &);
|
|
OP_op_T (BITOR, L, long, lvec2, lvec4, |);
|
|
OP_op_T (BITXOR, L, long, lvec2, lvec4, ^);
|
|
OP_uop_T (BITNOT, L, long, lvec2, lvec4, ~);
|
|
// 1 1101
|
|
OP_cmp_T (GE, u, int, ivec2, ivec4, >=, uint, uivec2, uivec4);
|
|
case OP_MOVE_I:
|
|
memmove (op_c, op_a, st->b * sizeof (pr_type_t));
|
|
break;
|
|
case OP_MOVE_P:
|
|
memmove (pr->pr_globals + OPC(ptr), pr->pr_globals + OPA(ptr),
|
|
OPB(uint) * sizeof (pr_type_t));
|
|
break;
|
|
case OP_MOVE_PI:
|
|
memmove (pr->pr_globals + OPC(ptr), pr->pr_globals + OPA(ptr),
|
|
st->b * sizeof (pr_type_t));
|
|
break;
|
|
case OP_STATE_ft:
|
|
{
|
|
int self = *pr->globals.self;
|
|
int nextthink = pr->fields.nextthink + self;
|
|
int frame = pr->fields.frame + self;
|
|
int think = pr->fields.think + self;
|
|
float time = *pr->globals.ftime + 0.1;
|
|
PR_PTR (float, &pr->pr_edict_area[nextthink]) = time;
|
|
PR_PTR (float, &pr->pr_edict_area[frame]) = OPA(float);
|
|
PR_PTR (func, &pr->pr_edict_area[think]) = OPB(func);
|
|
}
|
|
break;
|
|
OP_cmp_T (GE, U, long, lvec2, lvec4, >=, ulong, ulvec2, ulvec4);
|
|
case OP_MEMSET_I:
|
|
pr_memset (op_c, OPA(int), st->b);
|
|
break;
|
|
case OP_MEMSET_P:
|
|
pr_memset (pr->pr_globals + OPC(ptr), OPA(int), OPB(uint));
|
|
break;
|
|
case OP_MEMSET_PI:
|
|
pr_memset (pr->pr_globals + OPC(ptr), OPA(int), st->b);
|
|
break;
|
|
case OP_STATE_ftt:
|
|
{
|
|
int self = *pr->globals.self;
|
|
int nextthink = pr->fields.nextthink + self;
|
|
int frame = pr->fields.frame + self;
|
|
int think = pr->fields.think + self;
|
|
float time = *pr->globals.ftime + OPC(float);
|
|
PR_PTR (float, &pr->pr_edict_area[nextthink]) = time;
|
|
PR_PTR (float, &pr->pr_edict_area[frame]) = OPA(float);
|
|
PR_PTR (func, &pr->pr_edict_area[think]) = OPB(func);
|
|
}
|
|
break;
|
|
// 1 1110
|
|
OP_cmp_T (LE, u, int, ivec2, ivec4, <=, uint, uivec2, uivec4);
|
|
case OP_IFZ:
|
|
if (!OPC(int)) {
|
|
pr->pr_xstatement = pr_jump_mode (pr, st, 0);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_IFB:
|
|
if (OPC(int) < 0) {
|
|
pr->pr_xstatement = pr_jump_mode (pr, st, 0);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_IFA:
|
|
if (OPC(int) > 0) {
|
|
pr->pr_xstatement = pr_jump_mode (pr, st, 0);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_STATE_dt:
|
|
{
|
|
int self = *pr->globals.self;
|
|
int nextthink = pr->fields.nextthink + self;
|
|
int frame = pr->fields.frame + self;
|
|
int think = pr->fields.think + self;
|
|
double time = *pr->globals.dtime + 0.1;
|
|
PR_PTR (double, &pr->pr_edict_area[nextthink]) = time;
|
|
PR_PTR (int, &pr->pr_edict_area[frame]) = OPA(int);
|
|
PR_PTR (func, &pr->pr_edict_area[think]) = OPB(func);
|
|
}
|
|
break;
|
|
OP_cmp_T (LE, U, long, lvec2, lvec4, <=, ulong, ulvec2, ulvec4);
|
|
case OP_IFNZ:
|
|
if (OPC(int)) {
|
|
pr->pr_xstatement = pr_jump_mode (pr, st, 0);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_IFAE:
|
|
if (OPC(int) >= 0) {
|
|
pr->pr_xstatement = pr_jump_mode (pr, st, 0);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_IFBE:
|
|
if (OPC(int) <= 0) {
|
|
pr->pr_xstatement = pr_jump_mode (pr, st, 0);
|
|
st = pr->pr_statements + pr->pr_xstatement;
|
|
}
|
|
break;
|
|
case OP_STATE_dtt:
|
|
{
|
|
int self = *pr->globals.self;
|
|
int nextthink = pr->fields.nextthink + self;
|
|
int frame = pr->fields.frame + self;
|
|
int think = pr->fields.think + self;
|
|
double time = *pr->globals.dtime + OPC(double);
|
|
PR_PTR (double, &pr->pr_edict_area[nextthink]) = time;
|
|
PR_PTR (int, &pr->pr_edict_area[frame]) = OPA(int);
|
|
PR_PTR (func, &pr->pr_edict_area[think]) = OPB(func);
|
|
}
|
|
break;
|
|
// 1 1111
|
|
case OP_LEA_A:
|
|
case OP_LEA_B:
|
|
case OP_LEA_C:
|
|
case OP_LEA_D:
|
|
mm = pr_address_mode (pr, st, (st_op - OP_LEA_A));
|
|
OPC(ptr) = mm - pr->pr_globals;
|
|
break;
|
|
case OP_QV4MUL_F:
|
|
OPC(vec4) = qvmulf (OPA(vec4), OPB(vec4));
|
|
break;
|
|
case OP_V4QMUL_F:
|
|
OPC(vec4) = vqmulf (OPA(vec4), OPB(vec4));
|
|
break;
|
|
case OP_QV4MUL_D:
|
|
OPC(dvec4) = qvmuld (OPA(dvec4), OPB(dvec4));
|
|
break;
|
|
case OP_V4QMUL_D:
|
|
OPC(dvec4) = vqmuld (OPA(dvec4), OPB(dvec4));
|
|
break;
|
|
case OP_BITAND_F:
|
|
OPC(float) = (int) OPA(float) & (int) OPB(float);
|
|
break;
|
|
case OP_BITOR_F:
|
|
OPC(float) = (int) OPA(float) | (int) OPB(float);
|
|
break;
|
|
case OP_BITXOR_F:
|
|
OPC(float) = (int) OPA(float) ^ (int) OPB(float);
|
|
break;
|
|
case OP_BITNOT_F:
|
|
OPC(float) = ~ (int) OPA(float);
|
|
break;
|
|
case OP_CONV:
|
|
switch (st->b) {
|
|
#include "libs/gamecode/pr_convert.cinc"
|
|
default:
|
|
PR_RunError (pr, "invalid conversion code: %04o",
|
|
st->b);
|
|
}
|
|
break;
|
|
case OP_WITH:
|
|
pr_with (pr, st);
|
|
break;
|
|
case OP_EXTEND:
|
|
switch (st->b) {
|
|
#include "libs/gamecode/pr_extend.cinc"
|
|
default:
|
|
PR_RunError (pr, "invalid extend code: %04o", st->b);
|
|
}
|
|
break;
|
|
#define OP_hop2(vec, op) ((vec)[0] op (vec)[1])
|
|
#define OP_hop3(vec, op) ((vec)[0] op (vec)[1] op (vec)[2])
|
|
#define OP_hop4(vec, op) ((vec)[0] op (vec)[1] op (vec)[2] op (vec)[3])
|
|
case OP_HOPS:
|
|
switch (st->b) {
|
|
#include "libs/gamecode/pr_hops.cinc"
|
|
default:
|
|
PR_RunError (pr, "invalid hops code: %04o", st->b);
|
|
}
|
|
break;
|
|
default:
|
|
PR_RunError (pr, "Bad opcode o%03o", st->op & OP_MASK);
|
|
}
|
|
if (pr->watch && pr->watch->value != old_val.value) {
|
|
if (!pr->wp_conditional
|
|
|| pr->watch->value == pr->wp_val.value) {
|
|
if (pr->debug_handler) {
|
|
pr->debug_handler (prd_watchpoint, 0, pr->debug_data);
|
|
} else {
|
|
PR_RunError (pr, "watchpoint hit: %d -> %d",
|
|
old_val.value, pr->watch->value);
|
|
}
|
|
}
|
|
old_val.value = pr->watch->value;
|
|
}
|
|
}
|
|
exit_program:
|
|
(void)0;//FIXME for clang
|
|
}
|
|
/*
|
|
PR_ExecuteProgram
|
|
|
|
The interpretation main loop
|
|
*/
|
|
VISIBLE void
|
|
PR_ExecuteProgram (progs_t *pr, pr_func_t fnum)
|
|
{
|
|
Sys_PushSignalHook (signal_hook, pr);
|
|
Sys_PushErrorHandler (error_handler, pr);
|
|
|
|
if (pr->debug_handler) {
|
|
pr->debug_handler (prd_subenter, &fnum, pr->debug_data);
|
|
}
|
|
|
|
int exitdepth = pr->pr_depth;
|
|
if (!PR_CallFunction (pr, fnum, pr->pr_return)) {
|
|
// called a builtin instead of progs code
|
|
goto exit_program;
|
|
}
|
|
if (pr->progs->version < PROG_VERSION) {
|
|
pr_exec_quakec (pr, exitdepth);
|
|
} else {
|
|
pr_exec_ruamoko (pr, exitdepth);
|
|
}
|
|
exit_program:
|
|
if (pr->debug_handler) {
|
|
pr->debug_handler (prd_subexit, 0, pr->debug_data);
|
|
}
|
|
pr->pr_argc = 0;
|
|
Sys_PopErrorHandler ();
|
|
Sys_PopSignalHook ();
|
|
}
|