91f87b3109
git-svn-id: https://yap.svn.sf.net/svnroot/yap/trunk@1047 b08c6af1-5177-4d33-ba66-4b1c6b8b522a
2931 lines
80 KiB
C
2931 lines
80 KiB
C
/*************************************************************************
|
|
* *
|
|
* YAP Prolog *
|
|
* *
|
|
* Yap Prolog was developed at NCCUP - Universidade do Porto *
|
|
* *
|
|
* Copyright L.Damas, V.S.Costa and Universidade do Porto 1985-1997 *
|
|
* *
|
|
**************************************************************************
|
|
* *
|
|
* File: amasm.c *
|
|
* comments: abstract machine assembler *
|
|
* *
|
|
* Last rev: $Date: 2004-04-22 20:07:04 $ *
|
|
* $Log: not supported by cvs2svn $
|
|
* Revision 1.59 2004/03/31 01:03:09 vsc
|
|
* support expand group of clauses
|
|
*
|
|
* Revision 1.58 2004/03/10 14:59:55 vsc
|
|
* optimise -> for type tests
|
|
* *
|
|
* *
|
|
*************************************************************************/
|
|
#ifdef SCCS
|
|
static char SccsId[] = "@(#)amasm.c 1.3 3/15/90";
|
|
|
|
#endif
|
|
|
|
#include "Yap.h"
|
|
#include "yapio.h"
|
|
#include "compile.h"
|
|
#include "clause.h"
|
|
#ifdef YAPOR
|
|
#include "or.macros.h"
|
|
#endif /* YAPOR */
|
|
#if HAVE_STRING_H
|
|
#include <string.h>
|
|
#endif
|
|
|
|
/* info on compare built-ins */
|
|
#define TYPE_XX 0
|
|
#define TYPE_CX 1
|
|
#define TYPE_XC 2
|
|
|
|
typedef struct cmp_op_info_struct {
|
|
wamreg x1_arg, x2_arg;
|
|
Int c_arg;
|
|
int c_type;
|
|
struct clause_info_struct *cl_info;
|
|
} cmp_op_info;
|
|
|
|
typedef struct clause_info_struct {
|
|
int alloc_found, dealloc_found;
|
|
CELL commit_lab;
|
|
struct pred_entry *CurrentPred;
|
|
} clause_info;
|
|
|
|
STATIC_PROTO(OPREG Var_Ref, (Ventry *, int));
|
|
STATIC_PROTO(wamreg emit_xreg, (CELL));
|
|
STATIC_PROTO(yslot emit_yreg, (CELL));
|
|
STATIC_PROTO(wamreg emit_x, (CELL));
|
|
STATIC_PROTO(yslot emit_y, (Ventry *));
|
|
STATIC_PROTO(yamop *emit_a, (CELL));
|
|
STATIC_PROTO(CELL *emit_bmlabel, (CELL, struct intermediates *));
|
|
STATIC_PROTO(yamop *emit_ilabel, (CELL, struct intermediates *));
|
|
STATIC_PROTO(Functor emit_f, (CELL));
|
|
STATIC_PROTO(CELL emit_c, (CELL));
|
|
STATIC_PROTO(COUNT emit_count, (CELL));
|
|
STATIC_PROTO(OPCODE emit_op, (op_numbers));
|
|
STATIC_PROTO(yamop *a_cl, (op_numbers, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_cle, (op_numbers, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_e, (op_numbers, yamop *, int));
|
|
STATIC_PROTO(yamop *a_ue, (op_numbers, op_numbers, yamop *, int));
|
|
STATIC_PROTO(yamop *a_v, (op_numbers, yamop *, int, struct PSEUDO *));
|
|
STATIC_PROTO(yamop *a_uv, (Ventry *,op_numbers, op_numbers, yamop *, int));
|
|
STATIC_PROTO(yamop *a_vr, (op_numbers, yamop *, int, struct PSEUDO *));
|
|
STATIC_PROTO(yamop *a_rv, (op_numbers, OPREG, yamop *, int, struct PSEUDO *));
|
|
STATIC_PROTO(yamop *a_vv, (op_numbers, op_numbers, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_glist, (int *, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(void a_pair, (CELL *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_f, (CELL, op_numbers, yamop *, int));
|
|
STATIC_PROTO(yamop *a_c, (CELL, op_numbers, yamop *, int));
|
|
STATIC_PROTO(yamop *a_uc, (CELL, op_numbers, op_numbers, yamop *, int));
|
|
STATIC_PROTO(yamop *a_n, (op_numbers, int, yamop *, int));
|
|
STATIC_PROTO(yamop *a_un, (op_numbers, op_numbers, int, yamop *, int));
|
|
STATIC_PROTO(yamop *a_nc, (CELL, op_numbers, int, yamop *, int));
|
|
STATIC_PROTO(yamop *a_unc, (CELL, op_numbers, op_numbers, int, yamop *, int));
|
|
STATIC_PROTO(yamop *a_r, (CELL, op_numbers, yamop *, int));
|
|
STATIC_PROTO(yamop *a_p, (op_numbers, clause_info *, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_pl, (op_numbers, PredEntry *, yamop *, int));
|
|
STATIC_PROTO(yamop *a_l, (CELL, op_numbers, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_hx, (op_numbers, union clause_obj *, int, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_if, (op_numbers, union clause_obj *, int, yamop *, int, struct intermediates *cip));
|
|
STATIC_PROTO(yamop *a_cut, (clause_info *,yamop *, int, struct intermediates *));
|
|
#ifdef YAPOR
|
|
STATIC_PROTO(yamop *a_try, (op_numbers, CELL, CELL, *clause_info, int, int, yamop *, int));
|
|
STATIC_PROTO(yamop *a_either, (op_numbers, CELL, CELL, int, int, yamop *, int, struct intermediates *));
|
|
#else
|
|
STATIC_PROTO(yamop *a_try, (op_numbers, CELL, CELL, clause_info *, yamop *, int));
|
|
STATIC_PROTO(yamop *a_either, (op_numbers, CELL, CELL, yamop *, int, struct intermediates *));
|
|
#endif /* YAPOR */
|
|
STATIC_PROTO(yamop *a_gl, (op_numbers, clause_info *, yamop *, int, struct PSEUDO *));
|
|
STATIC_PROTO(yamop *a_bfunc, (CELL, clause_info *, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(wamreg compile_cmp_flags, (char *));
|
|
STATIC_PROTO(yamop *a_igl, (CELL, op_numbers, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_xigl, (op_numbers, yamop *, int, struct PSEUDO *));
|
|
STATIC_PROTO(yamop *a_ucons, (int *, compiler_vm_op, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_uvar, (yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_wvar, (yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *do_pass, (int, yamop **, int, int *, struct intermediates *, UInt));
|
|
#ifdef DEBUG_OPCODES
|
|
STATIC_PROTO(void DumpOpCodes, (void));
|
|
#endif
|
|
#ifdef SFUNC
|
|
STATIC_PROTO(void a_vsf, (int, yamop *, int, struct PSEUDO *));
|
|
STATIC_PROTO(void a_asf, (int, yamop *, int, struct PSEUDO *));
|
|
#endif
|
|
STATIC_PROTO(yamop *check_alloc, (clause_info *, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_deallocate, (clause_info *, yamop *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_bmap, (yamop *, int, struct PSEUDO *));
|
|
STATIC_PROTO(void a_fetch_vv, (cmp_op_info *, int, struct intermediates *));
|
|
STATIC_PROTO(void a_fetch_cv, (cmp_op_info *, int, struct intermediates *));
|
|
STATIC_PROTO(void a_fetch_vc, (cmp_op_info *, int, struct intermediates *));
|
|
STATIC_PROTO(yamop *a_f2, (int, cmp_op_info *, yamop *, int, struct intermediates *));
|
|
|
|
#define CELLSIZE sizeof(CELL)
|
|
|
|
#define GONEXT(TYPE) code_p = ((yamop *)(&(code_p->u.TYPE.next)))
|
|
|
|
inline static yslot
|
|
emit_y(Ventry *ve)
|
|
{
|
|
#if MSHIFTOFFS
|
|
return(-FixedEnvSize - ((ve->NoOfVE) & MaskVarAdrs) - 1);
|
|
#else
|
|
return(-FixedEnvSize - (((ve->NoOfVE) & MaskVarAdrs) * CELLSIZE) - CELLSIZE);
|
|
#endif
|
|
}
|
|
|
|
inline static OPREG
|
|
Var_Ref(Ventry *ve, int is_y_var)
|
|
{
|
|
if (is_y_var) {
|
|
#if MSHIFTOFFS
|
|
return -FixedEnvSize - ((ve->NoOfVE) & MaskVarAdrs) - 1;
|
|
#else
|
|
return -FixedEnvSize - (((ve->NoOfVE) & MaskVarAdrs) * CELLSIZE) - CELLSIZE;
|
|
#endif
|
|
}
|
|
else {
|
|
#if PRECOMPUTE_REGADDRESS
|
|
return (CELL) (XREGS + ((ve->NoOfVE) & MaskVarAdrs));
|
|
#else
|
|
#if MSHIFTOFFS
|
|
return ((ve->NoOfVE) & MaskVarAdrs);
|
|
#else
|
|
return CELLSIZE * ((ve->NoOfVE) & MaskVarAdrs);
|
|
#endif
|
|
#endif /* PRECOMPUTE_REGADDRESS */
|
|
}
|
|
}
|
|
|
|
#define is_void_var() (((Ventry *) (cip->cpc->rnd1))->KindOfVE == VoidVar)
|
|
#define is_a_void(X) (((Ventry *) (X))->KindOfVE == VoidVar)
|
|
|
|
#define is_temp_var() (((Ventry *) (cip->cpc->rnd1))->KindOfVE == TempVar)
|
|
#define is_atemp_var(p) (((Ventry *) (p->rnd1))->KindOfVE == TempVar)
|
|
|
|
#define no_ref_var() (((Ventry *) (cip->cpc->rnd1))->NoOfVE == 1)
|
|
#define no_ref(X) (((Ventry *) (X))->NoOfVE == 1)
|
|
|
|
inline static yamop *
|
|
fill_small(CELL w, yamop *code_p, int pass_no)
|
|
{
|
|
SMALLUNSGN *ptr = ((SMALLUNSGN *) (code_p));
|
|
|
|
if (pass_no)
|
|
*ptr = (SMALLUNSGN) w;
|
|
return (yamop *) (++ptr);
|
|
}
|
|
|
|
inline static yamop *
|
|
fill_a(CELL a, yamop *code_p, int pass_no)
|
|
{
|
|
CELL *ptr = ((CELL *) (code_p));
|
|
|
|
if (pass_no)
|
|
*ptr = a;
|
|
return (yamop *) (++ptr);
|
|
}
|
|
|
|
inline static wamreg
|
|
emit_xreg(CELL w)
|
|
{
|
|
return (wamreg) w;
|
|
}
|
|
|
|
inline static yslot
|
|
emit_yreg(CELL w)
|
|
{
|
|
return (yslot) w;
|
|
}
|
|
|
|
inline static wamreg
|
|
emit_x(CELL xarg)
|
|
{
|
|
#if PRECOMPUTE_REGADDRESS
|
|
return (emit_xreg((CELL) (XREGS + xarg)));
|
|
#else
|
|
#if MSHIFTOFFS
|
|
return (emit_xreg(xarg));
|
|
#else
|
|
return (emit_xreg(CELLSIZE * (xarg)));
|
|
#endif
|
|
#endif /* PRECOMPUTE_REGADDRESS */
|
|
}
|
|
|
|
wamreg
|
|
Yap_emit_x(CELL xarg)
|
|
{
|
|
return emit_x(xarg);
|
|
}
|
|
|
|
inline static yamop *
|
|
emit_a(CELL a)
|
|
{
|
|
return ((yamop *) (a));
|
|
}
|
|
|
|
inline static struct pred_entry *
|
|
emit_pe(struct pred_entry *a)
|
|
{
|
|
return a;
|
|
}
|
|
|
|
inline static yamop *
|
|
emit_ilabel(register CELL addr, struct intermediates *cip)
|
|
{
|
|
if (addr & 1)
|
|
return (emit_a(Unsigned(cip->code_addr) + cip->label_offset[addr]));
|
|
else {
|
|
return (emit_a(addr));
|
|
}
|
|
}
|
|
|
|
inline static CELL *
|
|
emit_bmlabel(register CELL addr, struct intermediates *cip)
|
|
{
|
|
return (CELL *)(emit_a(Unsigned(cip->code_addr) + cip->label_offset[addr]));
|
|
}
|
|
|
|
inline static Functor
|
|
emit_f(CELL a)
|
|
{
|
|
return (Functor) (a);
|
|
}
|
|
|
|
inline static CELL
|
|
emit_c(CELL a)
|
|
{
|
|
return a;
|
|
}
|
|
|
|
static inline COUNT
|
|
emit_count(CELL count)
|
|
{
|
|
return count;
|
|
}
|
|
|
|
#ifdef DEBUG_OPCODES
|
|
inline static void
|
|
DumpOpCodes(void)
|
|
{
|
|
int i = 0, j;
|
|
|
|
while (i < 30) {
|
|
for (j = i; j <= _std_top; j += 25)
|
|
fprintf(Yap_stderr, "%5d %6lx", j, absmadr(j));
|
|
fputc('\n',Yap_stderr);
|
|
++i;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
static inline OPCODE
|
|
emit_op(op_numbers op)
|
|
{
|
|
return absmadr((Int) op);
|
|
}
|
|
|
|
static OPCODE
|
|
opcode(op_numbers op)
|
|
{
|
|
return (emit_op(op));
|
|
}
|
|
|
|
OPCODE
|
|
Yap_opcode(op_numbers op)
|
|
{
|
|
return opcode(op);
|
|
}
|
|
|
|
static void
|
|
add_clref(CELL clause_code, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
LogUpdClause *cl = ClauseCodeToLogUpdClause(clause_code);
|
|
cl->ClRefCount++;
|
|
}
|
|
}
|
|
|
|
static yamop *
|
|
a_cl(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.l.l = cip->code_addr;
|
|
}
|
|
GONEXT(l);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_lucl(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.Ill.I = (LogUpdIndex *)cip->code_addr;
|
|
code_p->u.Ill.l1 = emit_ilabel(cip->cpc->rnd1, cip);
|
|
code_p->u.Ill.l2 = emit_ilabel(cip->cpc->rnd2, cip);
|
|
code_p->u.Ill.s = cip->cpc->rnd3;
|
|
}
|
|
GONEXT(Ill);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_cle(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
LogUpdClause *cl = (LogUpdClause *)cip->code_addr;
|
|
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.EC.ClTrail = 0;
|
|
code_p->u.EC.ClENV = 0;
|
|
code_p->u.EC.ClRefs = 0;
|
|
code_p->u.EC.ClBase = cip->code_addr;
|
|
cl->ClExt = code_p;
|
|
cl->ClFlags |= LogUpdRuleMask;
|
|
}
|
|
GONEXT(EC);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_e(op_numbers opcode, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
}
|
|
GONEXT(e);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_ue(op_numbers opcode, op_numbers opcodew, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.o.opcw = emit_op(opcodew);
|
|
}
|
|
GONEXT(o);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_v(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
Ventry *ve = (Ventry *) cpc->rnd1;
|
|
OPREG var_offset;
|
|
int is_y_var = (ve->KindOfVE == PermVar);
|
|
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
if (ve->KindOfVE == PermVar) {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.y.y = emit_yreg(var_offset);
|
|
}
|
|
GONEXT(y);
|
|
}
|
|
else {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.x.x = emit_xreg(var_offset);
|
|
}
|
|
GONEXT(x);
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_uv(Ventry *ve, op_numbers opcode, op_numbers opcodew, yamop *code_p, int pass_no)
|
|
{
|
|
OPREG var_offset;
|
|
int is_y_var = (ve->KindOfVE == PermVar);
|
|
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
if (is_y_var) {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.oy.opcw = emit_op((op_numbers)((int)opcodew + is_y_var));
|
|
code_p->u.oy.y = emit_yreg(var_offset);
|
|
}
|
|
GONEXT(oy);
|
|
}
|
|
else {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.ox.opcw = emit_op((op_numbers)((int)opcodew + is_y_var));
|
|
code_p->u.ox.x = emit_xreg(var_offset);
|
|
}
|
|
GONEXT(ox);
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_vv(op_numbers opcode, op_numbers opcodew, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
Ventry *ve = (Ventry *) cip->cpc->rnd1;
|
|
int is_y_var = (ve->KindOfVE == PermVar);
|
|
|
|
if (pass_no) {
|
|
OPREG var_offset = Var_Ref(ve, is_y_var);
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.oxx.opcw = emit_op(opcodew);
|
|
code_p->u.oxx.xl = emit_xreg(var_offset);
|
|
}
|
|
cip->cpc = cip->cpc->nextInst;
|
|
if (pass_no) {
|
|
OPREG var_offset;
|
|
int is_y_var;
|
|
|
|
ve = (Ventry *) cip->cpc->rnd1;
|
|
is_y_var = (ve->KindOfVE == PermVar);
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
code_p->u.oxx.xr = emit_xreg(var_offset);
|
|
}
|
|
GONEXT(oxx);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_vr(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
Ventry *ve = (Ventry *) cpc->rnd1;
|
|
int is_y_var = (ve->KindOfVE == PermVar);
|
|
|
|
if (is_y_var) {
|
|
if (pass_no) {
|
|
OPREG var_offset;
|
|
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.yx.y = emit_yreg(var_offset);
|
|
code_p->u.yx.x = emit_x(cpc->rnd2);
|
|
}
|
|
GONEXT(yx);
|
|
}
|
|
else {
|
|
if (pass_no) {
|
|
OPREG var_offset;
|
|
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.xx.xl = emit_xreg(var_offset);
|
|
code_p->u.xx.xr = emit_x(cpc->rnd2);
|
|
}
|
|
GONEXT(xx);
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_rv(op_numbers opcode, OPREG var_offset, yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
Ventry *ve = (Ventry *) cpc->rnd1;
|
|
int is_y_var = (ve->KindOfVE == PermVar);
|
|
|
|
if (is_y_var) {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.xy.x = emit_x(cpc->rnd2);
|
|
code_p->u.xy.y = emit_yreg(var_offset);
|
|
}
|
|
GONEXT(xy);
|
|
}
|
|
else {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.xx.xl = emit_x(cpc->rnd2);
|
|
code_p->u.xx.xr = emit_xreg(var_offset);
|
|
}
|
|
GONEXT(xx);
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
#ifdef SFUNC
|
|
|
|
/* vsc: I don't understand these instructions */
|
|
|
|
inline static void
|
|
a_vsf(int opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
Ventry *ve = (Ventry *) cpc->rnd1;
|
|
OPREG var_offset;
|
|
int is_y_var = (ve->KindOfVE == PermVar);
|
|
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
if (is_y_var) {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.fy.f = emit_f(cpc->rnd2);
|
|
code_p->u.fy.a = ArityOfFunctor(emit_f(cpc->rnd2));
|
|
code_p->u.fy.y = emit_yreg(var_offset);
|
|
}
|
|
GONEXT(fy);
|
|
}
|
|
else {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.fx.f = emit_f(cpc->rnd2);
|
|
code_p->u.fx.a = ArityOfFunctor(emit_f(cpc->rnd2));
|
|
code_p->u.fx.x = emit_xreg(var_offset);
|
|
}
|
|
GONEXT(fx);
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
inline static void
|
|
a_asf(int opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
|
|
code_p->u.fn.f = emit_f(cpc->rnd2);
|
|
code_p->u.fn.a = ArityOfFunctor(emit_f(cpc->rnd2));
|
|
code_p->u.fn.n = emit_count(cpc->rnd1);
|
|
}
|
|
GONEXT(fn);
|
|
return code_p;
|
|
}
|
|
#endif
|
|
|
|
inline static void
|
|
a_pair(CELL *seq_ptr, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
CELL lab, lab0 = seq_ptr[1];
|
|
lab = (CELL) emit_ilabel(lab0, cip);
|
|
seq_ptr[0] = (CELL) emit_a(seq_ptr[0]);
|
|
seq_ptr[1] = lab;
|
|
}
|
|
}
|
|
|
|
inline static yamop *
|
|
a_n(op_numbers opcode, int count, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.s.s = count;
|
|
}
|
|
GONEXT(s);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_un(op_numbers opcode, op_numbers opcodew, int count, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.os.opcw = emit_op(opcodew);
|
|
code_p->u.os.s = count;
|
|
}
|
|
GONEXT(os);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_f(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
Functor f = emit_f(rnd1);
|
|
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.f.f = f;
|
|
code_p->u.f.a = ArityOfFunctor(f);
|
|
}
|
|
GONEXT(f);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_uf(CELL rnd1, op_numbers opcode, op_numbers opcodew, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
Functor f = emit_f(rnd1);
|
|
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.of.opcw = emit_op(opcodew);
|
|
code_p->u.of.f = f;
|
|
code_p->u.of.a = ArityOfFunctor(f);
|
|
}
|
|
GONEXT(of);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_c(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.c.c = emit_c(rnd1);
|
|
}
|
|
GONEXT(c);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_uc(CELL rnd1, op_numbers opcode, op_numbers opcode_w, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.oc.opcw = emit_op(opcode_w);
|
|
code_p->u.oc.c = emit_c(rnd1);
|
|
}
|
|
GONEXT(oc);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_blob(CELL rnd1, op_numbers opcode, int *clause_has_blobsp, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.c.c =
|
|
AbsAppl((CELL *)(Unsigned(cip->code_addr) + cip->label_offset[rnd1]));
|
|
}
|
|
*clause_has_blobsp = TRUE;
|
|
GONEXT(c);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_ublob(CELL rnd1, op_numbers opcode, op_numbers opcode_w, int *clause_has_blobsp, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.oc.opcw = emit_op(opcode_w);
|
|
code_p->u.oc.c =
|
|
AbsAppl((CELL *)(Unsigned(cip->code_addr) + cip->label_offset[rnd1]));
|
|
|
|
}
|
|
*clause_has_blobsp = TRUE;
|
|
GONEXT(oc);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_nc(CELL rnd1, op_numbers opcode, int i, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.sc.s = i;
|
|
code_p->u.sc.c = emit_c(rnd1);
|
|
}
|
|
GONEXT(sc);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_unc(CELL rnd1, op_numbers opcode, op_numbers opcodew, int i, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.osc.opcw = emit_op(opcodew);
|
|
code_p->u.osc.s = i;
|
|
code_p->u.osc.c = emit_c(rnd1);
|
|
}
|
|
GONEXT(osc);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_rf(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.xf.x = emit_x(cpc->rnd2);
|
|
code_p->u.xf.f = emit_f(cpc->rnd1);
|
|
code_p->u.xf.a = ArityOfFunctor(emit_f(cpc->rnd1));
|
|
}
|
|
GONEXT(xf);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_rc(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.xc.x = emit_x(cpc->rnd2);
|
|
code_p->u.xc.c = emit_c(cpc->rnd1);
|
|
}
|
|
GONEXT(xc);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_rb(op_numbers opcode, int *clause_has_blobsp, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.xc.x = emit_x(cip->cpc->rnd2);
|
|
code_p->u.xc.c = AbsAppl((CELL *)(Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1]));
|
|
}
|
|
*clause_has_blobsp = TRUE;
|
|
GONEXT(xc);
|
|
return code_p;
|
|
}
|
|
|
|
inline static yamop *
|
|
a_r(CELL arnd2, op_numbers opcode, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.x.x = emit_x(arnd2);
|
|
}
|
|
GONEXT(x);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
check_alloc(clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (clinfo->alloc_found == 2) {
|
|
if (clinfo->CurrentPred->PredFlags & LogUpdatePredFlag)
|
|
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
|
|
code_p = a_e(_allocate, code_p, pass_no);
|
|
clinfo->alloc_found = 1;
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_p(op_numbers opcode, clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{ /* emit opcode & predicate code address */
|
|
Prop fe = (Prop) (cip->cpc->rnd1);
|
|
CELL Flags = RepPredProp(fe)->PredFlags;
|
|
if (Flags & AsmPredFlag) {
|
|
op_numbers op;
|
|
|
|
code_p = check_alloc(clinfo, code_p, pass_no, cip);
|
|
switch (Flags & 0x7f) {
|
|
case _equal:
|
|
op = _p_equal;
|
|
break;
|
|
case _dif:
|
|
op = _p_dif;
|
|
break;
|
|
case _eq:
|
|
op = _p_eq;
|
|
break;
|
|
case _functor:
|
|
op = _p_functor;
|
|
break;
|
|
default:
|
|
op = _p_equal; /* just to make some compilers happy */
|
|
Yap_Error(SYSTEM_ERROR, TermNil, "internal assembler error for built-in (%d)", (Flags & 0x7f));
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
}
|
|
return a_e(op, code_p, pass_no);
|
|
}
|
|
if (Flags & CPredFlag) {
|
|
code_p = check_alloc(clinfo, code_p, pass_no, cip);
|
|
if (clinfo->commit_lab && (Flags & TestPredFlag)) {
|
|
if (pass_no) {
|
|
if (Flags & UserCPredFlag) {
|
|
Yap_Error(SYSTEM_ERROR, TermNil,
|
|
"user defined predicate cannot be a test predicate");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
} else
|
|
code_p->opc = emit_op(_call_c_wfail);
|
|
code_p->u.sdl.s =
|
|
emit_count(-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2);
|
|
code_p->u.sdl.l =
|
|
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
|
|
code_p->u.sdl.p =
|
|
emit_pe(RepPredProp(fe));
|
|
}
|
|
GONEXT(sdl);
|
|
clinfo->commit_lab = 0;
|
|
} else {
|
|
if (pass_no) {
|
|
if (Flags & UserCPredFlag) {
|
|
code_p->opc = emit_op(_call_usercpred);
|
|
} else {
|
|
if (RepPredProp(fe)->FunctorOfPred == FunctorExecuteInMod) {
|
|
code_p->opc = emit_op(_p_execute);
|
|
} else {
|
|
code_p->opc = emit_op(_call_cpred);
|
|
}
|
|
}
|
|
code_p->u.sla.s = emit_count(-Signed(RealEnvSize) - CELLSIZE
|
|
* (cip->cpc->rnd2));
|
|
if (RepPredProp(fe)->FunctorOfPred != FunctorExecuteInMod) {
|
|
code_p->u.sla.sla_u.p = RepPredProp(fe);
|
|
} else {
|
|
code_p->u.sla.sla_u.mod = cip->cpc->rnd4;
|
|
}
|
|
code_p->u.sla.p0 = clinfo->CurrentPred;
|
|
if (cip->cpc->rnd2) {
|
|
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->arnds[1], cip);
|
|
} else {
|
|
/* there is no bitmap as there are no variables in the environment */
|
|
code_p->u.sla.bmap = NULL;
|
|
}
|
|
}
|
|
GONEXT(sla);
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
if (opcode == _call && clinfo->alloc_found == 2) {
|
|
if (clinfo->CurrentPred->PredFlags & LogUpdatePredFlag)
|
|
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_fcall);
|
|
}
|
|
clinfo->alloc_found = 1;
|
|
}
|
|
else {
|
|
code_p = check_alloc(clinfo, code_p, pass_no, cip);
|
|
if (pass_no)
|
|
code_p->opc = emit_op(opcode);
|
|
}
|
|
if (opcode == _call) {
|
|
if (pass_no) {
|
|
code_p->u.sla.s = emit_count(-Signed(RealEnvSize) - CELLSIZE *
|
|
cip->cpc->rnd2);
|
|
code_p->u.sla.sla_u.p = RepPredProp(fe);
|
|
code_p->u.sla.p0 = clinfo->CurrentPred;
|
|
if (cip->cpc->rnd2)
|
|
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->arnds[1], cip);
|
|
else
|
|
/* there is no bitmap as there are no variables in the environment */
|
|
code_p->u.sla.bmap = NULL;
|
|
}
|
|
GONEXT(sla);
|
|
}
|
|
else {
|
|
if (pass_no)
|
|
code_p->u.p.p = RepPredProp(fe);
|
|
GONEXT(p);
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
/*
|
|
emit a false call so that the garbage collector and friends will find
|
|
reasonable information on the stack.
|
|
*/
|
|
static yamop *
|
|
a_empty_call(clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (clinfo->alloc_found == 1 && !clinfo->dealloc_found) {
|
|
/* we have a solid environment under us, just trust it */
|
|
if (pass_no)
|
|
code_p->opc = emit_op(_call);
|
|
}
|
|
else {
|
|
/** oops, our environment is crap */
|
|
if (pass_no)
|
|
code_p->opc = emit_op(_fcall);
|
|
}
|
|
if (pass_no) {
|
|
PredEntry *pe = RepPredProp(Yap_GetPredPropByAtom(AtomTrue,0));
|
|
code_p->u.sla.s = emit_count(-Signed(RealEnvSize) - CELLSIZE *
|
|
cip->cpc->rnd2);
|
|
code_p->u.sla.sla_u.p = pe;
|
|
code_p->u.sla.p0 = clinfo->CurrentPred;
|
|
if (cip->cpc->rnd2)
|
|
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->rnd1, cip);
|
|
else
|
|
/* there is no bitmap as there are no variables in the environment */
|
|
code_p->u.sla.bmap = NULL;
|
|
}
|
|
GONEXT(sla);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_l(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.l.l = emit_a(Unsigned(cip->code_addr) + cip->label_offset[rnd1]);
|
|
}
|
|
GONEXT(l);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_il(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.l.l = emit_ilabel(rnd1, cip);
|
|
}
|
|
GONEXT(l);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_pl(op_numbers opcode, PredEntry *pred, yamop *code_p, int pass_no)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.p.p = pred;
|
|
}
|
|
GONEXT(p);
|
|
return code_p;
|
|
}
|
|
|
|
static wamreg
|
|
compile_cmp_flags(char *s)
|
|
{
|
|
if (strcmp(s,"=<") == 0)
|
|
return(EQ_OK_IN_CMP|LT_OK_IN_CMP);
|
|
if (strcmp(s,"<") == 0)
|
|
return(LT_OK_IN_CMP);
|
|
if (strcmp(s,">=") == 0)
|
|
return(EQ_OK_IN_CMP|GT_OK_IN_CMP);
|
|
if (strcmp(s,">") == 0)
|
|
return(GT_OK_IN_CMP);
|
|
if (strcmp(s,"=:=") == 0)
|
|
return(EQ_OK_IN_CMP);
|
|
if (strcmp(s,"=\\=") == 0)
|
|
return(GT_OK_IN_CMP|LT_OK_IN_CMP);
|
|
Yap_Error(SYSTEM_ERROR, TermNil, "internal assembler error in flags for %s", s);
|
|
return(0);
|
|
}
|
|
|
|
wamreg
|
|
Yap_compile_cmp_flags(PredEntry *pred)
|
|
{
|
|
return compile_cmp_flags(RepAtom(NameOfFunctor(pred->FunctorOfPred))->StrOfAE);
|
|
}
|
|
|
|
static yamop *
|
|
a_bfunc(CELL pred, clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
Ventry *ve = (Ventry *) cip->cpc->rnd1;
|
|
OPREG var_offset;
|
|
int is_y_var = (ve->KindOfVE == PermVar);
|
|
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
if (ve->KindOfVE == PermVar) {
|
|
yslot v1 = emit_yreg(var_offset);
|
|
cip->cpc = cip->cpc->nextInst;
|
|
ve = (Ventry *) cip->cpc->rnd1;
|
|
is_y_var = (ve->KindOfVE == PermVar);
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
if (is_y_var) {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_call_bfunc_yy);
|
|
code_p->u.llyy.p = RepPredProp(((Prop)pred));
|
|
if (clinfo->commit_lab) {
|
|
code_p->u.llyy.f =
|
|
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
|
|
clinfo->commit_lab = 0;
|
|
} else {
|
|
code_p->u.llyy.f = FAILCODE;
|
|
}
|
|
code_p->u.llyy.y1 = v1;
|
|
code_p->u.llyy.y2 = emit_yreg(var_offset);
|
|
code_p->u.llyy.flags = compile_cmp_flags(RepAtom(NameOfFunctor(RepPredProp(((Prop)pred))->FunctorOfPred))->StrOfAE);
|
|
}
|
|
GONEXT(llyy);
|
|
} else {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_call_bfunc_yx);
|
|
code_p->u.llxy.p = RepPredProp(((Prop)pred));
|
|
if (clinfo->commit_lab) {
|
|
code_p->u.llxy.f =
|
|
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
|
|
clinfo->commit_lab = 0;
|
|
} else {
|
|
code_p->u.llxy.f = FAILCODE;
|
|
}
|
|
code_p->u.llxy.x = emit_xreg(var_offset);
|
|
code_p->u.llxy.y = v1;
|
|
code_p->u.llxy.flags = compile_cmp_flags(RepAtom(NameOfFunctor(RepPredProp(((Prop)pred))->FunctorOfPred))->StrOfAE);
|
|
}
|
|
GONEXT(llxy);
|
|
}
|
|
} else {
|
|
wamreg x1 = emit_xreg(var_offset);
|
|
OPREG var_offset;
|
|
|
|
cip->cpc = cip->cpc->nextInst;
|
|
ve = (Ventry *) cip->cpc->rnd1;
|
|
is_y_var = (ve->KindOfVE == PermVar);
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
if (is_y_var) {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_call_bfunc_xy);
|
|
code_p->u.llxy.p = RepPredProp(((Prop)pred));
|
|
if (clinfo->commit_lab) {
|
|
code_p->u.llxy.f =
|
|
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
|
|
clinfo->commit_lab = 0;
|
|
} else {
|
|
code_p->u.llxy.f = FAILCODE;
|
|
}
|
|
code_p->u.llxy.x = x1;
|
|
code_p->u.llxy.y = emit_yreg(var_offset);
|
|
code_p->u.llxy.flags = compile_cmp_flags(RepAtom(NameOfFunctor(RepPredProp(((Prop)pred))->FunctorOfPred))->StrOfAE);
|
|
}
|
|
GONEXT(llxy);
|
|
} else {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_call_bfunc_xx);
|
|
code_p->u.llxx.p = RepPredProp(((Prop)pred));
|
|
if (clinfo->commit_lab) {
|
|
code_p->u.llxx.f =
|
|
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
|
|
clinfo->commit_lab = 0;
|
|
} else {
|
|
code_p->u.llxx.f = FAILCODE;
|
|
}
|
|
code_p->u.llxx.x1 = x1;
|
|
code_p->u.llxx.x2 = emit_xreg(var_offset);
|
|
code_p->u.llxx.flags = compile_cmp_flags(RepAtom(NameOfFunctor(RepPredProp(((Prop)pred))->FunctorOfPred))->StrOfAE);
|
|
}
|
|
GONEXT(llxx);
|
|
}
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_igl(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.l.l = emit_ilabel(rnd1, cip);
|
|
}
|
|
GONEXT(l);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_xigl(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.xl.x = emit_x(cpc->rnd2);
|
|
code_p->u.xl.l = emit_a(cpc->rnd1);
|
|
}
|
|
GONEXT(xl);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_4sw(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
CELL *seq_ptr;
|
|
|
|
if (opcode == _switch_on_type &&
|
|
cip->cpc->nextInst != NULL &&
|
|
cip->cpc->nextInst->op == label_op &&
|
|
cip->cpc->arnds[1] == cip->cpc->nextInst->rnd1 &&
|
|
!(cip->cpc->arnds[0] & 1) &&
|
|
cip->cpc->nextInst->nextInst != NULL &&
|
|
cip->cpc->nextInst->nextInst->op == if_c_op &&
|
|
cip->cpc->nextInst->nextInst->rnd1 == 1 &&
|
|
cip->cpc->nextInst->nextInst->arnds[1] == TermNil &&
|
|
cip->cpc->nextInst->nextInst->arnds[0] == cip->cpc->arnds[2]) {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_switch_list_nl);
|
|
seq_ptr = cip->cpc->arnds;
|
|
code_p->u.ollll.pop = ((yamop *)(seq_ptr[0]))->opc;
|
|
code_p->u.ollll.l1 = emit_ilabel(seq_ptr[0], cip);
|
|
code_p->u.ollll.l2 = emit_ilabel(cip->cpc->nextInst->nextInst->arnds[2], cip);
|
|
code_p->u.ollll.l3 = emit_ilabel(seq_ptr[2], cip);
|
|
code_p->u.ollll.l4 = emit_ilabel(seq_ptr[3], cip);
|
|
}
|
|
GONEXT(ollll);
|
|
cip->cpc = cip->cpc->nextInst->nextInst;
|
|
} else {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
seq_ptr = cip->cpc->arnds;
|
|
code_p->u.llll.l1 = emit_ilabel(seq_ptr[0], cip);
|
|
code_p->u.llll.l2 = emit_ilabel(seq_ptr[1], cip);
|
|
code_p->u.llll.l3 = emit_ilabel(seq_ptr[2], cip);
|
|
code_p->u.llll.l4 = emit_ilabel(seq_ptr[3], cip);
|
|
}
|
|
GONEXT(llll);
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_4sw_x(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
CELL *seq_ptr;
|
|
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.xllll.x = emit_x(cip->cpc->rnd2);
|
|
cip->cpc = cip->cpc->nextInst;
|
|
seq_ptr = cip->cpc->arnds;
|
|
code_p->u.xllll.l1 = emit_ilabel(seq_ptr[0], cip);
|
|
code_p->u.xllll.l2 = emit_ilabel(seq_ptr[1], cip);
|
|
code_p->u.xllll.l3 = emit_ilabel(seq_ptr[2], cip);
|
|
code_p->u.xllll.l4 = emit_ilabel(seq_ptr[3], cip);
|
|
} else {
|
|
/* skip one */
|
|
cip->cpc = cip->cpc->nextInst;
|
|
}
|
|
GONEXT(xllll);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_4sw_s(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
CELL *seq_ptr;
|
|
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.sllll.s = cip->cpc->rnd2;
|
|
cip->cpc = cip->cpc->nextInst;
|
|
seq_ptr = cip->cpc->arnds;
|
|
code_p->u.sllll.l1 = emit_ilabel(seq_ptr[0], cip);
|
|
code_p->u.sllll.l2 = emit_ilabel(seq_ptr[1], cip);
|
|
code_p->u.sllll.l3 = emit_ilabel(seq_ptr[2], cip);
|
|
code_p->u.sllll.l4 = emit_ilabel(seq_ptr[3], cip);
|
|
} else {
|
|
/* skip one */
|
|
cip->cpc = cip->cpc->nextInst;
|
|
}
|
|
GONEXT(sllll);
|
|
return code_p;
|
|
}
|
|
|
|
static void
|
|
init_log_upd_table(LogUpdIndex *ic, union clause_obj *cl_u)
|
|
{
|
|
/* insert myself in the indexing code chain */
|
|
ic->SiblingIndex = cl_u->lui.ChildIndex;
|
|
ic->ChildIndex = NULL;
|
|
ic->ClRefCount = 0;
|
|
ic->u.ParentIndex = (LogUpdIndex *)cl_u;
|
|
INIT_LOCK(ic->ClLock);
|
|
cl_u->lui.ChildIndex = ic;
|
|
cl_u->lui.ClRefCount++;
|
|
}
|
|
|
|
static void
|
|
init_static_table(StaticIndex *ic, union clause_obj *cl_u)
|
|
{
|
|
/* insert myself in the indexing code chain */
|
|
ic->SiblingIndex = cl_u->si.ChildIndex;
|
|
ic->ChildIndex = NULL;
|
|
cl_u->si.ChildIndex = ic;
|
|
}
|
|
|
|
static yamop *
|
|
a_hx(op_numbers opcode, union clause_obj *cl_u, int log_update, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
register CELL i, imax;
|
|
register CELL *seq_ptr = (CELL *)cip->cpc->rnd2;
|
|
int j = 0;
|
|
|
|
imax = cip->cpc->rnd1;
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.sssl.s = emit_c(imax);
|
|
code_p->u.sssl.l = emit_a(cip->cpc->rnd2);
|
|
if (log_update) {
|
|
init_log_upd_table(ClauseCodeToLogUpdIndex(cip->cpc->rnd2), cl_u);
|
|
} else {
|
|
init_static_table(ClauseCodeToStaticIndex(cip->cpc->rnd2), cl_u);
|
|
}
|
|
}
|
|
if (pass_no) {
|
|
for (i = 0; i < imax; i++) {
|
|
yamop *ipc = (yamop *)seq_ptr[1];
|
|
a_pair(seq_ptr, pass_no, cip);
|
|
if (ipc != FAILCODE) {
|
|
j++;
|
|
}
|
|
seq_ptr += 2;
|
|
}
|
|
code_p->u.sssl.e = j;
|
|
code_p->u.sssl.w = 0;
|
|
}
|
|
GONEXT(sssl);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_if(op_numbers opcode, union clause_obj *cl_u, int log_update, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
register CELL i, imax;
|
|
register CELL *seq_ptr = (CELL *)cip->cpc->rnd2;
|
|
|
|
imax = cip->cpc->rnd1;
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.sssl.s = code_p->u.sssl.e = emit_count(imax);
|
|
code_p->u.sssl.w = 0;
|
|
code_p->u.sssl.l = emit_a(cip->cpc->rnd2);
|
|
if (log_update) {
|
|
init_log_upd_table(ClauseCodeToLogUpdIndex(cip->cpc->rnd2), cl_u);
|
|
} else {
|
|
init_static_table(ClauseCodeToStaticIndex(cip->cpc->rnd2), cl_u);
|
|
}
|
|
}
|
|
GONEXT(sssl);
|
|
if (pass_no) {
|
|
CELL lab, lab0;
|
|
for (i = 0; i < imax; i++) {
|
|
a_pair(seq_ptr, pass_no, cip);
|
|
seq_ptr += 2;
|
|
}
|
|
lab0 = seq_ptr[1];
|
|
lab = (CELL) emit_ilabel(lab0, cip);
|
|
seq_ptr[1] = lab;
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_ifnot(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
CELL *seq_ptr = cip->cpc->arnds;
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.clll.c = seq_ptr[0]; /* tag */
|
|
code_p->u.clll.l1 = emit_ilabel(seq_ptr[1], cip); /* success point */
|
|
code_p->u.clll.l2 = emit_ilabel(seq_ptr[2], cip); /* fail point */
|
|
code_p->u.clll.l3 = emit_ilabel(seq_ptr[3], cip); /* delay point */
|
|
}
|
|
GONEXT(clll);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_cut(clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
code_p = check_alloc(clinfo, code_p, pass_no, cip);
|
|
if (clinfo->dealloc_found) {
|
|
return a_e(_cut_e, code_p, pass_no);
|
|
} else if (clinfo->alloc_found) {
|
|
return a_e(_cut, code_p, pass_no);
|
|
} else {
|
|
return a_e(_cut_t, code_p, pass_no);
|
|
}
|
|
}
|
|
|
|
static yamop *
|
|
#ifdef YAPOR
|
|
a_try(op_numbers opcode, CELL lab, CELL opr, clause_info *clinfo, int nofalts, int hascut, yamop *code_p, int pass_no)
|
|
#else
|
|
a_try(op_numbers opcode, CELL lab, CELL opr, clause_info *clinfo, yamop *code_p, int pass_no)
|
|
#endif /* YAPOR */
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.ld.d = emit_a(lab);
|
|
code_p->u.ld.s = emit_count(opr);
|
|
code_p->u.ld.p = clinfo->CurrentPred;
|
|
#ifdef TABLING
|
|
code_p->u.ld.te = clinfo->CurrentPred->TableOfPred;
|
|
#endif
|
|
#ifdef YAPOR
|
|
INIT_YAMOP_LTT(code_p, nofalts);
|
|
if (hascut)
|
|
PUT_YAMOP_CUT(code_p);
|
|
if (clinfo->CurrentPred->PredFlags & SequentialPredFlag)
|
|
PUT_YAMOP_SEQ(code_p);
|
|
#endif /* YAPOR */
|
|
}
|
|
GONEXT(ld);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
#ifdef YAPOR
|
|
a_either(op_numbers opcode, CELL opr, CELL lab, int nofalts, int hascut, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
#else
|
|
a_either(op_numbers opcode, CELL opr, CELL lab, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
#endif /* YAPOR */
|
|
{
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(opcode);
|
|
code_p->u.sla.s = emit_count(opr);
|
|
code_p->u.sla.sla_u.l = emit_a(lab);
|
|
code_p->u.sla.p0 = cip->CurrentPred;
|
|
#ifdef YAPOR
|
|
INIT_YAMOP_LTT(code_p, nofalts);
|
|
if (hascut)
|
|
PUT_YAMOP_CUT(code_p);
|
|
if (clinfo->CurrentPred->PredFlags & SequentialPredFlag)
|
|
PUT_YAMOP_SEQ(code_p);
|
|
if(opcode != _or_last) {
|
|
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->arnds[1], cip);
|
|
}
|
|
#else
|
|
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->arnds[1], cip);
|
|
#endif /* YAPOR */
|
|
}
|
|
GONEXT(sla);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_gl(op_numbers opcode, clause_info *clinfo, yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
#ifdef YAPOR
|
|
return a_try(opcode, cpc->rnd1, IPredArity, clinfo, cpc->rnd2 >> 1, cpc->rnd2 & 1, code_p, pass_no);
|
|
#else
|
|
return a_try(opcode, cpc->rnd1, IPredArity, clinfo, code_p, pass_no);
|
|
#endif /* YAPOR */
|
|
}
|
|
|
|
/*
|
|
* optimizes several unify_cons for the same constant. It must be avoided for
|
|
* the head of the first argument, because of indexing
|
|
*/
|
|
static yamop *
|
|
a_ucons(int *do_not_optimise_uatomp, compiler_vm_op opcode, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
#if AGGREGATE_OPS
|
|
PInstr *np = cip->cpc->nextInst;
|
|
register int i = 0;
|
|
CELL my_cons = cip->cpc->rnd1;
|
|
|
|
|
|
if (*do_not_optimise_uatomp) {
|
|
*do_not_optimise_uatomp = FALSE;
|
|
if (opcode == unify_atom_op)
|
|
return a_uc(cip->cpc->rnd1, _unify_atom, _unify_atom_write, code_p, pass_no);
|
|
else
|
|
return a_c(cip->cpc->rnd1, _write_atom, code_p, pass_no);
|
|
}
|
|
else {
|
|
while (np->op == opcode && np->rnd1 == my_cons) {
|
|
i++;
|
|
cip->cpc = np;
|
|
np = np->nextInst;
|
|
}
|
|
if (i == 0) {
|
|
if (opcode == unify_atom_op)
|
|
return a_uc(cip->cpc->rnd1, _unify_atom, _unify_atom_write, code_p, pass_no);
|
|
else
|
|
return a_c(cip->cpc->rnd1, _write_atom, code_p, pass_no);
|
|
} else {
|
|
if (opcode == unify_atom_op)
|
|
return a_unc(cip->cpc->rnd1, _unify_n_atoms, _unify_n_atoms_write, i + 1, code_p, pass_no);
|
|
else
|
|
return a_nc(cip->cpc->rnd1, _write_n_atoms, i + 1, code_p, pass_no);
|
|
}
|
|
}
|
|
#else
|
|
*do_not_optimise_uatomp = FALSE;
|
|
if (opcode == unify_atom_op)
|
|
return a_uc(cip->cpc->rnd1, _unify_atom, _unify_atom_write, code_p);
|
|
else
|
|
return a_c(cip->cpc->rnd1, _write_atom, code_p);
|
|
#endif
|
|
}
|
|
|
|
static yamop *
|
|
a_uvar(yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (!is_void_var()) {
|
|
#if AGGREGATE_OPS
|
|
if (is_temp_var()) {
|
|
PInstr *np = cip->cpc->nextInst;
|
|
|
|
if (np->op == unify_var_op &&
|
|
is_atemp_var(np)) {
|
|
return a_vv(_unify_x_var2, _unify_x_var2_write, code_p, pass_no, cip);
|
|
}
|
|
else if (np->op == unify_last_var_op &&
|
|
is_atemp_var(np)) {
|
|
return a_vv(_unify_l_x_var2,
|
|
_unify_l_x_var2_write, code_p, pass_no, cip);
|
|
}
|
|
}
|
|
#endif /* AGGREGATE_OPS */
|
|
return a_uv((Ventry *) cip->cpc->rnd1, _unify_x_var, _unify_x_var_write, code_p, pass_no);
|
|
}
|
|
else {
|
|
#if AGGREGATE_OPS
|
|
int i = 1;
|
|
PInstr *np = cip->cpc->nextInst;
|
|
|
|
/* skip void vars */
|
|
while (np->op == unify_var_op && is_a_void(np->rnd1)) {
|
|
i++;
|
|
cip->cpc = np;
|
|
np = np->nextInst;
|
|
}
|
|
if (np->op == unify_last_var_op &&
|
|
is_a_void(np->rnd1)) {
|
|
if (i == 0)
|
|
code_p = a_ue(_unify_l_void, _unify_l_void_write, code_p, pass_no);
|
|
else
|
|
code_p = a_un(_unify_l_n_voids, _unify_l_n_voids_write, i + 1, code_p, pass_no);
|
|
cip->cpc = np;
|
|
}
|
|
else if (i == 1)
|
|
return a_ue(_unify_void, _unify_void_write, code_p, pass_no);
|
|
else {
|
|
return a_un(_unify_n_voids, _unify_n_voids_write, i, code_p, pass_no);
|
|
}
|
|
#else
|
|
return a_ue(_unify_void, _unify_void_write);
|
|
#endif
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_wvar(yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (!no_ref_var())
|
|
return a_v(_write_x_var, code_p, pass_no, cip->cpc);
|
|
else {
|
|
#if AGGREGATE_OPS
|
|
int i = 0;
|
|
PInstr *np = cip->cpc->nextInst;
|
|
|
|
while (np->op == write_var_op && no_ref(np->rnd1)) {
|
|
i++;
|
|
cip->cpc = np;
|
|
np = np->nextInst;
|
|
}
|
|
if (i == 0)
|
|
return a_e(_write_void, code_p, pass_no);
|
|
else {
|
|
return a_n(_write_n_voids, i + 1, code_p, pass_no);
|
|
}
|
|
#else
|
|
return a_e(_write_void, pass_no);
|
|
#endif
|
|
}
|
|
}
|
|
|
|
static yamop *
|
|
a_glist(int *do_not_optimise_uatomp, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
#if AGGREGATE_OPS
|
|
PInstr *pnext = cip->cpc->nextInst;
|
|
|
|
if (cip->cpc->rnd2 != 1 && pnext->op == unify_val_op) {
|
|
Ventry *ve = (Ventry *) pnext->rnd1;
|
|
int is_y_var;
|
|
OPREG var_offset;
|
|
|
|
pnext->rnd2 = cip->cpc->rnd2;
|
|
cip->cpc = pnext;
|
|
is_y_var = (ve->KindOfVE == PermVar);
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
return a_rv(_glist_valx, var_offset, code_p, pass_no, cip->cpc);
|
|
} else if (cip->cpc->rnd2 == 1 && pnext->op == unify_atom_op) {
|
|
*do_not_optimise_uatomp = TRUE;
|
|
return a_r(cip->cpc->rnd2, _get_list, code_p, pass_no);
|
|
} else if (cip->cpc->rnd2 != 1 && pnext->op == unify_var_op
|
|
&& is_a_void(pnext->rnd1)) {
|
|
PInstr *ppnext = pnext->nextInst;
|
|
|
|
if (ppnext && (ppnext->op == unify_last_var_op
|
|
|| ppnext->op == unify_last_val_op)) {
|
|
Ventry *ve = (Ventry *) ppnext->rnd1;
|
|
int is_y_var = (ve->KindOfVE == PermVar);
|
|
OPREG var_offset;
|
|
|
|
ppnext->rnd2 = cip->cpc->rnd2;
|
|
cip->cpc = ppnext;
|
|
var_offset = Var_Ref(ve, is_y_var);
|
|
return a_rv((op_numbers)((int)_gl_void_varx + (cip->cpc->op == unify_last_var_op ? 0 : 2)), var_offset, code_p, pass_no, cip->cpc);
|
|
} else {
|
|
return a_r(cip->cpc->rnd2, _get_list, code_p, pass_no);
|
|
}
|
|
} else
|
|
#endif /* AGGREGATE_OPS */
|
|
return a_r(cip->cpc->rnd2, _get_list, code_p, pass_no);
|
|
}
|
|
|
|
#define NEXTOPC (cip->cpc->nextInst)->op
|
|
|
|
static yamop *
|
|
a_deallocate(clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
if (clinfo->alloc_found == 2) {
|
|
/* this should never happen */
|
|
if (clinfo->CurrentPred->PredFlags & LogUpdatePredFlag)
|
|
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
|
|
code_p = a_e(_allocate, code_p, pass_no);
|
|
}
|
|
if (NEXTOPC == execute_op) {
|
|
cip->cpc = cip->cpc->nextInst;
|
|
code_p = a_p(_dexecute, clinfo, code_p, pass_no, cip);
|
|
} else
|
|
code_p = a_e(_deallocate, code_p, pass_no);
|
|
clinfo->dealloc_found = TRUE;
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_bmap(yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
/* how much space do we need to reserve */
|
|
int i, max = (cpc->rnd1)/(8*sizeof(CELL));
|
|
for (i = 0; i <= max; i++)
|
|
code_p = fill_a(cpc->arnds[i], code_p, pass_no);
|
|
return code_p;
|
|
}
|
|
|
|
static yamop *
|
|
a_bregs(yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
/* how much space do we need to reserve */
|
|
int i, max = (cpc->rnd1)/(8*sizeof(CELL));
|
|
code_p = fill_a(cpc->rnd1, code_p, pass_no);
|
|
for (i = 0; i <= max; i++)
|
|
code_p = fill_a(cpc->arnds[i], code_p, pass_no);
|
|
return code_p;
|
|
}
|
|
|
|
|
|
static yamop *
|
|
copy_blob(yamop *code_p, int pass_no, struct PSEUDO *cpc)
|
|
{
|
|
/* copy the blob to code space, making no effort to align if a double */
|
|
int max = cpc->rnd1, i;
|
|
for (i = 0; i < max; i++)
|
|
code_p = fill_a(cpc->arnds[i], code_p, pass_no);
|
|
return code_p;
|
|
}
|
|
|
|
|
|
static void
|
|
a_fetch_vv(cmp_op_info *cmp_info, int pass_no, struct intermediates *cip)
|
|
{
|
|
/* the next three instructions must be a get_val, get_val, and BIP */
|
|
if (pass_no == 0) {
|
|
PInstr *p = cip->cpc->nextInst;
|
|
Ventry *ve;
|
|
ve = (Ventry *) p->rnd1;
|
|
if (ve->KindOfVE != PermVar)
|
|
p->op = nop_op;
|
|
p = p->nextInst;
|
|
ve = (Ventry *) p->rnd1;
|
|
if (ve->KindOfVE != PermVar)
|
|
p->op = nop_op;
|
|
} else {
|
|
PInstr *p = cip->cpc->nextInst;
|
|
Ventry *ve;
|
|
|
|
cmp_info->c_type = TYPE_XX;
|
|
ve = (Ventry *) p->rnd1;
|
|
if (ve->KindOfVE == PermVar) {
|
|
/* don't get rid of get_val_op */
|
|
cmp_info->x1_arg = emit_x(p->rnd2);
|
|
} else {
|
|
/* and use it directly as a second argument */
|
|
cmp_info->x1_arg = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
}
|
|
p = p->nextInst;
|
|
ve = (Ventry *) p->rnd1;
|
|
if (ve->KindOfVE == PermVar) {
|
|
/* don't get rid of get_val_op */
|
|
cmp_info->x2_arg = emit_x(p->rnd2);
|
|
} else {
|
|
/* and use it directly as a second argument */
|
|
cmp_info->x2_arg = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
a_fetch_vc(cmp_op_info *cmp_info, int pass_no, struct intermediates *cip)
|
|
{
|
|
/* the next two instructions must be a get_val and BIP */
|
|
if (pass_no == 0) {
|
|
PInstr *p = cip->cpc->nextInst;
|
|
Ventry *ve;
|
|
ve = (Ventry *) p->rnd1;
|
|
if (ve->KindOfVE != PermVar)
|
|
p->op = nop_op;
|
|
} else {
|
|
PInstr *p = cip->cpc->nextInst;
|
|
Ventry *ve;
|
|
|
|
cmp_info->c_type = TYPE_XC;
|
|
cmp_info->c_arg = (Int)(cip->cpc->rnd1);
|
|
ve = (Ventry *) p->rnd1;
|
|
if (ve->KindOfVE == PermVar) {
|
|
/* don't get rid of get_val_op */
|
|
cmp_info->x1_arg = emit_x(p->rnd2);
|
|
} else {
|
|
/* and use it directly as a second argument */
|
|
cmp_info->x1_arg = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void
|
|
a_fetch_cv(cmp_op_info *cmp_info, int pass_no, struct intermediates *cip)
|
|
{
|
|
/* the next two instructions must be a get_val and BIP */
|
|
if (pass_no == 0) {
|
|
PInstr *p = cip->cpc->nextInst;
|
|
Ventry *ve;
|
|
ve = (Ventry *) p->rnd1;
|
|
if (ve->KindOfVE != PermVar)
|
|
p->op = nop_op;
|
|
} else {
|
|
PInstr *p = cip->cpc->nextInst;
|
|
Ventry *ve;
|
|
|
|
cmp_info->c_type = TYPE_CX;
|
|
cmp_info->c_arg = (Int)(cip->cpc->rnd1);
|
|
ve = (Ventry *) p->rnd1;
|
|
if (ve->KindOfVE == PermVar) {
|
|
/* don't get rid of get_val_op */
|
|
cmp_info->x1_arg = emit_x(p->rnd2);
|
|
} else {
|
|
/* and use it directly as a second argument */
|
|
cmp_info->x1_arg = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
}
|
|
}
|
|
}
|
|
|
|
static yamop *
|
|
a_f2(int var, cmp_op_info *cmp_info, yamop *code_p, int pass_no, struct intermediates *cip)
|
|
{
|
|
Int opc = cip->cpc->rnd2;
|
|
Ventry *ve = (Ventry *)(cip->cpc->rnd1);
|
|
int is_y_var = (ve->KindOfVE == PermVar);
|
|
|
|
if (opc <= _primitive) {
|
|
if (is_y_var) {
|
|
if (pass_no) {
|
|
code_p->u.yF.y = emit_y(ve);
|
|
switch (opc) {
|
|
case _atom:
|
|
code_p->opc = opcode(_p_atom_y);
|
|
break;
|
|
case _atomic:
|
|
code_p->opc = opcode(_p_atomic_y);
|
|
break;
|
|
case _compound:
|
|
code_p->opc = opcode(_p_compound_y);
|
|
break;
|
|
case _float:
|
|
code_p->opc = opcode(_p_float_y);
|
|
break;
|
|
case _integer:
|
|
code_p->opc = opcode(_p_integer_y);
|
|
break;
|
|
case _nonvar:
|
|
code_p->opc = opcode(_p_nonvar_y);
|
|
break;
|
|
case _number:
|
|
code_p->opc = opcode(_p_number_y);
|
|
break;
|
|
case _var:
|
|
code_p->opc = opcode(_p_var_y);
|
|
break;
|
|
case _db_ref:
|
|
code_p->opc = opcode(_p_db_ref_y);
|
|
break;
|
|
case _cut_by:
|
|
code_p->opc = opcode(_p_cut_by_y);
|
|
break;
|
|
case _primitive:
|
|
code_p->opc = opcode(_p_primitive_y);
|
|
break;
|
|
}
|
|
if (cmp_info->cl_info->commit_lab) {
|
|
code_p->u.yF.F =
|
|
emit_a(Unsigned(cip->code_addr) + cip->label_offset[cmp_info->cl_info->commit_lab]);
|
|
cmp_info->cl_info->commit_lab = 0;
|
|
} else {
|
|
code_p->u.yF.F = FAILCODE;
|
|
}
|
|
}
|
|
GONEXT(yF);
|
|
return code_p;
|
|
} else {
|
|
if (pass_no) {
|
|
code_p->u.xF.x = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
switch (opc) {
|
|
case _atom:
|
|
code_p->opc = opcode(_p_atom_x);
|
|
break;
|
|
case _atomic:
|
|
code_p->opc = opcode(_p_atomic_x);
|
|
break;
|
|
case _compound:
|
|
code_p->opc = opcode(_p_compound_x);
|
|
break;
|
|
case _float:
|
|
code_p->opc = opcode(_p_float_x);
|
|
break;
|
|
case _integer:
|
|
code_p->opc = opcode(_p_integer_x);
|
|
break;
|
|
case _nonvar:
|
|
code_p->opc = opcode(_p_nonvar_x);
|
|
break;
|
|
case _number:
|
|
code_p->opc = opcode(_p_number_x);
|
|
break;
|
|
case _var:
|
|
code_p->opc = opcode(_p_var_x);
|
|
break;
|
|
case _db_ref:
|
|
code_p->opc = opcode(_p_db_ref_x);
|
|
break;
|
|
case _cut_by:
|
|
code_p->opc = opcode(_p_cut_by_x);
|
|
break;
|
|
case _primitive:
|
|
code_p->opc = opcode(_p_primitive_x);
|
|
break;
|
|
}
|
|
if (cmp_info->cl_info->commit_lab) {
|
|
code_p->u.xF.F =
|
|
emit_a(Unsigned(cip->code_addr) + cip->label_offset[cmp_info->cl_info->commit_lab]);
|
|
cmp_info->cl_info->commit_lab = 0;
|
|
} else {
|
|
code_p->u.xF.F = FAILCODE;
|
|
}
|
|
}
|
|
GONEXT(xF);
|
|
return code_p;
|
|
}
|
|
}
|
|
if (opc == _functor && cip->cpc->nextInst->op == f_var_op) {
|
|
Ventry *nve;
|
|
|
|
cip->cpc = cip->cpc->nextInst;
|
|
nve = (Ventry *)(cip->cpc->rnd1);
|
|
if (is_y_var) {
|
|
if (nve->KindOfVE == PermVar) {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_p_func2f_yy);
|
|
code_p->u.yyx.y1 = emit_y(ve);
|
|
code_p->u.yyx.y2 = emit_y(nve);
|
|
code_p->u.yyx.x = cmp_info->x1_arg;
|
|
}
|
|
GONEXT(yyx);
|
|
return code_p;
|
|
} else {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_p_func2f_yx);
|
|
code_p->u.yxx.y = emit_y(ve);
|
|
code_p->u.yxx.x1 = emit_x(nve->NoOfVE & MaskVarAdrs);
|
|
code_p->u.yxx.x2 = cmp_info->x1_arg;
|
|
}
|
|
GONEXT(yxx);
|
|
return code_p;
|
|
}
|
|
} else {
|
|
if (nve->KindOfVE == PermVar) {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_p_func2f_xy);
|
|
code_p->u.xyx.x1 = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
code_p->u.xyx.y2 = emit_y(nve);
|
|
code_p->u.xyx.x = cmp_info->x1_arg;
|
|
}
|
|
GONEXT(xyx);
|
|
return code_p;
|
|
} else {
|
|
if (pass_no) {
|
|
code_p->opc = emit_op(_p_func2f_xx);
|
|
code_p->u.xxx.x1 = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
code_p->u.xxx.x2 = emit_x(nve->NoOfVE & MaskVarAdrs);
|
|
code_p->u.xxx.x = cmp_info->x1_arg;
|
|
}
|
|
GONEXT(xxx);
|
|
return code_p;
|
|
}
|
|
}
|
|
}
|
|
if (is_y_var) {
|
|
switch (cmp_info->c_type) {
|
|
case TYPE_XX:
|
|
if (pass_no) {
|
|
switch (opc) {
|
|
case _plus:
|
|
code_p->opc = emit_op(_p_plus_y_vv);
|
|
break;
|
|
case _minus:
|
|
code_p->opc = emit_op(_p_minus_y_vv);
|
|
break;
|
|
case _times:
|
|
code_p->opc = emit_op(_p_times_y_vv);
|
|
break;
|
|
case _div:
|
|
code_p->opc = emit_op(_p_div_y_vv);
|
|
break;
|
|
case _and:
|
|
code_p->opc = emit_op(_p_and_y_vv);
|
|
break;
|
|
case _or:
|
|
code_p->opc = emit_op(_p_or_y_vv);
|
|
break;
|
|
case _sll:
|
|
code_p->opc = emit_op(_p_sll_y_vv);
|
|
break;
|
|
case _slr:
|
|
code_p->opc = emit_op(_p_slr_y_vv);
|
|
break;
|
|
case _arg:
|
|
code_p->opc = emit_op(_p_arg_y_vv);
|
|
break;
|
|
case _functor:
|
|
code_p->opc = emit_op(_p_func2s_y_vv);
|
|
break;
|
|
}
|
|
code_p->u.yxx.y = emit_y(ve);
|
|
code_p->u.yxx.x1 = cmp_info->x1_arg;
|
|
code_p->u.yxx.x2 = cmp_info->x2_arg;
|
|
}
|
|
GONEXT(yxx);
|
|
break;
|
|
case TYPE_CX:
|
|
if (pass_no) {
|
|
switch (opc) {
|
|
case _plus:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for +/2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _minus:
|
|
code_p->opc = emit_op(_p_minus_y_cv);
|
|
break;
|
|
case _times:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for */2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _div:
|
|
code_p->opc = emit_op(_p_div_y_cv);
|
|
break;
|
|
case _and:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for /\\/2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _or:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for \\//2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _sll:
|
|
code_p->opc = emit_op(_p_sll_y_cv);
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _slr:
|
|
code_p->opc = emit_op(_p_slr_y_cv);
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _arg:
|
|
code_p->opc = emit_op(_p_arg_y_cv);
|
|
break;
|
|
case _functor:
|
|
code_p->opc = emit_op(_p_func2s_y_cv);
|
|
break;
|
|
}
|
|
code_p->u.ycx.y = emit_y(ve);
|
|
code_p->u.ycx.c = cmp_info->c_arg;
|
|
code_p->u.ycx.xi = cmp_info->x1_arg;
|
|
}
|
|
GONEXT(ycx);
|
|
break;
|
|
case TYPE_XC:
|
|
if (pass_no) {
|
|
switch (opc) {
|
|
case _plus:
|
|
code_p->opc = emit_op(_p_plus_y_vc);
|
|
break;
|
|
case _minus:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x2_arg, "internal assembler error XC for -/2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _times:
|
|
code_p->opc = emit_op(_p_times_y_vc);
|
|
break;
|
|
case _div:
|
|
code_p->opc = emit_op(_p_div_y_vc);
|
|
break;
|
|
case _and:
|
|
code_p->opc = emit_op(_p_and_y_vc);
|
|
break;
|
|
case _or:
|
|
code_p->opc = emit_op(_p_or_y_vc);
|
|
break;
|
|
case _sll:
|
|
code_p->opc = emit_op(_p_sll_y_vc);
|
|
break;
|
|
case _slr:
|
|
code_p->opc = emit_op(_p_slr_y_vc);
|
|
break;
|
|
case _arg:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x2_arg, "internal assembler error for arg/3");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _functor:
|
|
code_p->opc = emit_op(_p_func2s_y_vc);
|
|
break;
|
|
}
|
|
code_p->u.yxc.y = emit_y(ve);
|
|
code_p->u.yxc.c = cmp_info->c_arg;
|
|
code_p->u.yxc.xi = cmp_info->x1_arg;
|
|
}
|
|
GONEXT(yxc);
|
|
break;
|
|
}
|
|
} else {
|
|
switch (cmp_info->c_type) {
|
|
case TYPE_XX:
|
|
if (pass_no) {
|
|
switch (opc) {
|
|
case _plus:
|
|
code_p->opc = emit_op(_p_plus_vv);
|
|
break;
|
|
case _minus:
|
|
code_p->opc = emit_op(_p_minus_vv);
|
|
break;
|
|
case _times:
|
|
code_p->opc = emit_op(_p_times_vv);
|
|
break;
|
|
case _div:
|
|
code_p->opc = emit_op(_p_div_vv);
|
|
break;
|
|
case _and:
|
|
code_p->opc = emit_op(_p_and_vv);
|
|
break;
|
|
case _or:
|
|
code_p->opc = emit_op(_p_or_vv);
|
|
break;
|
|
case _sll:
|
|
code_p->opc = emit_op(_p_sll_vv);
|
|
break;
|
|
case _slr:
|
|
code_p->opc = emit_op(_p_slr_vv);
|
|
break;
|
|
case _arg:
|
|
code_p->opc = emit_op(_p_arg_vv);
|
|
break;
|
|
case _functor:
|
|
code_p->opc = emit_op(_p_func2s_vv);
|
|
break;
|
|
}
|
|
code_p->u.xxx.x = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
code_p->u.xxx.x1 = cmp_info->x1_arg;
|
|
code_p->u.xxx.x2 = cmp_info->x2_arg;
|
|
}
|
|
GONEXT(xxx);
|
|
break;
|
|
case TYPE_CX:
|
|
if (pass_no) {
|
|
switch (opc) {
|
|
case _plus:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for +/2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _minus:
|
|
code_p->opc = emit_op(_p_minus_cv);
|
|
break;
|
|
case _times:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for */2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _div:
|
|
code_p->opc = emit_op(_p_div_cv);
|
|
break;
|
|
case _and:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for /\\/2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _or:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for \\//2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _sll:
|
|
code_p->opc = emit_op(_p_sll_cv);
|
|
break;
|
|
case _slr:
|
|
code_p->opc = emit_op(_p_slr_cv);
|
|
break;
|
|
case _arg:
|
|
code_p->opc = emit_op(_p_arg_cv);
|
|
break;
|
|
case _functor:
|
|
code_p->opc = emit_op(_p_func2s_cv);
|
|
break;
|
|
}
|
|
code_p->u.xxc.x = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
code_p->u.xxc.c = cmp_info->c_arg;
|
|
code_p->u.xxc.xi = cmp_info->x1_arg;
|
|
}
|
|
GONEXT(xxc);
|
|
break;
|
|
case TYPE_XC:
|
|
if (pass_no) {
|
|
switch (opc) {
|
|
case _plus:
|
|
code_p->opc = emit_op(_p_plus_vc);
|
|
break;
|
|
case _minus:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x2_arg, "internal assembler error XC for -/2");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _times:
|
|
code_p->opc = emit_op(_p_times_vc);
|
|
break;
|
|
case _div:
|
|
code_p->opc = emit_op(_p_div_vc);
|
|
break;
|
|
case _and:
|
|
code_p->opc = emit_op(_p_and_vc);
|
|
break;
|
|
case _or:
|
|
code_p->opc = emit_op(_p_or_vc);
|
|
break;
|
|
case _sll:
|
|
code_p->opc = emit_op(_p_sll_vc);
|
|
break;
|
|
case _slr:
|
|
code_p->opc = emit_op(_p_slr_vc);
|
|
break;
|
|
case _arg:
|
|
Yap_Error(SYSTEM_ERROR, cmp_info->x2_arg, "internal assembler error for arg/3");
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
break;
|
|
case _functor:
|
|
code_p->opc = emit_op(_p_func2s_vc);
|
|
break;
|
|
}
|
|
code_p->u.xcx.x = emit_x(ve->NoOfVE & MaskVarAdrs);
|
|
code_p->u.xcx.c = cmp_info->c_arg;
|
|
code_p->u.xcx.xi = cmp_info->x1_arg;
|
|
}
|
|
GONEXT(xcx);
|
|
break;
|
|
}
|
|
}
|
|
return code_p;
|
|
}
|
|
|
|
#define TRYOP(G,P) (IPredArity<5 ? (op_numbers)((int)(P)+(IPredArity*3)) : (G))
|
|
#ifdef YAPOR
|
|
#define TRYCODE(G,P) a_try(TRYOP(G,P), Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], IPredArity, &clinfo, cip->cpc->rnd2 >> 1, cip->cpc->rnd2 & 1, code_p, pass_no)
|
|
#define TABLE_TRYCODE(G) a_try(G, (CELL)emit_ilabel(cip->cpc->rnd1, code_addr, cip), IPredArity, cip->cpc->rnd2 >> 1, cip->cpc->rnd2 & 1, code_p, pass_no)
|
|
#else
|
|
#define TRYCODE(G,P) a_try(TRYOP(G,P), Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], IPredArity, &clinfo, code_p, pass_no)
|
|
#define TABLE_TRYCODE(G) a_try(G, (CELL)emit_ilabel(cip->cpc->rnd1, code_addr, cip), IPredArity, code_p, pass_no)
|
|
#endif /* YAPOR */
|
|
|
|
static yamop *
|
|
do_pass(int pass_no, yamop **entry_codep, int assembling, int *clause_has_blobsp, struct intermediates *cip, UInt size)
|
|
{
|
|
#ifdef YAPOR
|
|
#define EITHER_INST 50
|
|
yamop *either_inst[EITHER_INST];
|
|
int either_cont = 0;
|
|
#endif /* YAPOR */
|
|
int log_update;
|
|
int dynamic;
|
|
#ifdef TABLING
|
|
int tabled;
|
|
#endif
|
|
int ystop_found = FALSE;
|
|
union clause_obj *cl_u;
|
|
yamop *code_p;
|
|
cmp_op_info cmp_info;
|
|
clause_info clinfo;
|
|
int do_not_optimise_uatom;
|
|
|
|
|
|
code_p = cip->code_addr;
|
|
cl_u = (union clause_obj *)code_p;
|
|
cip->cpc = cip->CodeStart;
|
|
clinfo.alloc_found = clinfo.dealloc_found = FALSE;
|
|
clinfo.commit_lab = 0L;
|
|
clinfo.CurrentPred = cip->CurrentPred;
|
|
cmp_info.c_type = TYPE_XX;
|
|
cmp_info.cl_info = &clinfo;
|
|
do_not_optimise_uatom = FALSE;
|
|
|
|
/* Space while for the clause flags */
|
|
log_update = cip->CurrentPred->PredFlags & LogUpdatePredFlag;
|
|
dynamic = cip->CurrentPred->PredFlags & DynamicPredFlag;
|
|
#ifdef TABLING
|
|
tabled = cip->CurrentPred->PredFlags & TabledPredFlag;
|
|
#endif
|
|
if (assembling == ASSEMBLING_CLAUSE) {
|
|
if (log_update) {
|
|
if (pass_no) {
|
|
cl_u->luc.Id = FunctorDBRef;
|
|
cl_u->luc.ClFlags = LogUpdMask;
|
|
cl_u->luc.ClRefCount = 0;
|
|
cl_u->luc.ClPred = cip->CurrentPred;
|
|
cl_u->luc.ClSize = size;
|
|
if (*clause_has_blobsp) {
|
|
cl_u->luc.ClFlags |= HasBlobsMask;
|
|
}
|
|
cl_u->luc.ClExt = NULL;
|
|
cl_u->luc.ClPrev = cl_u->luc.ClNext = NULL;
|
|
#if defined(YAPOR) || defined(THREADS)
|
|
INIT_LOCK(cl_u->luc.ClLock);
|
|
INIT_CLREF_COUNT(&(cl_u->luc));
|
|
#endif
|
|
}
|
|
code_p = cl_u->luc.ClCode;
|
|
} else if (dynamic) {
|
|
if (pass_no) {
|
|
cl_u->ic.ClFlags = DynamicMask;
|
|
if (*clause_has_blobsp) {
|
|
cl_u->ic.ClFlags |= HasBlobsMask;
|
|
}
|
|
cl_u->ic.ClSize = size;
|
|
cl_u->ic.ClRefCount = 0;
|
|
#if defined(YAPOR) || defined(THREADS)
|
|
INIT_LOCK(cl_u->ic.ClLock);
|
|
INIT_CLREF_COUNT(&(cl_u->ic));
|
|
#endif
|
|
}
|
|
code_p = cl_u->ic.ClCode;
|
|
} else {
|
|
/* static clause */
|
|
if (pass_no) {
|
|
cl_u->sc.Id = FunctorDBRef;
|
|
cl_u->sc.ClFlags = StaticMask;
|
|
cl_u->sc.ClNext = NULL;
|
|
cl_u->sc.ClSize = size;
|
|
if (*clause_has_blobsp) {
|
|
cl_u->sc.ClFlags |= HasBlobsMask;
|
|
}
|
|
}
|
|
code_p = cl_u->sc.ClCode;
|
|
}
|
|
IPredArity = cip->cpc->rnd2; /* number of args */
|
|
*entry_codep = code_p;
|
|
if (dynamic) {
|
|
#ifdef YAPOR
|
|
a_try(TRYOP(_try_me, _try_me0), 0, IPredArity, &clinfo, 1, 0, code_p, pass_no);
|
|
#else
|
|
a_try(TRYOP(_try_me, _try_me0), 0, IPredArity, &clinfo, code_p, pass_no);
|
|
#endif /* YAPOR */
|
|
}
|
|
} else {
|
|
/* index code */
|
|
if (log_update) {
|
|
if (pass_no) {
|
|
cl_u->lui.ClFlags = LogUpdMask|IndexedPredFlag|IndexMask|SwitchRootMask;
|
|
cl_u->lui.ChildIndex = NULL;
|
|
cl_u->lui.SiblingIndex = NULL;
|
|
cl_u->lui.u.pred = cip->CurrentPred;
|
|
cl_u->lui.ClSize = size;
|
|
cl_u->lui.ClRefCount = 0;
|
|
INIT_LOCK(cl_u->lui.ClLock);
|
|
#if defined(YAPOR) || defined(THREADS)
|
|
INIT_CLREF_COUNT(&(cl_u->lui));
|
|
#endif
|
|
}
|
|
code_p = cl_u->lui.ClCode;
|
|
*entry_codep = code_p;
|
|
#if defined(YAPOR) || defined(THREADS)
|
|
if (assembling == ASSEMBLING_INDEX &&
|
|
!(cip->CurrentPred->PredFlags & ThreadLocalPredFlag)) {
|
|
if (pass_no) {
|
|
code_p->opc = opcode(_lock_lu);
|
|
code_p->u.p.p = cip->CurrentPred;
|
|
}
|
|
GONEXT(p);
|
|
}
|
|
#endif
|
|
} else {
|
|
if (pass_no) {
|
|
cl_u->si.ClSize = size;
|
|
cl_u->si.ClFlags = IndexMask;
|
|
cl_u->si.ChildIndex = NULL;
|
|
cl_u->si.SiblingIndex = NULL;
|
|
}
|
|
code_p = cl_u->si.ClCode;
|
|
*entry_codep = code_p;
|
|
}
|
|
}
|
|
while (cip->cpc) {
|
|
|
|
switch ((int) cip->cpc->op) {
|
|
#ifdef YAPOR
|
|
case sync_op:
|
|
code_p = a_try(_sync, cip->cpc->rnd1, cip->cpc->rnd2, 1, Zero, code_p);
|
|
break;
|
|
#endif /* YAPOR */
|
|
#ifdef TABLING
|
|
case table_new_answer_op:
|
|
code_p = a_n(_table_new_answer, (int) cip->cpc->rnd2, code_p);
|
|
break;
|
|
case table_try_single_op:
|
|
code_p = a_gl(_table_try_single, code_p, clinfo, code_p, pass_no, cpc);
|
|
break;
|
|
#endif /* TABLING */
|
|
#ifdef TABLING_INNER_CUTS
|
|
case clause_with_cut_op:
|
|
code_p = a_e(_clause_with_cut, code_p);
|
|
break;
|
|
#endif /* TABLING_INNER_CUTS */
|
|
#ifdef SFUNC
|
|
case get_s_f_op:
|
|
code_p = a_rf(_get_s_f, code_p, cip->cpc);
|
|
break;
|
|
case put_s_f_op:
|
|
code_p = a_rf(_put_s_f, code_p, cip->cpc);
|
|
break;
|
|
case unify_s_f_op:
|
|
code_p = a_d(_unify_s_f, code_p);
|
|
break;
|
|
case write_s_f_op:
|
|
code_p = a_f(cip->cpc->rnd1, _write_s_f);
|
|
break;
|
|
case unify_s_var_op:
|
|
code_p = a_vsf(_unify_s_xvar);
|
|
break;
|
|
case write_s_var_op:
|
|
code_p = a_vsf(_write_s_xvar);
|
|
break;
|
|
case unify_s_val_op:
|
|
code_p = a_vsf(_unify_s_xval);
|
|
break;
|
|
case write_s_val_op:
|
|
code_p = a_vsf(_write_s_xval);
|
|
break;
|
|
case unify_s_a_op:
|
|
code_p = a_asf(_unify_s_a);
|
|
break;
|
|
case write_s_a_op:
|
|
code_p = a_asf(_write_s_a);
|
|
break;
|
|
case get_s_end_op:
|
|
code_p = a_n(_get_s_end, Unsigned(0));
|
|
break;
|
|
case put_s_end_op:
|
|
code_p = a_n(_put_s_end, Unsigned(0));
|
|
break;
|
|
case unify_s_end_op:
|
|
code_p = a_n(_write_s_end, Unsigned(0));
|
|
break;
|
|
case write_s_end_op:
|
|
code_p = a_n(_write_s_end, Unsigned(0));
|
|
break;
|
|
#endif
|
|
case get_var_op:
|
|
code_p = a_vr(_get_x_var, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case put_var_op:
|
|
code_p = a_vr(_put_x_var, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case get_val_op:
|
|
code_p = a_vr(_get_x_val, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case put_val_op:
|
|
code_p = a_vr(_put_x_val, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case get_num_op:
|
|
case get_atom_op:
|
|
code_p = a_rc(_get_atom, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case get_float_op:
|
|
code_p = a_rb(_get_float, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case get_longint_op:
|
|
code_p = a_rb(_get_longint, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case get_bigint_op:
|
|
code_p = a_rb(_get_bigint, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case put_num_op:
|
|
case put_atom_op:
|
|
code_p = a_rc(_put_atom, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case put_float_op:
|
|
case put_longint_op:
|
|
case put_bigint_op:
|
|
code_p = a_rb(_put_atom, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case get_list_op:
|
|
code_p = a_glist(&do_not_optimise_uatom, code_p, pass_no, cip);
|
|
break;
|
|
case put_list_op:
|
|
code_p = a_r(cip->cpc->rnd2, _put_list, code_p, pass_no);
|
|
break;
|
|
case get_struct_op:
|
|
code_p = a_rf(_get_struct, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case put_struct_op:
|
|
code_p = a_rf(_put_struct, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case put_unsafe_op:
|
|
code_p = a_vr((op_numbers)((int)_put_unsafe - 1), code_p, pass_no, cip->cpc);
|
|
break;
|
|
case unify_var_op:
|
|
code_p = a_uvar(code_p, pass_no, cip);
|
|
break;
|
|
case unify_last_var_op:
|
|
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_l_x_var, _unify_l_x_var_write, code_p, pass_no);
|
|
break;
|
|
case write_var_op:
|
|
code_p = a_wvar(code_p, pass_no, cip);
|
|
break;
|
|
case unify_local_op:
|
|
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_x_loc, _unify_x_loc_write, code_p, pass_no);
|
|
break;
|
|
case unify_val_op:
|
|
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_x_val, _unify_x_val_write, code_p, pass_no);
|
|
break;
|
|
case unify_last_local_op:
|
|
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_l_x_loc, _unify_l_x_loc_write, code_p, pass_no);
|
|
break;
|
|
case unify_last_val_op:
|
|
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_l_x_val, _unify_l_x_val_write, code_p, pass_no);
|
|
break;
|
|
case write_local_op:
|
|
code_p = a_v(_write_x_loc, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case write_val_op:
|
|
code_p = a_v(_write_x_val, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case unify_num_op:
|
|
case unify_atom_op:
|
|
code_p = a_ucons(&do_not_optimise_uatom, unify_atom_op, code_p, pass_no, cip);
|
|
break;
|
|
case unify_float_op:
|
|
code_p = a_ublob(cip->cpc->rnd1, _unify_float, _unify_atom_write, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case unify_longint_op:
|
|
code_p = a_ublob(cip->cpc->rnd1, _unify_longint, _unify_atom_write, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case unify_bigint_op:
|
|
code_p = a_ublob(cip->cpc->rnd1, _unify_bigint, _unify_atom_write, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case unify_last_num_op:
|
|
case unify_last_atom_op:
|
|
code_p = a_uc(cip->cpc->rnd1, _unify_l_atom, _unify_l_atom_write, code_p, pass_no);
|
|
break;
|
|
case unify_last_float_op:
|
|
code_p = a_ublob(cip->cpc->rnd1, _unify_l_float, _unify_l_atom_write, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case unify_last_longint_op:
|
|
code_p = a_ublob(cip->cpc->rnd1, _unify_l_longint, _unify_l_atom_write, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case unify_last_bigint_op:
|
|
code_p = a_ublob(cip->cpc->rnd1, _unify_l_bigint, _unify_l_atom_write, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case write_num_op:
|
|
case write_atom_op:
|
|
code_p = a_ucons(&do_not_optimise_uatom, write_atom_op, code_p, pass_no, cip);
|
|
break;
|
|
case write_float_op:
|
|
case write_longint_op:
|
|
case write_bigint_op:
|
|
code_p = a_blob(cip->cpc->rnd1, _write_atom, clause_has_blobsp, code_p, pass_no, cip);
|
|
break;
|
|
case unify_list_op:
|
|
code_p = a_ue(_unify_list, _unify_list_write, code_p, pass_no);
|
|
break;
|
|
case unify_last_list_op:
|
|
code_p = a_ue(_unify_l_list, _unify_l_list_write, code_p, pass_no);
|
|
break;
|
|
case write_list_op:
|
|
code_p = a_e(_write_list, code_p, pass_no);
|
|
break;
|
|
case write_last_list_op:
|
|
code_p = a_e(_write_l_list, code_p, pass_no);
|
|
break;
|
|
case unify_struct_op:
|
|
code_p = a_uf(cip->cpc->rnd1, _unify_struct, _unify_struct_write, code_p, pass_no);
|
|
break;
|
|
case unify_last_struct_op:
|
|
code_p = a_uf(cip->cpc->rnd1, _unify_l_struc, _unify_l_struc_write, code_p, pass_no);
|
|
break;
|
|
case write_struct_op:
|
|
code_p = a_f(cip->cpc->rnd1, _write_struct, code_p, pass_no);
|
|
break;
|
|
case write_last_struct_op:
|
|
code_p = a_f(cip->cpc->rnd1, _write_l_struc, code_p, pass_no);
|
|
break;
|
|
case save_b_op:
|
|
case patch_b_op:
|
|
code_p = a_v(_save_b_x, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case commit_b_op:
|
|
code_p = a_v(_commit_b_x, code_p, pass_no, cip->cpc);
|
|
#ifdef YAPOR
|
|
if (pass_no)
|
|
PUT_YAMOP_CUT(*entry_codep);
|
|
#endif /* YAPOR */
|
|
break;
|
|
case save_pair_op:
|
|
code_p = a_uv((Ventry *) cip->cpc->rnd1, _save_pair_x, _save_pair_x_write, code_p, pass_no);
|
|
break;
|
|
case save_appl_op:
|
|
code_p = a_uv((Ventry *) cip->cpc->rnd1, _save_appl_x, _save_appl_x_write, code_p, pass_no);
|
|
break;
|
|
case fail_op:
|
|
code_p = a_e(_op_fail, code_p, pass_no);
|
|
break;
|
|
case cut_op:
|
|
code_p = a_cut(&clinfo, code_p, pass_no, cip);
|
|
#ifdef YAPOR
|
|
if (pass_no)
|
|
PUT_YAMOP_CUT(*entry_codep);
|
|
#endif /* YAPOR */
|
|
break;
|
|
case cutexit_op:
|
|
code_p = a_cut(&clinfo, code_p, pass_no, cip);
|
|
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
|
|
*clause_has_blobsp &&
|
|
!clinfo.alloc_found)
|
|
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
|
|
#if THREADS
|
|
else
|
|
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
|
|
!(cip->CurrentPred->PredFlags & ThreadLocalPredFlag))
|
|
code_p = a_e(_unlock_lu, code_p, pass_no);
|
|
#endif
|
|
code_p = a_e(_procceed, code_p, pass_no);
|
|
#ifdef YAPOR
|
|
if (pass_no)
|
|
PUT_YAMOP_CUT(*entry_codep);
|
|
#endif /* YAPOR */
|
|
break;
|
|
case allocate_op:
|
|
clinfo.alloc_found = 2;
|
|
break;
|
|
case deallocate_op:
|
|
code_p = a_deallocate(&clinfo, code_p, pass_no, cip);
|
|
break;
|
|
case tryme_op:
|
|
#ifdef TABLING
|
|
if (tabled)
|
|
code_p = TABLE_TRYCODE(_table_try_me);
|
|
else
|
|
#endif
|
|
code_p = TRYCODE(_try_me, _try_me0);
|
|
break;
|
|
case retryme_op:
|
|
#ifdef TABLING
|
|
if (tabled)
|
|
code_p = TABLE_TRYCODE(_table_retry_me);
|
|
else
|
|
#endif
|
|
code_p = TRYCODE(_retry_me, _retry_me0);
|
|
break;
|
|
case trustme_op:
|
|
if (log_update &&
|
|
(assembling == ASSEMBLING_INDEX ||
|
|
assembling == ASSEMBLING_EINDEX)) {
|
|
code_p = a_cl(_trust_logical_pred, code_p, pass_no, cip);
|
|
}
|
|
#ifdef TABLING
|
|
if (tabled)
|
|
code_p = TABLE_TRYCODE(_table_trust_me);
|
|
else
|
|
#endif
|
|
code_p = TRYCODE(_trust_me, _trust_me0);
|
|
break;
|
|
case enter_lu_op:
|
|
code_p = a_lucl(_enter_lu_pred, code_p, pass_no, cip);
|
|
break;
|
|
case try_op:
|
|
#ifdef TABLING
|
|
if (tabled)
|
|
code_p = a_gl(_table_try, &clinfo, code_p, pass_no, cip->cpc);
|
|
else
|
|
#endif
|
|
code_p = a_gl(_try_clause, &clinfo, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case retry_op:
|
|
if (log_update) {
|
|
add_clref(cip->cpc->rnd1, pass_no);
|
|
}
|
|
#ifdef TABLING
|
|
if (tabled)
|
|
code_p = a_gl(_table_retry, &clinfo, code_p, pass_no, cip->cpc);
|
|
else
|
|
#endif
|
|
code_p = a_gl(_retry, &clinfo, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case trust_op:
|
|
if (log_update) {
|
|
add_clref(cip->cpc->rnd1, pass_no);
|
|
code_p = a_cl(_trust_logical_pred, code_p, pass_no, cip);
|
|
}
|
|
#ifdef TABLING
|
|
if (tabled)
|
|
code_p = a_gl(_table_trust, &clinfo, code_p, pass_no, cpc);
|
|
else
|
|
#endif
|
|
code_p = a_gl(_trust, &clinfo, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case try_in_op:
|
|
code_p = a_il(cip->cpc->rnd1, _try_in, code_p, pass_no, cip);
|
|
break;
|
|
case jump_op:
|
|
/* don't assemble jumps to next instruction */
|
|
if (cip->cpc->nextInst == NULL ||
|
|
cip->cpc->nextInst->op != label_op ||
|
|
cip->cpc->rnd1 != cip->cpc->nextInst->rnd1) {
|
|
code_p = a_l(cip->cpc->rnd1, _jump, code_p, pass_no, cip);
|
|
}
|
|
break;
|
|
case jumpi_op:
|
|
code_p = a_il(cip->cpc->rnd1, _jump, code_p, pass_no, cip);
|
|
break;
|
|
case restore_tmps_op:
|
|
code_p = a_l(cip->cpc->rnd1, _move_back, code_p, pass_no, cip);
|
|
break;
|
|
case restore_tmps_and_skip_op:
|
|
code_p = a_l(cip->cpc->rnd1, _skip, code_p, pass_no, cip);
|
|
break;
|
|
case procceed_op:
|
|
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
|
|
*clause_has_blobsp &&
|
|
!clinfo.alloc_found)
|
|
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
|
|
#if THREADS
|
|
else
|
|
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
|
|
!(cip->CurrentPred->PredFlags & ThreadLocalPredFlag))
|
|
code_p = a_e(_unlock_lu, code_p, pass_no);
|
|
#endif
|
|
code_p = a_e(_procceed, code_p, pass_no);
|
|
break;
|
|
case call_op:
|
|
code_p = a_p(_call, &clinfo, code_p, pass_no, cip);
|
|
break;
|
|
case execute_op:
|
|
#if THREADS
|
|
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
|
|
!(cip->CurrentPred->PredFlags & ThreadLocalPredFlag))
|
|
code_p = a_e(_unlock_lu, code_p, pass_no);
|
|
#endif
|
|
code_p = a_p(_execute, &clinfo, code_p, pass_no, cip);
|
|
break;
|
|
case safe_call_op:
|
|
code_p = a_p(_call, &clinfo, code_p, pass_no, cip);
|
|
break;
|
|
case label_op:
|
|
if (!ystop_found &&
|
|
cip->cpc->nextInst != NULL &&
|
|
(cip->cpc->nextInst->op == mark_initialised_pvars_op ||
|
|
cip->cpc->nextInst->op == blob_op)) {
|
|
ystop_found = TRUE;
|
|
code_p = a_e(_Ystop, code_p, pass_no);
|
|
}
|
|
if (!pass_no) {
|
|
if (CellPtr(cip->label_offset+cip->cpc->rnd1) > ASP-256) {
|
|
Yap_Error_Size = 256+((char *)(cip->label_offset+cip->cpc->rnd1) - (char *)H);
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 3);
|
|
}
|
|
|
|
if ( (char *)(cip->label_offset+cip->cpc->rnd1) >= cip->freep)
|
|
cip->freep = (char *)(cip->label_offset+(cip->cpc->rnd1+1));
|
|
cip->label_offset[cip->cpc->rnd1] = (CELL) code_p;
|
|
}
|
|
/* reset dealloc_found in case there was a branch */
|
|
clinfo.dealloc_found = FALSE;
|
|
break;
|
|
case pop_op:
|
|
if (cip->cpc->rnd1 == 1)
|
|
code_p = a_e(_pop, code_p, pass_no);
|
|
else {
|
|
code_p = a_n(_pop_n, 2 * CELLSIZE * (cip->cpc->rnd1 - 1), code_p, pass_no);
|
|
}
|
|
break;
|
|
case either_op:
|
|
code_p = check_alloc(&clinfo, code_p, pass_no, cip);
|
|
#ifdef YAPOR
|
|
if (pass_no)
|
|
either_inst[either_cont++] = code_p;
|
|
code_p = a_either(_either,
|
|
-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2,
|
|
Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], 0, 0, code_p, pass_no, cip);
|
|
#else
|
|
code_p = a_either(_either,
|
|
-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2,
|
|
Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], code_p, pass_no, cip);
|
|
#endif /* YAPOR */
|
|
break;
|
|
case orelse_op:
|
|
#ifdef YAPOR
|
|
if (pass_no)
|
|
either_inst[either_cont++] = code_p;
|
|
code_p = a_either(_or_else,
|
|
-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2,
|
|
Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], 0, 0, code_p, pass_no, cip);
|
|
#else
|
|
code_p = a_either(_or_else,
|
|
-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2,
|
|
Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], code_p, pass_no, cip);
|
|
#endif /* YAPOR */
|
|
clinfo.dealloc_found = FALSE;
|
|
break;
|
|
case orlast_op:
|
|
#ifdef YAPOR
|
|
if (pass_no)
|
|
either_inst[either_cont++] = code_p;
|
|
code_p = a_either(_or_last, 0, 0, 0, 0, code_p, pass_no, cpc);
|
|
if (pass_no) {
|
|
int cont = 1;
|
|
do {
|
|
either_cont--;
|
|
PUT_YAMOP_LTT(either_inst[either_cont], cont++);
|
|
} while (either_inst[either_cont]->opc != opcode(_either));
|
|
}
|
|
#else
|
|
code_p = a_e(_or_last, code_p, pass_no);
|
|
#endif /* YAPOR */
|
|
clinfo.dealloc_found = FALSE;
|
|
break;
|
|
case cache_arg_op:
|
|
code_p = a_4sw_x(_switch_on_arg_type, code_p, pass_no, cip);
|
|
break;
|
|
case cache_sub_arg_op:
|
|
code_p = a_4sw_s(_switch_on_sub_arg_type, code_p, pass_no, cip);
|
|
break;
|
|
case jump_v_op:
|
|
code_p = a_igl(cip->cpc->rnd1, _jump_if_var, code_p, pass_no, cip);
|
|
break;
|
|
case jump_nv_op:
|
|
code_p = a_xigl(_jump_if_nonvar, code_p, pass_no, cip->cpc);
|
|
break;
|
|
case switch_on_type_op:
|
|
code_p = a_4sw(_switch_on_type, code_p, pass_no, cip);
|
|
break;
|
|
case switch_c_op:
|
|
code_p = a_hx(_switch_on_cons, cl_u, log_update, code_p, pass_no, cip);
|
|
break;
|
|
case switch_f_op:
|
|
code_p = a_hx(_switch_on_func, cl_u, log_update, code_p, pass_no, cip);
|
|
break;
|
|
case if_c_op:
|
|
if (cip->cpc->rnd1 == 1) {
|
|
code_p = a_if(_go_on_cons, cl_u, log_update, code_p, pass_no, cip);
|
|
} else {
|
|
code_p = a_if(_if_cons, cl_u, log_update, code_p, pass_no, cip);
|
|
}
|
|
break;
|
|
case if_f_op:
|
|
if (cip->cpc->rnd1 == 1) {
|
|
code_p = a_if(_go_on_func, cl_u, log_update, code_p, pass_no, cip);
|
|
} else {
|
|
code_p = a_if(_if_func, cl_u, log_update, code_p, pass_no, cip);
|
|
}
|
|
break;
|
|
case if_not_op:
|
|
code_p = a_ifnot(_if_not_then, code_p, pass_no, cip);
|
|
break;
|
|
case index_dbref_op:
|
|
code_p = a_e(_index_dbref, code_p, pass_no);
|
|
break;
|
|
case index_blob_op:
|
|
code_p = a_e(_index_blob, code_p, pass_no);
|
|
break;
|
|
case mark_initialised_pvars_op:
|
|
code_p = a_bmap(code_p, pass_no, cip->cpc);
|
|
break;
|
|
case mark_live_regs_op:
|
|
code_p = a_bregs(code_p, pass_no, cip->cpc);
|
|
break;
|
|
case commit_opt_op:
|
|
clinfo.commit_lab = cip->cpc->rnd1;
|
|
break;
|
|
case fetch_args_vv_op:
|
|
a_fetch_vv(&cmp_info, pass_no, cip);
|
|
break;
|
|
case fetch_args_vc_op:
|
|
a_fetch_vc(&cmp_info, pass_no, cip);
|
|
break;
|
|
case fetch_args_cv_op:
|
|
a_fetch_cv(&cmp_info, pass_no, cip);
|
|
break;
|
|
case f_val_op:
|
|
code_p = a_f2(FALSE, &cmp_info, code_p, pass_no, cip);
|
|
break;
|
|
case f_var_op:
|
|
code_p = a_f2(TRUE, &cmp_info, code_p, pass_no, cip);
|
|
break;
|
|
case enter_profiling_op:
|
|
code_p = a_pl(_enter_profiling, (PredEntry *)(cip->cpc->rnd1), code_p, pass_no);
|
|
break;
|
|
case retry_profiled_op:
|
|
code_p = a_pl(_retry_profiled, (PredEntry *)(cip->cpc->rnd1), code_p, pass_no);
|
|
break;
|
|
case count_call_op:
|
|
code_p = a_pl(_count_call, (PredEntry *)(cip->cpc->rnd1), code_p, pass_no);
|
|
break;
|
|
case count_retry_op:
|
|
code_p = a_pl(_count_retry, (PredEntry *)(cip->cpc->rnd1), code_p, pass_no);
|
|
break;
|
|
case fetch_args_for_bccall:
|
|
if (cip->cpc->nextInst->op != bccall_op) {
|
|
Yap_Error(SYSTEM_ERROR, TermNil, "compiling binary test", (int) cip->cpc->op);
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
}
|
|
code_p = a_bfunc(cip->cpc->nextInst->rnd2, &clinfo, code_p, pass_no, cip);
|
|
break;
|
|
case align_float_op:
|
|
/* install a blob */
|
|
#if SIZEOF_DOUBLE == 2*SIZEOF_LONG_INT
|
|
if (!((CELL)code_p & 0x4))
|
|
GONEXT(e);
|
|
#endif
|
|
break;
|
|
case blob_op:
|
|
/* install a blob */
|
|
code_p = copy_blob(code_p, pass_no, cip->cpc);
|
|
break;
|
|
case empty_call_op:
|
|
/* create an empty call */
|
|
code_p = a_empty_call(&clinfo, code_p, pass_no, cip);
|
|
break;
|
|
case push_or_op:
|
|
/* be sure to allocate if we have an ;, even if it is
|
|
compiled inline.
|
|
*/
|
|
code_p = check_alloc(&clinfo, code_p, pass_no, cip);
|
|
case pushpop_or_op:
|
|
case pop_or_op:
|
|
case nop_op:
|
|
case name_op:
|
|
break;
|
|
default:
|
|
Yap_Error(SYSTEM_ERROR, TermNil, "instruction %d found while assembling", (int) cip->cpc->op);
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch, 1);
|
|
}
|
|
cip->cpc = cip->cpc->nextInst;
|
|
}
|
|
if (!ystop_found)
|
|
code_p = a_e(_Ystop, code_p, pass_no);
|
|
return code_p;
|
|
}
|
|
|
|
yamop *
|
|
Yap_assemble(int mode, Term t, PredEntry *ap, int is_fact, struct intermediates *cip)
|
|
{
|
|
/*
|
|
* the assembly proccess is done in two passes: 1 - a first pass
|
|
* computes labels offsets and total code size 2 - the second pass
|
|
* produces the final version of the code
|
|
*/
|
|
UInt size = 0;
|
|
yamop *entry_code;
|
|
yamop *code_p;
|
|
int clause_has_blobs = FALSE;
|
|
|
|
cip->label_offset = (int *)cip->freep;
|
|
cip->code_addr = NULL;
|
|
code_p = do_pass(0, &entry_code, mode, &clause_has_blobs, cip, size);
|
|
size =
|
|
(CELL)NEXTOP(NEXTOP(NEXTOP((yamop *)(((DynamicClause *)NULL)->ClCode),ld),sla),e);
|
|
if ((CELL)code_p > size)
|
|
size = (CELL)code_p;
|
|
if (mode == ASSEMBLING_CLAUSE &&
|
|
ap->PredFlags & LogUpdatePredFlag &&
|
|
!is_fact) {
|
|
DBTerm *x;
|
|
LogUpdClause *cl;
|
|
CELL *h0 = H;
|
|
|
|
H = (CELL *)cip->freep;
|
|
while ((x = Yap_StoreTermInDBPlusExtraSpace(t, size)) == NULL) {
|
|
H = h0;
|
|
if (!Yap_growheap(TRUE, size, cip)) {
|
|
Yap_Error_TYPE = SYSTEM_ERROR;
|
|
return NULL;
|
|
}
|
|
h0 = H;
|
|
H = (CELL *)cip->freep;
|
|
}
|
|
H = h0;
|
|
cl = (LogUpdClause *)((CODEADDR)x-(UInt)size);
|
|
cl->ClSource = x;
|
|
cip->code_addr = (yamop *)cl;
|
|
} else if (mode == ASSEMBLING_CLAUSE &&
|
|
(ap->PredFlags & SourcePredFlag ||
|
|
(!ap->cs.p_code.NOfClauses && yap_flags[SOURCE_MODE_FLAG])) &&
|
|
!is_fact) {
|
|
DBTerm *x;
|
|
StaticClause *cl;
|
|
CELL *h0 = H;
|
|
|
|
H = (CELL *)cip->freep;
|
|
while ((x = Yap_StoreTermInDBPlusExtraSpace(t, size)) == NULL) {
|
|
switch (Yap_Error_TYPE) {
|
|
case OUT_OF_STACK_ERROR:
|
|
Yap_Error_Size = 256+((char *)cip->freep - (char *)H);
|
|
save_machine_regs();
|
|
longjmp(cip->CompilerBotch,3);
|
|
case OUT_OF_TRAIL_ERROR:
|
|
/* don't just return NULL */
|
|
H = h0;
|
|
ARG1 = t;
|
|
if (!Yap_growtrail(64 * 1024L)) {
|
|
return NULL;
|
|
}
|
|
Yap_Error_TYPE = YAP_NO_ERROR;
|
|
t = ARG1;
|
|
h0 = H;
|
|
H = (CELL *)cip->freep;
|
|
break;
|
|
case OUT_OF_HEAP_ERROR:
|
|
/* don't just return NULL */
|
|
H = h0;
|
|
ARG1 = t;
|
|
if (!Yap_growheap(TRUE, size, cip)) {
|
|
return NULL;
|
|
}
|
|
Yap_Error_TYPE = YAP_NO_ERROR;
|
|
t = ARG1;
|
|
h0 = H;
|
|
H = (CELL *)cip->freep;
|
|
break;
|
|
default:
|
|
return NULL;
|
|
}
|
|
}
|
|
H = h0;
|
|
cl = (StaticClause *)((CODEADDR)x-(UInt)size);
|
|
cip->code_addr = (yamop *)cl;
|
|
code_p = do_pass(1, &entry_code, mode, &clause_has_blobs, cip, size);
|
|
/* make sure we copy after second pass */
|
|
cl->usc.ClSource = x;
|
|
ProfEnd=code_p;
|
|
return entry_code;
|
|
} else {
|
|
while ((cip->code_addr = (yamop *) Yap_AllocCodeSpace(size)) == NULL) {
|
|
if (!Yap_growheap(TRUE, size, cip)) {
|
|
Yap_Error_TYPE = SYSTEM_ERROR;
|
|
return NULL;
|
|
}
|
|
}
|
|
}
|
|
code_p = do_pass(1, &entry_code, mode, &clause_has_blobs, cip, size);
|
|
ProfEnd=code_p;
|
|
return entry_code;
|
|
}
|
|
|
|
void
|
|
Yap_InitComma(void)
|
|
{
|
|
yamop *code_p = COMMA_CODE;
|
|
code_p->opc = opcode(_call);
|
|
code_p->u.sla.s = emit_count(-Signed(RealEnvSize) - sizeof(CELL) * 3);
|
|
code_p->u.sla.sla_u.p =
|
|
code_p->u.sla.p0 =
|
|
RepPredProp(PredPropByFunc(FunctorComma,0));
|
|
code_p->u.sla.bmap = NULL;
|
|
GONEXT(sla);
|
|
if (PRED_GOAL_EXPANSION_ON) {
|
|
Functor fp = Yap_MkFunctor(Yap_FullLookupAtom("$generate_pred_info"),4);
|
|
code_p->opc = emit_op(_call_cpred);
|
|
code_p->u.sla.s = emit_count(-Signed(RealEnvSize));
|
|
code_p->u.sla.sla_u.p = RepPredProp(Yap_GetPredPropByFunc(fp,0));
|
|
code_p->u.sla.bmap = NULL;
|
|
GONEXT(sla);
|
|
code_p->opc = emit_op(_call);
|
|
code_p->u.sla.s = emit_count(-Signed(RealEnvSize));
|
|
code_p->u.sla.sla_u.p = PredMetaCall;
|
|
code_p->u.sla.bmap = NULL;
|
|
GONEXT(sla);
|
|
code_p->opc = emit_op(_deallocate);
|
|
GONEXT(e);
|
|
code_p->opc = emit_op(_procceed);
|
|
GONEXT(e);
|
|
} else {
|
|
if (PROFILING) {
|
|
code_p->opc = opcode(_enter_a_profiling);
|
|
GONEXT(e);
|
|
}
|
|
if (CALL_COUNTING) {
|
|
code_p->opc = opcode(_count_a_call);
|
|
GONEXT(e);
|
|
}
|
|
code_p->opc = opcode(_p_execute_tail);
|
|
GONEXT(e);
|
|
}
|
|
}
|