This repository has been archived on 2023-08-20. You can view files and clone it, but cannot push or open issues or pull requests.
yap-6.3/C/amasm.c
2004-04-22 20:07:07 +00:00

2931 lines
80 KiB
C

/*************************************************************************
* *
* YAP Prolog *
* *
* Yap Prolog was developed at NCCUP - Universidade do Porto *
* *
* Copyright L.Damas, V.S.Costa and Universidade do Porto 1985-1997 *
* *
**************************************************************************
* *
* File: amasm.c *
* comments: abstract machine assembler *
* *
* Last rev: $Date: 2004-04-22 20:07:04 $ *
* $Log: not supported by cvs2svn $
* Revision 1.59 2004/03/31 01:03:09 vsc
* support expand group of clauses
*
* Revision 1.58 2004/03/10 14:59:55 vsc
* optimise -> for type tests
* *
* *
*************************************************************************/
#ifdef SCCS
static char SccsId[] = "@(#)amasm.c 1.3 3/15/90";
#endif
#include "Yap.h"
#include "yapio.h"
#include "compile.h"
#include "clause.h"
#ifdef YAPOR
#include "or.macros.h"
#endif /* YAPOR */
#if HAVE_STRING_H
#include <string.h>
#endif
/* info on compare built-ins */
#define TYPE_XX 0
#define TYPE_CX 1
#define TYPE_XC 2
typedef struct cmp_op_info_struct {
wamreg x1_arg, x2_arg;
Int c_arg;
int c_type;
struct clause_info_struct *cl_info;
} cmp_op_info;
typedef struct clause_info_struct {
int alloc_found, dealloc_found;
CELL commit_lab;
struct pred_entry *CurrentPred;
} clause_info;
STATIC_PROTO(OPREG Var_Ref, (Ventry *, int));
STATIC_PROTO(wamreg emit_xreg, (CELL));
STATIC_PROTO(yslot emit_yreg, (CELL));
STATIC_PROTO(wamreg emit_x, (CELL));
STATIC_PROTO(yslot emit_y, (Ventry *));
STATIC_PROTO(yamop *emit_a, (CELL));
STATIC_PROTO(CELL *emit_bmlabel, (CELL, struct intermediates *));
STATIC_PROTO(yamop *emit_ilabel, (CELL, struct intermediates *));
STATIC_PROTO(Functor emit_f, (CELL));
STATIC_PROTO(CELL emit_c, (CELL));
STATIC_PROTO(COUNT emit_count, (CELL));
STATIC_PROTO(OPCODE emit_op, (op_numbers));
STATIC_PROTO(yamop *a_cl, (op_numbers, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_cle, (op_numbers, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_e, (op_numbers, yamop *, int));
STATIC_PROTO(yamop *a_ue, (op_numbers, op_numbers, yamop *, int));
STATIC_PROTO(yamop *a_v, (op_numbers, yamop *, int, struct PSEUDO *));
STATIC_PROTO(yamop *a_uv, (Ventry *,op_numbers, op_numbers, yamop *, int));
STATIC_PROTO(yamop *a_vr, (op_numbers, yamop *, int, struct PSEUDO *));
STATIC_PROTO(yamop *a_rv, (op_numbers, OPREG, yamop *, int, struct PSEUDO *));
STATIC_PROTO(yamop *a_vv, (op_numbers, op_numbers, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_glist, (int *, yamop *, int, struct intermediates *));
STATIC_PROTO(void a_pair, (CELL *, int, struct intermediates *));
STATIC_PROTO(yamop *a_f, (CELL, op_numbers, yamop *, int));
STATIC_PROTO(yamop *a_c, (CELL, op_numbers, yamop *, int));
STATIC_PROTO(yamop *a_uc, (CELL, op_numbers, op_numbers, yamop *, int));
STATIC_PROTO(yamop *a_n, (op_numbers, int, yamop *, int));
STATIC_PROTO(yamop *a_un, (op_numbers, op_numbers, int, yamop *, int));
STATIC_PROTO(yamop *a_nc, (CELL, op_numbers, int, yamop *, int));
STATIC_PROTO(yamop *a_unc, (CELL, op_numbers, op_numbers, int, yamop *, int));
STATIC_PROTO(yamop *a_r, (CELL, op_numbers, yamop *, int));
STATIC_PROTO(yamop *a_p, (op_numbers, clause_info *, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_pl, (op_numbers, PredEntry *, yamop *, int));
STATIC_PROTO(yamop *a_l, (CELL, op_numbers, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_hx, (op_numbers, union clause_obj *, int, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_if, (op_numbers, union clause_obj *, int, yamop *, int, struct intermediates *cip));
STATIC_PROTO(yamop *a_cut, (clause_info *,yamop *, int, struct intermediates *));
#ifdef YAPOR
STATIC_PROTO(yamop *a_try, (op_numbers, CELL, CELL, *clause_info, int, int, yamop *, int));
STATIC_PROTO(yamop *a_either, (op_numbers, CELL, CELL, int, int, yamop *, int, struct intermediates *));
#else
STATIC_PROTO(yamop *a_try, (op_numbers, CELL, CELL, clause_info *, yamop *, int));
STATIC_PROTO(yamop *a_either, (op_numbers, CELL, CELL, yamop *, int, struct intermediates *));
#endif /* YAPOR */
STATIC_PROTO(yamop *a_gl, (op_numbers, clause_info *, yamop *, int, struct PSEUDO *));
STATIC_PROTO(yamop *a_bfunc, (CELL, clause_info *, yamop *, int, struct intermediates *));
STATIC_PROTO(wamreg compile_cmp_flags, (char *));
STATIC_PROTO(yamop *a_igl, (CELL, op_numbers, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_xigl, (op_numbers, yamop *, int, struct PSEUDO *));
STATIC_PROTO(yamop *a_ucons, (int *, compiler_vm_op, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_uvar, (yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_wvar, (yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *do_pass, (int, yamop **, int, int *, struct intermediates *, UInt));
#ifdef DEBUG_OPCODES
STATIC_PROTO(void DumpOpCodes, (void));
#endif
#ifdef SFUNC
STATIC_PROTO(void a_vsf, (int, yamop *, int, struct PSEUDO *));
STATIC_PROTO(void a_asf, (int, yamop *, int, struct PSEUDO *));
#endif
STATIC_PROTO(yamop *check_alloc, (clause_info *, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_deallocate, (clause_info *, yamop *, int, struct intermediates *));
STATIC_PROTO(yamop *a_bmap, (yamop *, int, struct PSEUDO *));
STATIC_PROTO(void a_fetch_vv, (cmp_op_info *, int, struct intermediates *));
STATIC_PROTO(void a_fetch_cv, (cmp_op_info *, int, struct intermediates *));
STATIC_PROTO(void a_fetch_vc, (cmp_op_info *, int, struct intermediates *));
STATIC_PROTO(yamop *a_f2, (int, cmp_op_info *, yamop *, int, struct intermediates *));
#define CELLSIZE sizeof(CELL)
#define GONEXT(TYPE) code_p = ((yamop *)(&(code_p->u.TYPE.next)))
inline static yslot
emit_y(Ventry *ve)
{
#if MSHIFTOFFS
return(-FixedEnvSize - ((ve->NoOfVE) & MaskVarAdrs) - 1);
#else
return(-FixedEnvSize - (((ve->NoOfVE) & MaskVarAdrs) * CELLSIZE) - CELLSIZE);
#endif
}
inline static OPREG
Var_Ref(Ventry *ve, int is_y_var)
{
if (is_y_var) {
#if MSHIFTOFFS
return -FixedEnvSize - ((ve->NoOfVE) & MaskVarAdrs) - 1;
#else
return -FixedEnvSize - (((ve->NoOfVE) & MaskVarAdrs) * CELLSIZE) - CELLSIZE;
#endif
}
else {
#if PRECOMPUTE_REGADDRESS
return (CELL) (XREGS + ((ve->NoOfVE) & MaskVarAdrs));
#else
#if MSHIFTOFFS
return ((ve->NoOfVE) & MaskVarAdrs);
#else
return CELLSIZE * ((ve->NoOfVE) & MaskVarAdrs);
#endif
#endif /* PRECOMPUTE_REGADDRESS */
}
}
#define is_void_var() (((Ventry *) (cip->cpc->rnd1))->KindOfVE == VoidVar)
#define is_a_void(X) (((Ventry *) (X))->KindOfVE == VoidVar)
#define is_temp_var() (((Ventry *) (cip->cpc->rnd1))->KindOfVE == TempVar)
#define is_atemp_var(p) (((Ventry *) (p->rnd1))->KindOfVE == TempVar)
#define no_ref_var() (((Ventry *) (cip->cpc->rnd1))->NoOfVE == 1)
#define no_ref(X) (((Ventry *) (X))->NoOfVE == 1)
inline static yamop *
fill_small(CELL w, yamop *code_p, int pass_no)
{
SMALLUNSGN *ptr = ((SMALLUNSGN *) (code_p));
if (pass_no)
*ptr = (SMALLUNSGN) w;
return (yamop *) (++ptr);
}
inline static yamop *
fill_a(CELL a, yamop *code_p, int pass_no)
{
CELL *ptr = ((CELL *) (code_p));
if (pass_no)
*ptr = a;
return (yamop *) (++ptr);
}
inline static wamreg
emit_xreg(CELL w)
{
return (wamreg) w;
}
inline static yslot
emit_yreg(CELL w)
{
return (yslot) w;
}
inline static wamreg
emit_x(CELL xarg)
{
#if PRECOMPUTE_REGADDRESS
return (emit_xreg((CELL) (XREGS + xarg)));
#else
#if MSHIFTOFFS
return (emit_xreg(xarg));
#else
return (emit_xreg(CELLSIZE * (xarg)));
#endif
#endif /* PRECOMPUTE_REGADDRESS */
}
wamreg
Yap_emit_x(CELL xarg)
{
return emit_x(xarg);
}
inline static yamop *
emit_a(CELL a)
{
return ((yamop *) (a));
}
inline static struct pred_entry *
emit_pe(struct pred_entry *a)
{
return a;
}
inline static yamop *
emit_ilabel(register CELL addr, struct intermediates *cip)
{
if (addr & 1)
return (emit_a(Unsigned(cip->code_addr) + cip->label_offset[addr]));
else {
return (emit_a(addr));
}
}
inline static CELL *
emit_bmlabel(register CELL addr, struct intermediates *cip)
{
return (CELL *)(emit_a(Unsigned(cip->code_addr) + cip->label_offset[addr]));
}
inline static Functor
emit_f(CELL a)
{
return (Functor) (a);
}
inline static CELL
emit_c(CELL a)
{
return a;
}
static inline COUNT
emit_count(CELL count)
{
return count;
}
#ifdef DEBUG_OPCODES
inline static void
DumpOpCodes(void)
{
int i = 0, j;
while (i < 30) {
for (j = i; j <= _std_top; j += 25)
fprintf(Yap_stderr, "%5d %6lx", j, absmadr(j));
fputc('\n',Yap_stderr);
++i;
}
}
#endif
static inline OPCODE
emit_op(op_numbers op)
{
return absmadr((Int) op);
}
static OPCODE
opcode(op_numbers op)
{
return (emit_op(op));
}
OPCODE
Yap_opcode(op_numbers op)
{
return opcode(op);
}
static void
add_clref(CELL clause_code, int pass_no)
{
if (pass_no) {
LogUpdClause *cl = ClauseCodeToLogUpdClause(clause_code);
cl->ClRefCount++;
}
}
static yamop *
a_cl(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.l.l = cip->code_addr;
}
GONEXT(l);
return code_p;
}
static yamop *
a_lucl(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.Ill.I = (LogUpdIndex *)cip->code_addr;
code_p->u.Ill.l1 = emit_ilabel(cip->cpc->rnd1, cip);
code_p->u.Ill.l2 = emit_ilabel(cip->cpc->rnd2, cip);
code_p->u.Ill.s = cip->cpc->rnd3;
}
GONEXT(Ill);
return code_p;
}
static yamop *
a_cle(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (pass_no) {
LogUpdClause *cl = (LogUpdClause *)cip->code_addr;
code_p->opc = emit_op(opcode);
code_p->u.EC.ClTrail = 0;
code_p->u.EC.ClENV = 0;
code_p->u.EC.ClRefs = 0;
code_p->u.EC.ClBase = cip->code_addr;
cl->ClExt = code_p;
cl->ClFlags |= LogUpdRuleMask;
}
GONEXT(EC);
return code_p;
}
inline static yamop *
a_e(op_numbers opcode, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
}
GONEXT(e);
return code_p;
}
inline static yamop *
a_ue(op_numbers opcode, op_numbers opcodew, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.o.opcw = emit_op(opcodew);
}
GONEXT(o);
return code_p;
}
inline static yamop *
a_v(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
Ventry *ve = (Ventry *) cpc->rnd1;
OPREG var_offset;
int is_y_var = (ve->KindOfVE == PermVar);
var_offset = Var_Ref(ve, is_y_var);
if (ve->KindOfVE == PermVar) {
if (pass_no) {
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.y.y = emit_yreg(var_offset);
}
GONEXT(y);
}
else {
if (pass_no) {
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.x.x = emit_xreg(var_offset);
}
GONEXT(x);
}
return code_p;
}
inline static yamop *
a_uv(Ventry *ve, op_numbers opcode, op_numbers opcodew, yamop *code_p, int pass_no)
{
OPREG var_offset;
int is_y_var = (ve->KindOfVE == PermVar);
var_offset = Var_Ref(ve, is_y_var);
if (is_y_var) {
if (pass_no) {
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.oy.opcw = emit_op((op_numbers)((int)opcodew + is_y_var));
code_p->u.oy.y = emit_yreg(var_offset);
}
GONEXT(oy);
}
else {
if (pass_no) {
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.ox.opcw = emit_op((op_numbers)((int)opcodew + is_y_var));
code_p->u.ox.x = emit_xreg(var_offset);
}
GONEXT(ox);
}
return code_p;
}
inline static yamop *
a_vv(op_numbers opcode, op_numbers opcodew, yamop *code_p, int pass_no, struct intermediates *cip)
{
Ventry *ve = (Ventry *) cip->cpc->rnd1;
int is_y_var = (ve->KindOfVE == PermVar);
if (pass_no) {
OPREG var_offset = Var_Ref(ve, is_y_var);
code_p->opc = emit_op(opcode);
code_p->u.oxx.opcw = emit_op(opcodew);
code_p->u.oxx.xl = emit_xreg(var_offset);
}
cip->cpc = cip->cpc->nextInst;
if (pass_no) {
OPREG var_offset;
int is_y_var;
ve = (Ventry *) cip->cpc->rnd1;
is_y_var = (ve->KindOfVE == PermVar);
var_offset = Var_Ref(ve, is_y_var);
code_p->u.oxx.xr = emit_xreg(var_offset);
}
GONEXT(oxx);
return code_p;
}
inline static yamop *
a_vr(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
Ventry *ve = (Ventry *) cpc->rnd1;
int is_y_var = (ve->KindOfVE == PermVar);
if (is_y_var) {
if (pass_no) {
OPREG var_offset;
var_offset = Var_Ref(ve, is_y_var);
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.yx.y = emit_yreg(var_offset);
code_p->u.yx.x = emit_x(cpc->rnd2);
}
GONEXT(yx);
}
else {
if (pass_no) {
OPREG var_offset;
var_offset = Var_Ref(ve, is_y_var);
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.xx.xl = emit_xreg(var_offset);
code_p->u.xx.xr = emit_x(cpc->rnd2);
}
GONEXT(xx);
}
return code_p;
}
inline static yamop *
a_rv(op_numbers opcode, OPREG var_offset, yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
Ventry *ve = (Ventry *) cpc->rnd1;
int is_y_var = (ve->KindOfVE == PermVar);
if (is_y_var) {
if (pass_no) {
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.xy.x = emit_x(cpc->rnd2);
code_p->u.xy.y = emit_yreg(var_offset);
}
GONEXT(xy);
}
else {
if (pass_no) {
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.xx.xl = emit_x(cpc->rnd2);
code_p->u.xx.xr = emit_xreg(var_offset);
}
GONEXT(xx);
}
return code_p;
}
#ifdef SFUNC
/* vsc: I don't understand these instructions */
inline static void
a_vsf(int opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
Ventry *ve = (Ventry *) cpc->rnd1;
OPREG var_offset;
int is_y_var = (ve->KindOfVE == PermVar);
var_offset = Var_Ref(ve, is_y_var);
if (is_y_var) {
if (pass_no) {
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.fy.f = emit_f(cpc->rnd2);
code_p->u.fy.a = ArityOfFunctor(emit_f(cpc->rnd2));
code_p->u.fy.y = emit_yreg(var_offset);
}
GONEXT(fy);
}
else {
if (pass_no) {
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.fx.f = emit_f(cpc->rnd2);
code_p->u.fx.a = ArityOfFunctor(emit_f(cpc->rnd2));
code_p->u.fx.x = emit_xreg(var_offset);
}
GONEXT(fx);
}
return code_p;
}
inline static void
a_asf(int opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
if (pass_no) {
code_p->opc = emit_op((op_numbers)((int)opcode + is_y_var));
code_p->u.fn.f = emit_f(cpc->rnd2);
code_p->u.fn.a = ArityOfFunctor(emit_f(cpc->rnd2));
code_p->u.fn.n = emit_count(cpc->rnd1);
}
GONEXT(fn);
return code_p;
}
#endif
inline static void
a_pair(CELL *seq_ptr, int pass_no, struct intermediates *cip)
{
if (pass_no) {
CELL lab, lab0 = seq_ptr[1];
lab = (CELL) emit_ilabel(lab0, cip);
seq_ptr[0] = (CELL) emit_a(seq_ptr[0]);
seq_ptr[1] = lab;
}
}
inline static yamop *
a_n(op_numbers opcode, int count, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.s.s = count;
}
GONEXT(s);
return code_p;
}
inline static yamop *
a_un(op_numbers opcode, op_numbers opcodew, int count, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.os.opcw = emit_op(opcodew);
code_p->u.os.s = count;
}
GONEXT(os);
return code_p;
}
inline static yamop *
a_f(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no)
{
if (pass_no) {
Functor f = emit_f(rnd1);
code_p->opc = emit_op(opcode);
code_p->u.f.f = f;
code_p->u.f.a = ArityOfFunctor(f);
}
GONEXT(f);
return code_p;
}
inline static yamop *
a_uf(CELL rnd1, op_numbers opcode, op_numbers opcodew, yamop *code_p, int pass_no)
{
if (pass_no) {
Functor f = emit_f(rnd1);
code_p->opc = emit_op(opcode);
code_p->u.of.opcw = emit_op(opcodew);
code_p->u.of.f = f;
code_p->u.of.a = ArityOfFunctor(f);
}
GONEXT(of);
return code_p;
}
inline static yamop *
a_c(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.c.c = emit_c(rnd1);
}
GONEXT(c);
return code_p;
}
inline static yamop *
a_uc(CELL rnd1, op_numbers opcode, op_numbers opcode_w, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.oc.opcw = emit_op(opcode_w);
code_p->u.oc.c = emit_c(rnd1);
}
GONEXT(oc);
return code_p;
}
inline static yamop *
a_blob(CELL rnd1, op_numbers opcode, int *clause_has_blobsp, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.c.c =
AbsAppl((CELL *)(Unsigned(cip->code_addr) + cip->label_offset[rnd1]));
}
*clause_has_blobsp = TRUE;
GONEXT(c);
return code_p;
}
inline static yamop *
a_ublob(CELL rnd1, op_numbers opcode, op_numbers opcode_w, int *clause_has_blobsp, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.oc.opcw = emit_op(opcode_w);
code_p->u.oc.c =
AbsAppl((CELL *)(Unsigned(cip->code_addr) + cip->label_offset[rnd1]));
}
*clause_has_blobsp = TRUE;
GONEXT(oc);
return code_p;
}
inline static yamop *
a_nc(CELL rnd1, op_numbers opcode, int i, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.sc.s = i;
code_p->u.sc.c = emit_c(rnd1);
}
GONEXT(sc);
return code_p;
}
inline static yamop *
a_unc(CELL rnd1, op_numbers opcode, op_numbers opcodew, int i, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.osc.opcw = emit_op(opcodew);
code_p->u.osc.s = i;
code_p->u.osc.c = emit_c(rnd1);
}
GONEXT(osc);
return code_p;
}
inline static yamop *
a_rf(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.xf.x = emit_x(cpc->rnd2);
code_p->u.xf.f = emit_f(cpc->rnd1);
code_p->u.xf.a = ArityOfFunctor(emit_f(cpc->rnd1));
}
GONEXT(xf);
return code_p;
}
inline static yamop *
a_rc(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.xc.x = emit_x(cpc->rnd2);
code_p->u.xc.c = emit_c(cpc->rnd1);
}
GONEXT(xc);
return code_p;
}
inline static yamop *
a_rb(op_numbers opcode, int *clause_has_blobsp, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.xc.x = emit_x(cip->cpc->rnd2);
code_p->u.xc.c = AbsAppl((CELL *)(Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1]));
}
*clause_has_blobsp = TRUE;
GONEXT(xc);
return code_p;
}
inline static yamop *
a_r(CELL arnd2, op_numbers opcode, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.x.x = emit_x(arnd2);
}
GONEXT(x);
return code_p;
}
static yamop *
check_alloc(clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (clinfo->alloc_found == 2) {
if (clinfo->CurrentPred->PredFlags & LogUpdatePredFlag)
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
code_p = a_e(_allocate, code_p, pass_no);
clinfo->alloc_found = 1;
}
return code_p;
}
static yamop *
a_p(op_numbers opcode, clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
{ /* emit opcode & predicate code address */
Prop fe = (Prop) (cip->cpc->rnd1);
CELL Flags = RepPredProp(fe)->PredFlags;
if (Flags & AsmPredFlag) {
op_numbers op;
code_p = check_alloc(clinfo, code_p, pass_no, cip);
switch (Flags & 0x7f) {
case _equal:
op = _p_equal;
break;
case _dif:
op = _p_dif;
break;
case _eq:
op = _p_eq;
break;
case _functor:
op = _p_functor;
break;
default:
op = _p_equal; /* just to make some compilers happy */
Yap_Error(SYSTEM_ERROR, TermNil, "internal assembler error for built-in (%d)", (Flags & 0x7f));
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
}
return a_e(op, code_p, pass_no);
}
if (Flags & CPredFlag) {
code_p = check_alloc(clinfo, code_p, pass_no, cip);
if (clinfo->commit_lab && (Flags & TestPredFlag)) {
if (pass_no) {
if (Flags & UserCPredFlag) {
Yap_Error(SYSTEM_ERROR, TermNil,
"user defined predicate cannot be a test predicate");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
} else
code_p->opc = emit_op(_call_c_wfail);
code_p->u.sdl.s =
emit_count(-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2);
code_p->u.sdl.l =
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
code_p->u.sdl.p =
emit_pe(RepPredProp(fe));
}
GONEXT(sdl);
clinfo->commit_lab = 0;
} else {
if (pass_no) {
if (Flags & UserCPredFlag) {
code_p->opc = emit_op(_call_usercpred);
} else {
if (RepPredProp(fe)->FunctorOfPred == FunctorExecuteInMod) {
code_p->opc = emit_op(_p_execute);
} else {
code_p->opc = emit_op(_call_cpred);
}
}
code_p->u.sla.s = emit_count(-Signed(RealEnvSize) - CELLSIZE
* (cip->cpc->rnd2));
if (RepPredProp(fe)->FunctorOfPred != FunctorExecuteInMod) {
code_p->u.sla.sla_u.p = RepPredProp(fe);
} else {
code_p->u.sla.sla_u.mod = cip->cpc->rnd4;
}
code_p->u.sla.p0 = clinfo->CurrentPred;
if (cip->cpc->rnd2) {
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->arnds[1], cip);
} else {
/* there is no bitmap as there are no variables in the environment */
code_p->u.sla.bmap = NULL;
}
}
GONEXT(sla);
}
return code_p;
}
if (opcode == _call && clinfo->alloc_found == 2) {
if (clinfo->CurrentPred->PredFlags & LogUpdatePredFlag)
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
if (pass_no) {
code_p->opc = emit_op(_fcall);
}
clinfo->alloc_found = 1;
}
else {
code_p = check_alloc(clinfo, code_p, pass_no, cip);
if (pass_no)
code_p->opc = emit_op(opcode);
}
if (opcode == _call) {
if (pass_no) {
code_p->u.sla.s = emit_count(-Signed(RealEnvSize) - CELLSIZE *
cip->cpc->rnd2);
code_p->u.sla.sla_u.p = RepPredProp(fe);
code_p->u.sla.p0 = clinfo->CurrentPred;
if (cip->cpc->rnd2)
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->arnds[1], cip);
else
/* there is no bitmap as there are no variables in the environment */
code_p->u.sla.bmap = NULL;
}
GONEXT(sla);
}
else {
if (pass_no)
code_p->u.p.p = RepPredProp(fe);
GONEXT(p);
}
return code_p;
}
/*
emit a false call so that the garbage collector and friends will find
reasonable information on the stack.
*/
static yamop *
a_empty_call(clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (clinfo->alloc_found == 1 && !clinfo->dealloc_found) {
/* we have a solid environment under us, just trust it */
if (pass_no)
code_p->opc = emit_op(_call);
}
else {
/** oops, our environment is crap */
if (pass_no)
code_p->opc = emit_op(_fcall);
}
if (pass_no) {
PredEntry *pe = RepPredProp(Yap_GetPredPropByAtom(AtomTrue,0));
code_p->u.sla.s = emit_count(-Signed(RealEnvSize) - CELLSIZE *
cip->cpc->rnd2);
code_p->u.sla.sla_u.p = pe;
code_p->u.sla.p0 = clinfo->CurrentPred;
if (cip->cpc->rnd2)
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->rnd1, cip);
else
/* there is no bitmap as there are no variables in the environment */
code_p->u.sla.bmap = NULL;
}
GONEXT(sla);
return code_p;
}
static yamop *
a_l(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.l.l = emit_a(Unsigned(cip->code_addr) + cip->label_offset[rnd1]);
}
GONEXT(l);
return code_p;
}
static yamop *
a_il(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.l.l = emit_ilabel(rnd1, cip);
}
GONEXT(l);
return code_p;
}
static yamop *
a_pl(op_numbers opcode, PredEntry *pred, yamop *code_p, int pass_no)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.p.p = pred;
}
GONEXT(p);
return code_p;
}
static wamreg
compile_cmp_flags(char *s)
{
if (strcmp(s,"=<") == 0)
return(EQ_OK_IN_CMP|LT_OK_IN_CMP);
if (strcmp(s,"<") == 0)
return(LT_OK_IN_CMP);
if (strcmp(s,">=") == 0)
return(EQ_OK_IN_CMP|GT_OK_IN_CMP);
if (strcmp(s,">") == 0)
return(GT_OK_IN_CMP);
if (strcmp(s,"=:=") == 0)
return(EQ_OK_IN_CMP);
if (strcmp(s,"=\\=") == 0)
return(GT_OK_IN_CMP|LT_OK_IN_CMP);
Yap_Error(SYSTEM_ERROR, TermNil, "internal assembler error in flags for %s", s);
return(0);
}
wamreg
Yap_compile_cmp_flags(PredEntry *pred)
{
return compile_cmp_flags(RepAtom(NameOfFunctor(pred->FunctorOfPred))->StrOfAE);
}
static yamop *
a_bfunc(CELL pred, clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
{
Ventry *ve = (Ventry *) cip->cpc->rnd1;
OPREG var_offset;
int is_y_var = (ve->KindOfVE == PermVar);
var_offset = Var_Ref(ve, is_y_var);
if (ve->KindOfVE == PermVar) {
yslot v1 = emit_yreg(var_offset);
cip->cpc = cip->cpc->nextInst;
ve = (Ventry *) cip->cpc->rnd1;
is_y_var = (ve->KindOfVE == PermVar);
var_offset = Var_Ref(ve, is_y_var);
if (is_y_var) {
if (pass_no) {
code_p->opc = emit_op(_call_bfunc_yy);
code_p->u.llyy.p = RepPredProp(((Prop)pred));
if (clinfo->commit_lab) {
code_p->u.llyy.f =
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
clinfo->commit_lab = 0;
} else {
code_p->u.llyy.f = FAILCODE;
}
code_p->u.llyy.y1 = v1;
code_p->u.llyy.y2 = emit_yreg(var_offset);
code_p->u.llyy.flags = compile_cmp_flags(RepAtom(NameOfFunctor(RepPredProp(((Prop)pred))->FunctorOfPred))->StrOfAE);
}
GONEXT(llyy);
} else {
if (pass_no) {
code_p->opc = emit_op(_call_bfunc_yx);
code_p->u.llxy.p = RepPredProp(((Prop)pred));
if (clinfo->commit_lab) {
code_p->u.llxy.f =
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
clinfo->commit_lab = 0;
} else {
code_p->u.llxy.f = FAILCODE;
}
code_p->u.llxy.x = emit_xreg(var_offset);
code_p->u.llxy.y = v1;
code_p->u.llxy.flags = compile_cmp_flags(RepAtom(NameOfFunctor(RepPredProp(((Prop)pred))->FunctorOfPred))->StrOfAE);
}
GONEXT(llxy);
}
} else {
wamreg x1 = emit_xreg(var_offset);
OPREG var_offset;
cip->cpc = cip->cpc->nextInst;
ve = (Ventry *) cip->cpc->rnd1;
is_y_var = (ve->KindOfVE == PermVar);
var_offset = Var_Ref(ve, is_y_var);
if (is_y_var) {
if (pass_no) {
code_p->opc = emit_op(_call_bfunc_xy);
code_p->u.llxy.p = RepPredProp(((Prop)pred));
if (clinfo->commit_lab) {
code_p->u.llxy.f =
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
clinfo->commit_lab = 0;
} else {
code_p->u.llxy.f = FAILCODE;
}
code_p->u.llxy.x = x1;
code_p->u.llxy.y = emit_yreg(var_offset);
code_p->u.llxy.flags = compile_cmp_flags(RepAtom(NameOfFunctor(RepPredProp(((Prop)pred))->FunctorOfPred))->StrOfAE);
}
GONEXT(llxy);
} else {
if (pass_no) {
code_p->opc = emit_op(_call_bfunc_xx);
code_p->u.llxx.p = RepPredProp(((Prop)pred));
if (clinfo->commit_lab) {
code_p->u.llxx.f =
emit_a(Unsigned(cip->code_addr) + cip->label_offset[clinfo->commit_lab]);
clinfo->commit_lab = 0;
} else {
code_p->u.llxx.f = FAILCODE;
}
code_p->u.llxx.x1 = x1;
code_p->u.llxx.x2 = emit_xreg(var_offset);
code_p->u.llxx.flags = compile_cmp_flags(RepAtom(NameOfFunctor(RepPredProp(((Prop)pred))->FunctorOfPred))->StrOfAE);
}
GONEXT(llxx);
}
}
return code_p;
}
static yamop *
a_igl(CELL rnd1, op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.l.l = emit_ilabel(rnd1, cip);
}
GONEXT(l);
return code_p;
}
static yamop *
a_xigl(op_numbers opcode, yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.xl.x = emit_x(cpc->rnd2);
code_p->u.xl.l = emit_a(cpc->rnd1);
}
GONEXT(xl);
return code_p;
}
static yamop *
a_4sw(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
CELL *seq_ptr;
if (opcode == _switch_on_type &&
cip->cpc->nextInst != NULL &&
cip->cpc->nextInst->op == label_op &&
cip->cpc->arnds[1] == cip->cpc->nextInst->rnd1 &&
!(cip->cpc->arnds[0] & 1) &&
cip->cpc->nextInst->nextInst != NULL &&
cip->cpc->nextInst->nextInst->op == if_c_op &&
cip->cpc->nextInst->nextInst->rnd1 == 1 &&
cip->cpc->nextInst->nextInst->arnds[1] == TermNil &&
cip->cpc->nextInst->nextInst->arnds[0] == cip->cpc->arnds[2]) {
if (pass_no) {
code_p->opc = emit_op(_switch_list_nl);
seq_ptr = cip->cpc->arnds;
code_p->u.ollll.pop = ((yamop *)(seq_ptr[0]))->opc;
code_p->u.ollll.l1 = emit_ilabel(seq_ptr[0], cip);
code_p->u.ollll.l2 = emit_ilabel(cip->cpc->nextInst->nextInst->arnds[2], cip);
code_p->u.ollll.l3 = emit_ilabel(seq_ptr[2], cip);
code_p->u.ollll.l4 = emit_ilabel(seq_ptr[3], cip);
}
GONEXT(ollll);
cip->cpc = cip->cpc->nextInst->nextInst;
} else {
if (pass_no) {
code_p->opc = emit_op(opcode);
seq_ptr = cip->cpc->arnds;
code_p->u.llll.l1 = emit_ilabel(seq_ptr[0], cip);
code_p->u.llll.l2 = emit_ilabel(seq_ptr[1], cip);
code_p->u.llll.l3 = emit_ilabel(seq_ptr[2], cip);
code_p->u.llll.l4 = emit_ilabel(seq_ptr[3], cip);
}
GONEXT(llll);
}
return code_p;
}
static yamop *
a_4sw_x(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
CELL *seq_ptr;
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.xllll.x = emit_x(cip->cpc->rnd2);
cip->cpc = cip->cpc->nextInst;
seq_ptr = cip->cpc->arnds;
code_p->u.xllll.l1 = emit_ilabel(seq_ptr[0], cip);
code_p->u.xllll.l2 = emit_ilabel(seq_ptr[1], cip);
code_p->u.xllll.l3 = emit_ilabel(seq_ptr[2], cip);
code_p->u.xllll.l4 = emit_ilabel(seq_ptr[3], cip);
} else {
/* skip one */
cip->cpc = cip->cpc->nextInst;
}
GONEXT(xllll);
return code_p;
}
static yamop *
a_4sw_s(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
CELL *seq_ptr;
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.sllll.s = cip->cpc->rnd2;
cip->cpc = cip->cpc->nextInst;
seq_ptr = cip->cpc->arnds;
code_p->u.sllll.l1 = emit_ilabel(seq_ptr[0], cip);
code_p->u.sllll.l2 = emit_ilabel(seq_ptr[1], cip);
code_p->u.sllll.l3 = emit_ilabel(seq_ptr[2], cip);
code_p->u.sllll.l4 = emit_ilabel(seq_ptr[3], cip);
} else {
/* skip one */
cip->cpc = cip->cpc->nextInst;
}
GONEXT(sllll);
return code_p;
}
static void
init_log_upd_table(LogUpdIndex *ic, union clause_obj *cl_u)
{
/* insert myself in the indexing code chain */
ic->SiblingIndex = cl_u->lui.ChildIndex;
ic->ChildIndex = NULL;
ic->ClRefCount = 0;
ic->u.ParentIndex = (LogUpdIndex *)cl_u;
INIT_LOCK(ic->ClLock);
cl_u->lui.ChildIndex = ic;
cl_u->lui.ClRefCount++;
}
static void
init_static_table(StaticIndex *ic, union clause_obj *cl_u)
{
/* insert myself in the indexing code chain */
ic->SiblingIndex = cl_u->si.ChildIndex;
ic->ChildIndex = NULL;
cl_u->si.ChildIndex = ic;
}
static yamop *
a_hx(op_numbers opcode, union clause_obj *cl_u, int log_update, yamop *code_p, int pass_no, struct intermediates *cip)
{
register CELL i, imax;
register CELL *seq_ptr = (CELL *)cip->cpc->rnd2;
int j = 0;
imax = cip->cpc->rnd1;
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.sssl.s = emit_c(imax);
code_p->u.sssl.l = emit_a(cip->cpc->rnd2);
if (log_update) {
init_log_upd_table(ClauseCodeToLogUpdIndex(cip->cpc->rnd2), cl_u);
} else {
init_static_table(ClauseCodeToStaticIndex(cip->cpc->rnd2), cl_u);
}
}
if (pass_no) {
for (i = 0; i < imax; i++) {
yamop *ipc = (yamop *)seq_ptr[1];
a_pair(seq_ptr, pass_no, cip);
if (ipc != FAILCODE) {
j++;
}
seq_ptr += 2;
}
code_p->u.sssl.e = j;
code_p->u.sssl.w = 0;
}
GONEXT(sssl);
return code_p;
}
static yamop *
a_if(op_numbers opcode, union clause_obj *cl_u, int log_update, yamop *code_p, int pass_no, struct intermediates *cip)
{
register CELL i, imax;
register CELL *seq_ptr = (CELL *)cip->cpc->rnd2;
imax = cip->cpc->rnd1;
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.sssl.s = code_p->u.sssl.e = emit_count(imax);
code_p->u.sssl.w = 0;
code_p->u.sssl.l = emit_a(cip->cpc->rnd2);
if (log_update) {
init_log_upd_table(ClauseCodeToLogUpdIndex(cip->cpc->rnd2), cl_u);
} else {
init_static_table(ClauseCodeToStaticIndex(cip->cpc->rnd2), cl_u);
}
}
GONEXT(sssl);
if (pass_no) {
CELL lab, lab0;
for (i = 0; i < imax; i++) {
a_pair(seq_ptr, pass_no, cip);
seq_ptr += 2;
}
lab0 = seq_ptr[1];
lab = (CELL) emit_ilabel(lab0, cip);
seq_ptr[1] = lab;
}
return code_p;
}
static yamop *
a_ifnot(op_numbers opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
CELL *seq_ptr = cip->cpc->arnds;
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.clll.c = seq_ptr[0]; /* tag */
code_p->u.clll.l1 = emit_ilabel(seq_ptr[1], cip); /* success point */
code_p->u.clll.l2 = emit_ilabel(seq_ptr[2], cip); /* fail point */
code_p->u.clll.l3 = emit_ilabel(seq_ptr[3], cip); /* delay point */
}
GONEXT(clll);
return code_p;
}
static yamop *
a_cut(clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
{
code_p = check_alloc(clinfo, code_p, pass_no, cip);
if (clinfo->dealloc_found) {
return a_e(_cut_e, code_p, pass_no);
} else if (clinfo->alloc_found) {
return a_e(_cut, code_p, pass_no);
} else {
return a_e(_cut_t, code_p, pass_no);
}
}
static yamop *
#ifdef YAPOR
a_try(op_numbers opcode, CELL lab, CELL opr, clause_info *clinfo, int nofalts, int hascut, yamop *code_p, int pass_no)
#else
a_try(op_numbers opcode, CELL lab, CELL opr, clause_info *clinfo, yamop *code_p, int pass_no)
#endif /* YAPOR */
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.ld.d = emit_a(lab);
code_p->u.ld.s = emit_count(opr);
code_p->u.ld.p = clinfo->CurrentPred;
#ifdef TABLING
code_p->u.ld.te = clinfo->CurrentPred->TableOfPred;
#endif
#ifdef YAPOR
INIT_YAMOP_LTT(code_p, nofalts);
if (hascut)
PUT_YAMOP_CUT(code_p);
if (clinfo->CurrentPred->PredFlags & SequentialPredFlag)
PUT_YAMOP_SEQ(code_p);
#endif /* YAPOR */
}
GONEXT(ld);
return code_p;
}
static yamop *
#ifdef YAPOR
a_either(op_numbers opcode, CELL opr, CELL lab, int nofalts, int hascut, yamop *code_p, int pass_no, struct intermediates *cip)
#else
a_either(op_numbers opcode, CELL opr, CELL lab, yamop *code_p, int pass_no, struct intermediates *cip)
#endif /* YAPOR */
{
if (pass_no) {
code_p->opc = emit_op(opcode);
code_p->u.sla.s = emit_count(opr);
code_p->u.sla.sla_u.l = emit_a(lab);
code_p->u.sla.p0 = cip->CurrentPred;
#ifdef YAPOR
INIT_YAMOP_LTT(code_p, nofalts);
if (hascut)
PUT_YAMOP_CUT(code_p);
if (clinfo->CurrentPred->PredFlags & SequentialPredFlag)
PUT_YAMOP_SEQ(code_p);
if(opcode != _or_last) {
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->arnds[1], cip);
}
#else
code_p->u.sla.bmap = emit_bmlabel(cip->cpc->arnds[1], cip);
#endif /* YAPOR */
}
GONEXT(sla);
return code_p;
}
static yamop *
a_gl(op_numbers opcode, clause_info *clinfo, yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
#ifdef YAPOR
return a_try(opcode, cpc->rnd1, IPredArity, clinfo, cpc->rnd2 >> 1, cpc->rnd2 & 1, code_p, pass_no);
#else
return a_try(opcode, cpc->rnd1, IPredArity, clinfo, code_p, pass_no);
#endif /* YAPOR */
}
/*
* optimizes several unify_cons for the same constant. It must be avoided for
* the head of the first argument, because of indexing
*/
static yamop *
a_ucons(int *do_not_optimise_uatomp, compiler_vm_op opcode, yamop *code_p, int pass_no, struct intermediates *cip)
{
#if AGGREGATE_OPS
PInstr *np = cip->cpc->nextInst;
register int i = 0;
CELL my_cons = cip->cpc->rnd1;
if (*do_not_optimise_uatomp) {
*do_not_optimise_uatomp = FALSE;
if (opcode == unify_atom_op)
return a_uc(cip->cpc->rnd1, _unify_atom, _unify_atom_write, code_p, pass_no);
else
return a_c(cip->cpc->rnd1, _write_atom, code_p, pass_no);
}
else {
while (np->op == opcode && np->rnd1 == my_cons) {
i++;
cip->cpc = np;
np = np->nextInst;
}
if (i == 0) {
if (opcode == unify_atom_op)
return a_uc(cip->cpc->rnd1, _unify_atom, _unify_atom_write, code_p, pass_no);
else
return a_c(cip->cpc->rnd1, _write_atom, code_p, pass_no);
} else {
if (opcode == unify_atom_op)
return a_unc(cip->cpc->rnd1, _unify_n_atoms, _unify_n_atoms_write, i + 1, code_p, pass_no);
else
return a_nc(cip->cpc->rnd1, _write_n_atoms, i + 1, code_p, pass_no);
}
}
#else
*do_not_optimise_uatomp = FALSE;
if (opcode == unify_atom_op)
return a_uc(cip->cpc->rnd1, _unify_atom, _unify_atom_write, code_p);
else
return a_c(cip->cpc->rnd1, _write_atom, code_p);
#endif
}
static yamop *
a_uvar(yamop *code_p, int pass_no, struct intermediates *cip)
{
if (!is_void_var()) {
#if AGGREGATE_OPS
if (is_temp_var()) {
PInstr *np = cip->cpc->nextInst;
if (np->op == unify_var_op &&
is_atemp_var(np)) {
return a_vv(_unify_x_var2, _unify_x_var2_write, code_p, pass_no, cip);
}
else if (np->op == unify_last_var_op &&
is_atemp_var(np)) {
return a_vv(_unify_l_x_var2,
_unify_l_x_var2_write, code_p, pass_no, cip);
}
}
#endif /* AGGREGATE_OPS */
return a_uv((Ventry *) cip->cpc->rnd1, _unify_x_var, _unify_x_var_write, code_p, pass_no);
}
else {
#if AGGREGATE_OPS
int i = 1;
PInstr *np = cip->cpc->nextInst;
/* skip void vars */
while (np->op == unify_var_op && is_a_void(np->rnd1)) {
i++;
cip->cpc = np;
np = np->nextInst;
}
if (np->op == unify_last_var_op &&
is_a_void(np->rnd1)) {
if (i == 0)
code_p = a_ue(_unify_l_void, _unify_l_void_write, code_p, pass_no);
else
code_p = a_un(_unify_l_n_voids, _unify_l_n_voids_write, i + 1, code_p, pass_no);
cip->cpc = np;
}
else if (i == 1)
return a_ue(_unify_void, _unify_void_write, code_p, pass_no);
else {
return a_un(_unify_n_voids, _unify_n_voids_write, i, code_p, pass_no);
}
#else
return a_ue(_unify_void, _unify_void_write);
#endif
}
return code_p;
}
static yamop *
a_wvar(yamop *code_p, int pass_no, struct intermediates *cip)
{
if (!no_ref_var())
return a_v(_write_x_var, code_p, pass_no, cip->cpc);
else {
#if AGGREGATE_OPS
int i = 0;
PInstr *np = cip->cpc->nextInst;
while (np->op == write_var_op && no_ref(np->rnd1)) {
i++;
cip->cpc = np;
np = np->nextInst;
}
if (i == 0)
return a_e(_write_void, code_p, pass_no);
else {
return a_n(_write_n_voids, i + 1, code_p, pass_no);
}
#else
return a_e(_write_void, pass_no);
#endif
}
}
static yamop *
a_glist(int *do_not_optimise_uatomp, yamop *code_p, int pass_no, struct intermediates *cip)
{
#if AGGREGATE_OPS
PInstr *pnext = cip->cpc->nextInst;
if (cip->cpc->rnd2 != 1 && pnext->op == unify_val_op) {
Ventry *ve = (Ventry *) pnext->rnd1;
int is_y_var;
OPREG var_offset;
pnext->rnd2 = cip->cpc->rnd2;
cip->cpc = pnext;
is_y_var = (ve->KindOfVE == PermVar);
var_offset = Var_Ref(ve, is_y_var);
return a_rv(_glist_valx, var_offset, code_p, pass_no, cip->cpc);
} else if (cip->cpc->rnd2 == 1 && pnext->op == unify_atom_op) {
*do_not_optimise_uatomp = TRUE;
return a_r(cip->cpc->rnd2, _get_list, code_p, pass_no);
} else if (cip->cpc->rnd2 != 1 && pnext->op == unify_var_op
&& is_a_void(pnext->rnd1)) {
PInstr *ppnext = pnext->nextInst;
if (ppnext && (ppnext->op == unify_last_var_op
|| ppnext->op == unify_last_val_op)) {
Ventry *ve = (Ventry *) ppnext->rnd1;
int is_y_var = (ve->KindOfVE == PermVar);
OPREG var_offset;
ppnext->rnd2 = cip->cpc->rnd2;
cip->cpc = ppnext;
var_offset = Var_Ref(ve, is_y_var);
return a_rv((op_numbers)((int)_gl_void_varx + (cip->cpc->op == unify_last_var_op ? 0 : 2)), var_offset, code_p, pass_no, cip->cpc);
} else {
return a_r(cip->cpc->rnd2, _get_list, code_p, pass_no);
}
} else
#endif /* AGGREGATE_OPS */
return a_r(cip->cpc->rnd2, _get_list, code_p, pass_no);
}
#define NEXTOPC (cip->cpc->nextInst)->op
static yamop *
a_deallocate(clause_info *clinfo, yamop *code_p, int pass_no, struct intermediates *cip)
{
if (clinfo->alloc_found == 2) {
/* this should never happen */
if (clinfo->CurrentPred->PredFlags & LogUpdatePredFlag)
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
code_p = a_e(_allocate, code_p, pass_no);
}
if (NEXTOPC == execute_op) {
cip->cpc = cip->cpc->nextInst;
code_p = a_p(_dexecute, clinfo, code_p, pass_no, cip);
} else
code_p = a_e(_deallocate, code_p, pass_no);
clinfo->dealloc_found = TRUE;
return code_p;
}
static yamop *
a_bmap(yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
/* how much space do we need to reserve */
int i, max = (cpc->rnd1)/(8*sizeof(CELL));
for (i = 0; i <= max; i++)
code_p = fill_a(cpc->arnds[i], code_p, pass_no);
return code_p;
}
static yamop *
a_bregs(yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
/* how much space do we need to reserve */
int i, max = (cpc->rnd1)/(8*sizeof(CELL));
code_p = fill_a(cpc->rnd1, code_p, pass_no);
for (i = 0; i <= max; i++)
code_p = fill_a(cpc->arnds[i], code_p, pass_no);
return code_p;
}
static yamop *
copy_blob(yamop *code_p, int pass_no, struct PSEUDO *cpc)
{
/* copy the blob to code space, making no effort to align if a double */
int max = cpc->rnd1, i;
for (i = 0; i < max; i++)
code_p = fill_a(cpc->arnds[i], code_p, pass_no);
return code_p;
}
static void
a_fetch_vv(cmp_op_info *cmp_info, int pass_no, struct intermediates *cip)
{
/* the next three instructions must be a get_val, get_val, and BIP */
if (pass_no == 0) {
PInstr *p = cip->cpc->nextInst;
Ventry *ve;
ve = (Ventry *) p->rnd1;
if (ve->KindOfVE != PermVar)
p->op = nop_op;
p = p->nextInst;
ve = (Ventry *) p->rnd1;
if (ve->KindOfVE != PermVar)
p->op = nop_op;
} else {
PInstr *p = cip->cpc->nextInst;
Ventry *ve;
cmp_info->c_type = TYPE_XX;
ve = (Ventry *) p->rnd1;
if (ve->KindOfVE == PermVar) {
/* don't get rid of get_val_op */
cmp_info->x1_arg = emit_x(p->rnd2);
} else {
/* and use it directly as a second argument */
cmp_info->x1_arg = emit_x(ve->NoOfVE & MaskVarAdrs);
}
p = p->nextInst;
ve = (Ventry *) p->rnd1;
if (ve->KindOfVE == PermVar) {
/* don't get rid of get_val_op */
cmp_info->x2_arg = emit_x(p->rnd2);
} else {
/* and use it directly as a second argument */
cmp_info->x2_arg = emit_x(ve->NoOfVE & MaskVarAdrs);
}
}
}
static void
a_fetch_vc(cmp_op_info *cmp_info, int pass_no, struct intermediates *cip)
{
/* the next two instructions must be a get_val and BIP */
if (pass_no == 0) {
PInstr *p = cip->cpc->nextInst;
Ventry *ve;
ve = (Ventry *) p->rnd1;
if (ve->KindOfVE != PermVar)
p->op = nop_op;
} else {
PInstr *p = cip->cpc->nextInst;
Ventry *ve;
cmp_info->c_type = TYPE_XC;
cmp_info->c_arg = (Int)(cip->cpc->rnd1);
ve = (Ventry *) p->rnd1;
if (ve->KindOfVE == PermVar) {
/* don't get rid of get_val_op */
cmp_info->x1_arg = emit_x(p->rnd2);
} else {
/* and use it directly as a second argument */
cmp_info->x1_arg = emit_x(ve->NoOfVE & MaskVarAdrs);
}
}
}
static void
a_fetch_cv(cmp_op_info *cmp_info, int pass_no, struct intermediates *cip)
{
/* the next two instructions must be a get_val and BIP */
if (pass_no == 0) {
PInstr *p = cip->cpc->nextInst;
Ventry *ve;
ve = (Ventry *) p->rnd1;
if (ve->KindOfVE != PermVar)
p->op = nop_op;
} else {
PInstr *p = cip->cpc->nextInst;
Ventry *ve;
cmp_info->c_type = TYPE_CX;
cmp_info->c_arg = (Int)(cip->cpc->rnd1);
ve = (Ventry *) p->rnd1;
if (ve->KindOfVE == PermVar) {
/* don't get rid of get_val_op */
cmp_info->x1_arg = emit_x(p->rnd2);
} else {
/* and use it directly as a second argument */
cmp_info->x1_arg = emit_x(ve->NoOfVE & MaskVarAdrs);
}
}
}
static yamop *
a_f2(int var, cmp_op_info *cmp_info, yamop *code_p, int pass_no, struct intermediates *cip)
{
Int opc = cip->cpc->rnd2;
Ventry *ve = (Ventry *)(cip->cpc->rnd1);
int is_y_var = (ve->KindOfVE == PermVar);
if (opc <= _primitive) {
if (is_y_var) {
if (pass_no) {
code_p->u.yF.y = emit_y(ve);
switch (opc) {
case _atom:
code_p->opc = opcode(_p_atom_y);
break;
case _atomic:
code_p->opc = opcode(_p_atomic_y);
break;
case _compound:
code_p->opc = opcode(_p_compound_y);
break;
case _float:
code_p->opc = opcode(_p_float_y);
break;
case _integer:
code_p->opc = opcode(_p_integer_y);
break;
case _nonvar:
code_p->opc = opcode(_p_nonvar_y);
break;
case _number:
code_p->opc = opcode(_p_number_y);
break;
case _var:
code_p->opc = opcode(_p_var_y);
break;
case _db_ref:
code_p->opc = opcode(_p_db_ref_y);
break;
case _cut_by:
code_p->opc = opcode(_p_cut_by_y);
break;
case _primitive:
code_p->opc = opcode(_p_primitive_y);
break;
}
if (cmp_info->cl_info->commit_lab) {
code_p->u.yF.F =
emit_a(Unsigned(cip->code_addr) + cip->label_offset[cmp_info->cl_info->commit_lab]);
cmp_info->cl_info->commit_lab = 0;
} else {
code_p->u.yF.F = FAILCODE;
}
}
GONEXT(yF);
return code_p;
} else {
if (pass_no) {
code_p->u.xF.x = emit_x(ve->NoOfVE & MaskVarAdrs);
switch (opc) {
case _atom:
code_p->opc = opcode(_p_atom_x);
break;
case _atomic:
code_p->opc = opcode(_p_atomic_x);
break;
case _compound:
code_p->opc = opcode(_p_compound_x);
break;
case _float:
code_p->opc = opcode(_p_float_x);
break;
case _integer:
code_p->opc = opcode(_p_integer_x);
break;
case _nonvar:
code_p->opc = opcode(_p_nonvar_x);
break;
case _number:
code_p->opc = opcode(_p_number_x);
break;
case _var:
code_p->opc = opcode(_p_var_x);
break;
case _db_ref:
code_p->opc = opcode(_p_db_ref_x);
break;
case _cut_by:
code_p->opc = opcode(_p_cut_by_x);
break;
case _primitive:
code_p->opc = opcode(_p_primitive_x);
break;
}
if (cmp_info->cl_info->commit_lab) {
code_p->u.xF.F =
emit_a(Unsigned(cip->code_addr) + cip->label_offset[cmp_info->cl_info->commit_lab]);
cmp_info->cl_info->commit_lab = 0;
} else {
code_p->u.xF.F = FAILCODE;
}
}
GONEXT(xF);
return code_p;
}
}
if (opc == _functor && cip->cpc->nextInst->op == f_var_op) {
Ventry *nve;
cip->cpc = cip->cpc->nextInst;
nve = (Ventry *)(cip->cpc->rnd1);
if (is_y_var) {
if (nve->KindOfVE == PermVar) {
if (pass_no) {
code_p->opc = emit_op(_p_func2f_yy);
code_p->u.yyx.y1 = emit_y(ve);
code_p->u.yyx.y2 = emit_y(nve);
code_p->u.yyx.x = cmp_info->x1_arg;
}
GONEXT(yyx);
return code_p;
} else {
if (pass_no) {
code_p->opc = emit_op(_p_func2f_yx);
code_p->u.yxx.y = emit_y(ve);
code_p->u.yxx.x1 = emit_x(nve->NoOfVE & MaskVarAdrs);
code_p->u.yxx.x2 = cmp_info->x1_arg;
}
GONEXT(yxx);
return code_p;
}
} else {
if (nve->KindOfVE == PermVar) {
if (pass_no) {
code_p->opc = emit_op(_p_func2f_xy);
code_p->u.xyx.x1 = emit_x(ve->NoOfVE & MaskVarAdrs);
code_p->u.xyx.y2 = emit_y(nve);
code_p->u.xyx.x = cmp_info->x1_arg;
}
GONEXT(xyx);
return code_p;
} else {
if (pass_no) {
code_p->opc = emit_op(_p_func2f_xx);
code_p->u.xxx.x1 = emit_x(ve->NoOfVE & MaskVarAdrs);
code_p->u.xxx.x2 = emit_x(nve->NoOfVE & MaskVarAdrs);
code_p->u.xxx.x = cmp_info->x1_arg;
}
GONEXT(xxx);
return code_p;
}
}
}
if (is_y_var) {
switch (cmp_info->c_type) {
case TYPE_XX:
if (pass_no) {
switch (opc) {
case _plus:
code_p->opc = emit_op(_p_plus_y_vv);
break;
case _minus:
code_p->opc = emit_op(_p_minus_y_vv);
break;
case _times:
code_p->opc = emit_op(_p_times_y_vv);
break;
case _div:
code_p->opc = emit_op(_p_div_y_vv);
break;
case _and:
code_p->opc = emit_op(_p_and_y_vv);
break;
case _or:
code_p->opc = emit_op(_p_or_y_vv);
break;
case _sll:
code_p->opc = emit_op(_p_sll_y_vv);
break;
case _slr:
code_p->opc = emit_op(_p_slr_y_vv);
break;
case _arg:
code_p->opc = emit_op(_p_arg_y_vv);
break;
case _functor:
code_p->opc = emit_op(_p_func2s_y_vv);
break;
}
code_p->u.yxx.y = emit_y(ve);
code_p->u.yxx.x1 = cmp_info->x1_arg;
code_p->u.yxx.x2 = cmp_info->x2_arg;
}
GONEXT(yxx);
break;
case TYPE_CX:
if (pass_no) {
switch (opc) {
case _plus:
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for +/2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _minus:
code_p->opc = emit_op(_p_minus_y_cv);
break;
case _times:
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for */2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _div:
code_p->opc = emit_op(_p_div_y_cv);
break;
case _and:
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for /\\/2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _or:
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for \\//2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _sll:
code_p->opc = emit_op(_p_sll_y_cv);
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _slr:
code_p->opc = emit_op(_p_slr_y_cv);
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _arg:
code_p->opc = emit_op(_p_arg_y_cv);
break;
case _functor:
code_p->opc = emit_op(_p_func2s_y_cv);
break;
}
code_p->u.ycx.y = emit_y(ve);
code_p->u.ycx.c = cmp_info->c_arg;
code_p->u.ycx.xi = cmp_info->x1_arg;
}
GONEXT(ycx);
break;
case TYPE_XC:
if (pass_no) {
switch (opc) {
case _plus:
code_p->opc = emit_op(_p_plus_y_vc);
break;
case _minus:
Yap_Error(SYSTEM_ERROR, cmp_info->x2_arg, "internal assembler error XC for -/2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _times:
code_p->opc = emit_op(_p_times_y_vc);
break;
case _div:
code_p->opc = emit_op(_p_div_y_vc);
break;
case _and:
code_p->opc = emit_op(_p_and_y_vc);
break;
case _or:
code_p->opc = emit_op(_p_or_y_vc);
break;
case _sll:
code_p->opc = emit_op(_p_sll_y_vc);
break;
case _slr:
code_p->opc = emit_op(_p_slr_y_vc);
break;
case _arg:
Yap_Error(SYSTEM_ERROR, cmp_info->x2_arg, "internal assembler error for arg/3");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _functor:
code_p->opc = emit_op(_p_func2s_y_vc);
break;
}
code_p->u.yxc.y = emit_y(ve);
code_p->u.yxc.c = cmp_info->c_arg;
code_p->u.yxc.xi = cmp_info->x1_arg;
}
GONEXT(yxc);
break;
}
} else {
switch (cmp_info->c_type) {
case TYPE_XX:
if (pass_no) {
switch (opc) {
case _plus:
code_p->opc = emit_op(_p_plus_vv);
break;
case _minus:
code_p->opc = emit_op(_p_minus_vv);
break;
case _times:
code_p->opc = emit_op(_p_times_vv);
break;
case _div:
code_p->opc = emit_op(_p_div_vv);
break;
case _and:
code_p->opc = emit_op(_p_and_vv);
break;
case _or:
code_p->opc = emit_op(_p_or_vv);
break;
case _sll:
code_p->opc = emit_op(_p_sll_vv);
break;
case _slr:
code_p->opc = emit_op(_p_slr_vv);
break;
case _arg:
code_p->opc = emit_op(_p_arg_vv);
break;
case _functor:
code_p->opc = emit_op(_p_func2s_vv);
break;
}
code_p->u.xxx.x = emit_x(ve->NoOfVE & MaskVarAdrs);
code_p->u.xxx.x1 = cmp_info->x1_arg;
code_p->u.xxx.x2 = cmp_info->x2_arg;
}
GONEXT(xxx);
break;
case TYPE_CX:
if (pass_no) {
switch (opc) {
case _plus:
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for +/2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _minus:
code_p->opc = emit_op(_p_minus_cv);
break;
case _times:
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for */2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _div:
code_p->opc = emit_op(_p_div_cv);
break;
case _and:
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for /\\/2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _or:
Yap_Error(SYSTEM_ERROR, cmp_info->x1_arg, "internal assembler error CX for \\//2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _sll:
code_p->opc = emit_op(_p_sll_cv);
break;
case _slr:
code_p->opc = emit_op(_p_slr_cv);
break;
case _arg:
code_p->opc = emit_op(_p_arg_cv);
break;
case _functor:
code_p->opc = emit_op(_p_func2s_cv);
break;
}
code_p->u.xxc.x = emit_x(ve->NoOfVE & MaskVarAdrs);
code_p->u.xxc.c = cmp_info->c_arg;
code_p->u.xxc.xi = cmp_info->x1_arg;
}
GONEXT(xxc);
break;
case TYPE_XC:
if (pass_no) {
switch (opc) {
case _plus:
code_p->opc = emit_op(_p_plus_vc);
break;
case _minus:
Yap_Error(SYSTEM_ERROR, cmp_info->x2_arg, "internal assembler error XC for -/2");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _times:
code_p->opc = emit_op(_p_times_vc);
break;
case _div:
code_p->opc = emit_op(_p_div_vc);
break;
case _and:
code_p->opc = emit_op(_p_and_vc);
break;
case _or:
code_p->opc = emit_op(_p_or_vc);
break;
case _sll:
code_p->opc = emit_op(_p_sll_vc);
break;
case _slr:
code_p->opc = emit_op(_p_slr_vc);
break;
case _arg:
Yap_Error(SYSTEM_ERROR, cmp_info->x2_arg, "internal assembler error for arg/3");
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
break;
case _functor:
code_p->opc = emit_op(_p_func2s_vc);
break;
}
code_p->u.xcx.x = emit_x(ve->NoOfVE & MaskVarAdrs);
code_p->u.xcx.c = cmp_info->c_arg;
code_p->u.xcx.xi = cmp_info->x1_arg;
}
GONEXT(xcx);
break;
}
}
return code_p;
}
#define TRYOP(G,P) (IPredArity<5 ? (op_numbers)((int)(P)+(IPredArity*3)) : (G))
#ifdef YAPOR
#define TRYCODE(G,P) a_try(TRYOP(G,P), Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], IPredArity, &clinfo, cip->cpc->rnd2 >> 1, cip->cpc->rnd2 & 1, code_p, pass_no)
#define TABLE_TRYCODE(G) a_try(G, (CELL)emit_ilabel(cip->cpc->rnd1, code_addr, cip), IPredArity, cip->cpc->rnd2 >> 1, cip->cpc->rnd2 & 1, code_p, pass_no)
#else
#define TRYCODE(G,P) a_try(TRYOP(G,P), Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], IPredArity, &clinfo, code_p, pass_no)
#define TABLE_TRYCODE(G) a_try(G, (CELL)emit_ilabel(cip->cpc->rnd1, code_addr, cip), IPredArity, code_p, pass_no)
#endif /* YAPOR */
static yamop *
do_pass(int pass_no, yamop **entry_codep, int assembling, int *clause_has_blobsp, struct intermediates *cip, UInt size)
{
#ifdef YAPOR
#define EITHER_INST 50
yamop *either_inst[EITHER_INST];
int either_cont = 0;
#endif /* YAPOR */
int log_update;
int dynamic;
#ifdef TABLING
int tabled;
#endif
int ystop_found = FALSE;
union clause_obj *cl_u;
yamop *code_p;
cmp_op_info cmp_info;
clause_info clinfo;
int do_not_optimise_uatom;
code_p = cip->code_addr;
cl_u = (union clause_obj *)code_p;
cip->cpc = cip->CodeStart;
clinfo.alloc_found = clinfo.dealloc_found = FALSE;
clinfo.commit_lab = 0L;
clinfo.CurrentPred = cip->CurrentPred;
cmp_info.c_type = TYPE_XX;
cmp_info.cl_info = &clinfo;
do_not_optimise_uatom = FALSE;
/* Space while for the clause flags */
log_update = cip->CurrentPred->PredFlags & LogUpdatePredFlag;
dynamic = cip->CurrentPred->PredFlags & DynamicPredFlag;
#ifdef TABLING
tabled = cip->CurrentPred->PredFlags & TabledPredFlag;
#endif
if (assembling == ASSEMBLING_CLAUSE) {
if (log_update) {
if (pass_no) {
cl_u->luc.Id = FunctorDBRef;
cl_u->luc.ClFlags = LogUpdMask;
cl_u->luc.ClRefCount = 0;
cl_u->luc.ClPred = cip->CurrentPred;
cl_u->luc.ClSize = size;
if (*clause_has_blobsp) {
cl_u->luc.ClFlags |= HasBlobsMask;
}
cl_u->luc.ClExt = NULL;
cl_u->luc.ClPrev = cl_u->luc.ClNext = NULL;
#if defined(YAPOR) || defined(THREADS)
INIT_LOCK(cl_u->luc.ClLock);
INIT_CLREF_COUNT(&(cl_u->luc));
#endif
}
code_p = cl_u->luc.ClCode;
} else if (dynamic) {
if (pass_no) {
cl_u->ic.ClFlags = DynamicMask;
if (*clause_has_blobsp) {
cl_u->ic.ClFlags |= HasBlobsMask;
}
cl_u->ic.ClSize = size;
cl_u->ic.ClRefCount = 0;
#if defined(YAPOR) || defined(THREADS)
INIT_LOCK(cl_u->ic.ClLock);
INIT_CLREF_COUNT(&(cl_u->ic));
#endif
}
code_p = cl_u->ic.ClCode;
} else {
/* static clause */
if (pass_no) {
cl_u->sc.Id = FunctorDBRef;
cl_u->sc.ClFlags = StaticMask;
cl_u->sc.ClNext = NULL;
cl_u->sc.ClSize = size;
if (*clause_has_blobsp) {
cl_u->sc.ClFlags |= HasBlobsMask;
}
}
code_p = cl_u->sc.ClCode;
}
IPredArity = cip->cpc->rnd2; /* number of args */
*entry_codep = code_p;
if (dynamic) {
#ifdef YAPOR
a_try(TRYOP(_try_me, _try_me0), 0, IPredArity, &clinfo, 1, 0, code_p, pass_no);
#else
a_try(TRYOP(_try_me, _try_me0), 0, IPredArity, &clinfo, code_p, pass_no);
#endif /* YAPOR */
}
} else {
/* index code */
if (log_update) {
if (pass_no) {
cl_u->lui.ClFlags = LogUpdMask|IndexedPredFlag|IndexMask|SwitchRootMask;
cl_u->lui.ChildIndex = NULL;
cl_u->lui.SiblingIndex = NULL;
cl_u->lui.u.pred = cip->CurrentPred;
cl_u->lui.ClSize = size;
cl_u->lui.ClRefCount = 0;
INIT_LOCK(cl_u->lui.ClLock);
#if defined(YAPOR) || defined(THREADS)
INIT_CLREF_COUNT(&(cl_u->lui));
#endif
}
code_p = cl_u->lui.ClCode;
*entry_codep = code_p;
#if defined(YAPOR) || defined(THREADS)
if (assembling == ASSEMBLING_INDEX &&
!(cip->CurrentPred->PredFlags & ThreadLocalPredFlag)) {
if (pass_no) {
code_p->opc = opcode(_lock_lu);
code_p->u.p.p = cip->CurrentPred;
}
GONEXT(p);
}
#endif
} else {
if (pass_no) {
cl_u->si.ClSize = size;
cl_u->si.ClFlags = IndexMask;
cl_u->si.ChildIndex = NULL;
cl_u->si.SiblingIndex = NULL;
}
code_p = cl_u->si.ClCode;
*entry_codep = code_p;
}
}
while (cip->cpc) {
switch ((int) cip->cpc->op) {
#ifdef YAPOR
case sync_op:
code_p = a_try(_sync, cip->cpc->rnd1, cip->cpc->rnd2, 1, Zero, code_p);
break;
#endif /* YAPOR */
#ifdef TABLING
case table_new_answer_op:
code_p = a_n(_table_new_answer, (int) cip->cpc->rnd2, code_p);
break;
case table_try_single_op:
code_p = a_gl(_table_try_single, code_p, clinfo, code_p, pass_no, cpc);
break;
#endif /* TABLING */
#ifdef TABLING_INNER_CUTS
case clause_with_cut_op:
code_p = a_e(_clause_with_cut, code_p);
break;
#endif /* TABLING_INNER_CUTS */
#ifdef SFUNC
case get_s_f_op:
code_p = a_rf(_get_s_f, code_p, cip->cpc);
break;
case put_s_f_op:
code_p = a_rf(_put_s_f, code_p, cip->cpc);
break;
case unify_s_f_op:
code_p = a_d(_unify_s_f, code_p);
break;
case write_s_f_op:
code_p = a_f(cip->cpc->rnd1, _write_s_f);
break;
case unify_s_var_op:
code_p = a_vsf(_unify_s_xvar);
break;
case write_s_var_op:
code_p = a_vsf(_write_s_xvar);
break;
case unify_s_val_op:
code_p = a_vsf(_unify_s_xval);
break;
case write_s_val_op:
code_p = a_vsf(_write_s_xval);
break;
case unify_s_a_op:
code_p = a_asf(_unify_s_a);
break;
case write_s_a_op:
code_p = a_asf(_write_s_a);
break;
case get_s_end_op:
code_p = a_n(_get_s_end, Unsigned(0));
break;
case put_s_end_op:
code_p = a_n(_put_s_end, Unsigned(0));
break;
case unify_s_end_op:
code_p = a_n(_write_s_end, Unsigned(0));
break;
case write_s_end_op:
code_p = a_n(_write_s_end, Unsigned(0));
break;
#endif
case get_var_op:
code_p = a_vr(_get_x_var, code_p, pass_no, cip->cpc);
break;
case put_var_op:
code_p = a_vr(_put_x_var, code_p, pass_no, cip->cpc);
break;
case get_val_op:
code_p = a_vr(_get_x_val, code_p, pass_no, cip->cpc);
break;
case put_val_op:
code_p = a_vr(_put_x_val, code_p, pass_no, cip->cpc);
break;
case get_num_op:
case get_atom_op:
code_p = a_rc(_get_atom, code_p, pass_no, cip->cpc);
break;
case get_float_op:
code_p = a_rb(_get_float, clause_has_blobsp, code_p, pass_no, cip);
break;
case get_longint_op:
code_p = a_rb(_get_longint, clause_has_blobsp, code_p, pass_no, cip);
break;
case get_bigint_op:
code_p = a_rb(_get_bigint, clause_has_blobsp, code_p, pass_no, cip);
break;
case put_num_op:
case put_atom_op:
code_p = a_rc(_put_atom, code_p, pass_no, cip->cpc);
break;
case put_float_op:
case put_longint_op:
case put_bigint_op:
code_p = a_rb(_put_atom, clause_has_blobsp, code_p, pass_no, cip);
break;
case get_list_op:
code_p = a_glist(&do_not_optimise_uatom, code_p, pass_no, cip);
break;
case put_list_op:
code_p = a_r(cip->cpc->rnd2, _put_list, code_p, pass_no);
break;
case get_struct_op:
code_p = a_rf(_get_struct, code_p, pass_no, cip->cpc);
break;
case put_struct_op:
code_p = a_rf(_put_struct, code_p, pass_no, cip->cpc);
break;
case put_unsafe_op:
code_p = a_vr((op_numbers)((int)_put_unsafe - 1), code_p, pass_no, cip->cpc);
break;
case unify_var_op:
code_p = a_uvar(code_p, pass_no, cip);
break;
case unify_last_var_op:
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_l_x_var, _unify_l_x_var_write, code_p, pass_no);
break;
case write_var_op:
code_p = a_wvar(code_p, pass_no, cip);
break;
case unify_local_op:
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_x_loc, _unify_x_loc_write, code_p, pass_no);
break;
case unify_val_op:
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_x_val, _unify_x_val_write, code_p, pass_no);
break;
case unify_last_local_op:
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_l_x_loc, _unify_l_x_loc_write, code_p, pass_no);
break;
case unify_last_val_op:
code_p = a_uv((Ventry *) cip->cpc->rnd1, _unify_l_x_val, _unify_l_x_val_write, code_p, pass_no);
break;
case write_local_op:
code_p = a_v(_write_x_loc, code_p, pass_no, cip->cpc);
break;
case write_val_op:
code_p = a_v(_write_x_val, code_p, pass_no, cip->cpc);
break;
case unify_num_op:
case unify_atom_op:
code_p = a_ucons(&do_not_optimise_uatom, unify_atom_op, code_p, pass_no, cip);
break;
case unify_float_op:
code_p = a_ublob(cip->cpc->rnd1, _unify_float, _unify_atom_write, clause_has_blobsp, code_p, pass_no, cip);
break;
case unify_longint_op:
code_p = a_ublob(cip->cpc->rnd1, _unify_longint, _unify_atom_write, clause_has_blobsp, code_p, pass_no, cip);
break;
case unify_bigint_op:
code_p = a_ublob(cip->cpc->rnd1, _unify_bigint, _unify_atom_write, clause_has_blobsp, code_p, pass_no, cip);
break;
case unify_last_num_op:
case unify_last_atom_op:
code_p = a_uc(cip->cpc->rnd1, _unify_l_atom, _unify_l_atom_write, code_p, pass_no);
break;
case unify_last_float_op:
code_p = a_ublob(cip->cpc->rnd1, _unify_l_float, _unify_l_atom_write, clause_has_blobsp, code_p, pass_no, cip);
break;
case unify_last_longint_op:
code_p = a_ublob(cip->cpc->rnd1, _unify_l_longint, _unify_l_atom_write, clause_has_blobsp, code_p, pass_no, cip);
break;
case unify_last_bigint_op:
code_p = a_ublob(cip->cpc->rnd1, _unify_l_bigint, _unify_l_atom_write, clause_has_blobsp, code_p, pass_no, cip);
break;
case write_num_op:
case write_atom_op:
code_p = a_ucons(&do_not_optimise_uatom, write_atom_op, code_p, pass_no, cip);
break;
case write_float_op:
case write_longint_op:
case write_bigint_op:
code_p = a_blob(cip->cpc->rnd1, _write_atom, clause_has_blobsp, code_p, pass_no, cip);
break;
case unify_list_op:
code_p = a_ue(_unify_list, _unify_list_write, code_p, pass_no);
break;
case unify_last_list_op:
code_p = a_ue(_unify_l_list, _unify_l_list_write, code_p, pass_no);
break;
case write_list_op:
code_p = a_e(_write_list, code_p, pass_no);
break;
case write_last_list_op:
code_p = a_e(_write_l_list, code_p, pass_no);
break;
case unify_struct_op:
code_p = a_uf(cip->cpc->rnd1, _unify_struct, _unify_struct_write, code_p, pass_no);
break;
case unify_last_struct_op:
code_p = a_uf(cip->cpc->rnd1, _unify_l_struc, _unify_l_struc_write, code_p, pass_no);
break;
case write_struct_op:
code_p = a_f(cip->cpc->rnd1, _write_struct, code_p, pass_no);
break;
case write_last_struct_op:
code_p = a_f(cip->cpc->rnd1, _write_l_struc, code_p, pass_no);
break;
case save_b_op:
case patch_b_op:
code_p = a_v(_save_b_x, code_p, pass_no, cip->cpc);
break;
case commit_b_op:
code_p = a_v(_commit_b_x, code_p, pass_no, cip->cpc);
#ifdef YAPOR
if (pass_no)
PUT_YAMOP_CUT(*entry_codep);
#endif /* YAPOR */
break;
case save_pair_op:
code_p = a_uv((Ventry *) cip->cpc->rnd1, _save_pair_x, _save_pair_x_write, code_p, pass_no);
break;
case save_appl_op:
code_p = a_uv((Ventry *) cip->cpc->rnd1, _save_appl_x, _save_appl_x_write, code_p, pass_no);
break;
case fail_op:
code_p = a_e(_op_fail, code_p, pass_no);
break;
case cut_op:
code_p = a_cut(&clinfo, code_p, pass_no, cip);
#ifdef YAPOR
if (pass_no)
PUT_YAMOP_CUT(*entry_codep);
#endif /* YAPOR */
break;
case cutexit_op:
code_p = a_cut(&clinfo, code_p, pass_no, cip);
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
*clause_has_blobsp &&
!clinfo.alloc_found)
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
#if THREADS
else
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
!(cip->CurrentPred->PredFlags & ThreadLocalPredFlag))
code_p = a_e(_unlock_lu, code_p, pass_no);
#endif
code_p = a_e(_procceed, code_p, pass_no);
#ifdef YAPOR
if (pass_no)
PUT_YAMOP_CUT(*entry_codep);
#endif /* YAPOR */
break;
case allocate_op:
clinfo.alloc_found = 2;
break;
case deallocate_op:
code_p = a_deallocate(&clinfo, code_p, pass_no, cip);
break;
case tryme_op:
#ifdef TABLING
if (tabled)
code_p = TABLE_TRYCODE(_table_try_me);
else
#endif
code_p = TRYCODE(_try_me, _try_me0);
break;
case retryme_op:
#ifdef TABLING
if (tabled)
code_p = TABLE_TRYCODE(_table_retry_me);
else
#endif
code_p = TRYCODE(_retry_me, _retry_me0);
break;
case trustme_op:
if (log_update &&
(assembling == ASSEMBLING_INDEX ||
assembling == ASSEMBLING_EINDEX)) {
code_p = a_cl(_trust_logical_pred, code_p, pass_no, cip);
}
#ifdef TABLING
if (tabled)
code_p = TABLE_TRYCODE(_table_trust_me);
else
#endif
code_p = TRYCODE(_trust_me, _trust_me0);
break;
case enter_lu_op:
code_p = a_lucl(_enter_lu_pred, code_p, pass_no, cip);
break;
case try_op:
#ifdef TABLING
if (tabled)
code_p = a_gl(_table_try, &clinfo, code_p, pass_no, cip->cpc);
else
#endif
code_p = a_gl(_try_clause, &clinfo, code_p, pass_no, cip->cpc);
break;
case retry_op:
if (log_update) {
add_clref(cip->cpc->rnd1, pass_no);
}
#ifdef TABLING
if (tabled)
code_p = a_gl(_table_retry, &clinfo, code_p, pass_no, cip->cpc);
else
#endif
code_p = a_gl(_retry, &clinfo, code_p, pass_no, cip->cpc);
break;
case trust_op:
if (log_update) {
add_clref(cip->cpc->rnd1, pass_no);
code_p = a_cl(_trust_logical_pred, code_p, pass_no, cip);
}
#ifdef TABLING
if (tabled)
code_p = a_gl(_table_trust, &clinfo, code_p, pass_no, cpc);
else
#endif
code_p = a_gl(_trust, &clinfo, code_p, pass_no, cip->cpc);
break;
case try_in_op:
code_p = a_il(cip->cpc->rnd1, _try_in, code_p, pass_no, cip);
break;
case jump_op:
/* don't assemble jumps to next instruction */
if (cip->cpc->nextInst == NULL ||
cip->cpc->nextInst->op != label_op ||
cip->cpc->rnd1 != cip->cpc->nextInst->rnd1) {
code_p = a_l(cip->cpc->rnd1, _jump, code_p, pass_no, cip);
}
break;
case jumpi_op:
code_p = a_il(cip->cpc->rnd1, _jump, code_p, pass_no, cip);
break;
case restore_tmps_op:
code_p = a_l(cip->cpc->rnd1, _move_back, code_p, pass_no, cip);
break;
case restore_tmps_and_skip_op:
code_p = a_l(cip->cpc->rnd1, _skip, code_p, pass_no, cip);
break;
case procceed_op:
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
*clause_has_blobsp &&
!clinfo.alloc_found)
code_p = a_cle(_alloc_for_logical_pred, code_p, pass_no, cip);
#if THREADS
else
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
!(cip->CurrentPred->PredFlags & ThreadLocalPredFlag))
code_p = a_e(_unlock_lu, code_p, pass_no);
#endif
code_p = a_e(_procceed, code_p, pass_no);
break;
case call_op:
code_p = a_p(_call, &clinfo, code_p, pass_no, cip);
break;
case execute_op:
#if THREADS
if (cip->CurrentPred->PredFlags & LogUpdatePredFlag &&
!(cip->CurrentPred->PredFlags & ThreadLocalPredFlag))
code_p = a_e(_unlock_lu, code_p, pass_no);
#endif
code_p = a_p(_execute, &clinfo, code_p, pass_no, cip);
break;
case safe_call_op:
code_p = a_p(_call, &clinfo, code_p, pass_no, cip);
break;
case label_op:
if (!ystop_found &&
cip->cpc->nextInst != NULL &&
(cip->cpc->nextInst->op == mark_initialised_pvars_op ||
cip->cpc->nextInst->op == blob_op)) {
ystop_found = TRUE;
code_p = a_e(_Ystop, code_p, pass_no);
}
if (!pass_no) {
if (CellPtr(cip->label_offset+cip->cpc->rnd1) > ASP-256) {
Yap_Error_Size = 256+((char *)(cip->label_offset+cip->cpc->rnd1) - (char *)H);
save_machine_regs();
longjmp(cip->CompilerBotch, 3);
}
if ( (char *)(cip->label_offset+cip->cpc->rnd1) >= cip->freep)
cip->freep = (char *)(cip->label_offset+(cip->cpc->rnd1+1));
cip->label_offset[cip->cpc->rnd1] = (CELL) code_p;
}
/* reset dealloc_found in case there was a branch */
clinfo.dealloc_found = FALSE;
break;
case pop_op:
if (cip->cpc->rnd1 == 1)
code_p = a_e(_pop, code_p, pass_no);
else {
code_p = a_n(_pop_n, 2 * CELLSIZE * (cip->cpc->rnd1 - 1), code_p, pass_no);
}
break;
case either_op:
code_p = check_alloc(&clinfo, code_p, pass_no, cip);
#ifdef YAPOR
if (pass_no)
either_inst[either_cont++] = code_p;
code_p = a_either(_either,
-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2,
Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], 0, 0, code_p, pass_no, cip);
#else
code_p = a_either(_either,
-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2,
Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], code_p, pass_no, cip);
#endif /* YAPOR */
break;
case orelse_op:
#ifdef YAPOR
if (pass_no)
either_inst[either_cont++] = code_p;
code_p = a_either(_or_else,
-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2,
Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], 0, 0, code_p, pass_no, cip);
#else
code_p = a_either(_or_else,
-Signed(RealEnvSize) - CELLSIZE * cip->cpc->rnd2,
Unsigned(cip->code_addr) + cip->label_offset[cip->cpc->rnd1], code_p, pass_no, cip);
#endif /* YAPOR */
clinfo.dealloc_found = FALSE;
break;
case orlast_op:
#ifdef YAPOR
if (pass_no)
either_inst[either_cont++] = code_p;
code_p = a_either(_or_last, 0, 0, 0, 0, code_p, pass_no, cpc);
if (pass_no) {
int cont = 1;
do {
either_cont--;
PUT_YAMOP_LTT(either_inst[either_cont], cont++);
} while (either_inst[either_cont]->opc != opcode(_either));
}
#else
code_p = a_e(_or_last, code_p, pass_no);
#endif /* YAPOR */
clinfo.dealloc_found = FALSE;
break;
case cache_arg_op:
code_p = a_4sw_x(_switch_on_arg_type, code_p, pass_no, cip);
break;
case cache_sub_arg_op:
code_p = a_4sw_s(_switch_on_sub_arg_type, code_p, pass_no, cip);
break;
case jump_v_op:
code_p = a_igl(cip->cpc->rnd1, _jump_if_var, code_p, pass_no, cip);
break;
case jump_nv_op:
code_p = a_xigl(_jump_if_nonvar, code_p, pass_no, cip->cpc);
break;
case switch_on_type_op:
code_p = a_4sw(_switch_on_type, code_p, pass_no, cip);
break;
case switch_c_op:
code_p = a_hx(_switch_on_cons, cl_u, log_update, code_p, pass_no, cip);
break;
case switch_f_op:
code_p = a_hx(_switch_on_func, cl_u, log_update, code_p, pass_no, cip);
break;
case if_c_op:
if (cip->cpc->rnd1 == 1) {
code_p = a_if(_go_on_cons, cl_u, log_update, code_p, pass_no, cip);
} else {
code_p = a_if(_if_cons, cl_u, log_update, code_p, pass_no, cip);
}
break;
case if_f_op:
if (cip->cpc->rnd1 == 1) {
code_p = a_if(_go_on_func, cl_u, log_update, code_p, pass_no, cip);
} else {
code_p = a_if(_if_func, cl_u, log_update, code_p, pass_no, cip);
}
break;
case if_not_op:
code_p = a_ifnot(_if_not_then, code_p, pass_no, cip);
break;
case index_dbref_op:
code_p = a_e(_index_dbref, code_p, pass_no);
break;
case index_blob_op:
code_p = a_e(_index_blob, code_p, pass_no);
break;
case mark_initialised_pvars_op:
code_p = a_bmap(code_p, pass_no, cip->cpc);
break;
case mark_live_regs_op:
code_p = a_bregs(code_p, pass_no, cip->cpc);
break;
case commit_opt_op:
clinfo.commit_lab = cip->cpc->rnd1;
break;
case fetch_args_vv_op:
a_fetch_vv(&cmp_info, pass_no, cip);
break;
case fetch_args_vc_op:
a_fetch_vc(&cmp_info, pass_no, cip);
break;
case fetch_args_cv_op:
a_fetch_cv(&cmp_info, pass_no, cip);
break;
case f_val_op:
code_p = a_f2(FALSE, &cmp_info, code_p, pass_no, cip);
break;
case f_var_op:
code_p = a_f2(TRUE, &cmp_info, code_p, pass_no, cip);
break;
case enter_profiling_op:
code_p = a_pl(_enter_profiling, (PredEntry *)(cip->cpc->rnd1), code_p, pass_no);
break;
case retry_profiled_op:
code_p = a_pl(_retry_profiled, (PredEntry *)(cip->cpc->rnd1), code_p, pass_no);
break;
case count_call_op:
code_p = a_pl(_count_call, (PredEntry *)(cip->cpc->rnd1), code_p, pass_no);
break;
case count_retry_op:
code_p = a_pl(_count_retry, (PredEntry *)(cip->cpc->rnd1), code_p, pass_no);
break;
case fetch_args_for_bccall:
if (cip->cpc->nextInst->op != bccall_op) {
Yap_Error(SYSTEM_ERROR, TermNil, "compiling binary test", (int) cip->cpc->op);
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
}
code_p = a_bfunc(cip->cpc->nextInst->rnd2, &clinfo, code_p, pass_no, cip);
break;
case align_float_op:
/* install a blob */
#if SIZEOF_DOUBLE == 2*SIZEOF_LONG_INT
if (!((CELL)code_p & 0x4))
GONEXT(e);
#endif
break;
case blob_op:
/* install a blob */
code_p = copy_blob(code_p, pass_no, cip->cpc);
break;
case empty_call_op:
/* create an empty call */
code_p = a_empty_call(&clinfo, code_p, pass_no, cip);
break;
case push_or_op:
/* be sure to allocate if we have an ;, even if it is
compiled inline.
*/
code_p = check_alloc(&clinfo, code_p, pass_no, cip);
case pushpop_or_op:
case pop_or_op:
case nop_op:
case name_op:
break;
default:
Yap_Error(SYSTEM_ERROR, TermNil, "instruction %d found while assembling", (int) cip->cpc->op);
save_machine_regs();
longjmp(cip->CompilerBotch, 1);
}
cip->cpc = cip->cpc->nextInst;
}
if (!ystop_found)
code_p = a_e(_Ystop, code_p, pass_no);
return code_p;
}
yamop *
Yap_assemble(int mode, Term t, PredEntry *ap, int is_fact, struct intermediates *cip)
{
/*
* the assembly proccess is done in two passes: 1 - a first pass
* computes labels offsets and total code size 2 - the second pass
* produces the final version of the code
*/
UInt size = 0;
yamop *entry_code;
yamop *code_p;
int clause_has_blobs = FALSE;
cip->label_offset = (int *)cip->freep;
cip->code_addr = NULL;
code_p = do_pass(0, &entry_code, mode, &clause_has_blobs, cip, size);
size =
(CELL)NEXTOP(NEXTOP(NEXTOP((yamop *)(((DynamicClause *)NULL)->ClCode),ld),sla),e);
if ((CELL)code_p > size)
size = (CELL)code_p;
if (mode == ASSEMBLING_CLAUSE &&
ap->PredFlags & LogUpdatePredFlag &&
!is_fact) {
DBTerm *x;
LogUpdClause *cl;
CELL *h0 = H;
H = (CELL *)cip->freep;
while ((x = Yap_StoreTermInDBPlusExtraSpace(t, size)) == NULL) {
H = h0;
if (!Yap_growheap(TRUE, size, cip)) {
Yap_Error_TYPE = SYSTEM_ERROR;
return NULL;
}
h0 = H;
H = (CELL *)cip->freep;
}
H = h0;
cl = (LogUpdClause *)((CODEADDR)x-(UInt)size);
cl->ClSource = x;
cip->code_addr = (yamop *)cl;
} else if (mode == ASSEMBLING_CLAUSE &&
(ap->PredFlags & SourcePredFlag ||
(!ap->cs.p_code.NOfClauses && yap_flags[SOURCE_MODE_FLAG])) &&
!is_fact) {
DBTerm *x;
StaticClause *cl;
CELL *h0 = H;
H = (CELL *)cip->freep;
while ((x = Yap_StoreTermInDBPlusExtraSpace(t, size)) == NULL) {
switch (Yap_Error_TYPE) {
case OUT_OF_STACK_ERROR:
Yap_Error_Size = 256+((char *)cip->freep - (char *)H);
save_machine_regs();
longjmp(cip->CompilerBotch,3);
case OUT_OF_TRAIL_ERROR:
/* don't just return NULL */
H = h0;
ARG1 = t;
if (!Yap_growtrail(64 * 1024L)) {
return NULL;
}
Yap_Error_TYPE = YAP_NO_ERROR;
t = ARG1;
h0 = H;
H = (CELL *)cip->freep;
break;
case OUT_OF_HEAP_ERROR:
/* don't just return NULL */
H = h0;
ARG1 = t;
if (!Yap_growheap(TRUE, size, cip)) {
return NULL;
}
Yap_Error_TYPE = YAP_NO_ERROR;
t = ARG1;
h0 = H;
H = (CELL *)cip->freep;
break;
default:
return NULL;
}
}
H = h0;
cl = (StaticClause *)((CODEADDR)x-(UInt)size);
cip->code_addr = (yamop *)cl;
code_p = do_pass(1, &entry_code, mode, &clause_has_blobs, cip, size);
/* make sure we copy after second pass */
cl->usc.ClSource = x;
ProfEnd=code_p;
return entry_code;
} else {
while ((cip->code_addr = (yamop *) Yap_AllocCodeSpace(size)) == NULL) {
if (!Yap_growheap(TRUE, size, cip)) {
Yap_Error_TYPE = SYSTEM_ERROR;
return NULL;
}
}
}
code_p = do_pass(1, &entry_code, mode, &clause_has_blobs, cip, size);
ProfEnd=code_p;
return entry_code;
}
void
Yap_InitComma(void)
{
yamop *code_p = COMMA_CODE;
code_p->opc = opcode(_call);
code_p->u.sla.s = emit_count(-Signed(RealEnvSize) - sizeof(CELL) * 3);
code_p->u.sla.sla_u.p =
code_p->u.sla.p0 =
RepPredProp(PredPropByFunc(FunctorComma,0));
code_p->u.sla.bmap = NULL;
GONEXT(sla);
if (PRED_GOAL_EXPANSION_ON) {
Functor fp = Yap_MkFunctor(Yap_FullLookupAtom("$generate_pred_info"),4);
code_p->opc = emit_op(_call_cpred);
code_p->u.sla.s = emit_count(-Signed(RealEnvSize));
code_p->u.sla.sla_u.p = RepPredProp(Yap_GetPredPropByFunc(fp,0));
code_p->u.sla.bmap = NULL;
GONEXT(sla);
code_p->opc = emit_op(_call);
code_p->u.sla.s = emit_count(-Signed(RealEnvSize));
code_p->u.sla.sla_u.p = PredMetaCall;
code_p->u.sla.bmap = NULL;
GONEXT(sla);
code_p->opc = emit_op(_deallocate);
GONEXT(e);
code_p->opc = emit_op(_procceed);
GONEXT(e);
} else {
if (PROFILING) {
code_p->opc = opcode(_enter_a_profiling);
GONEXT(e);
}
if (CALL_COUNTING) {
code_p->opc = opcode(_count_a_call);
GONEXT(e);
}
code_p->opc = opcode(_p_execute_tail);
GONEXT(e);
}
}