reducing = true;
HashSet<Phase.Node> s = new HashSet<Phase.Node>();
s.addAll(hash.values());
+ //while(pendingReduct.size()>0)
+ //pendingReduct.removeFirst().go();
for(Phase.Node n : s) n.queueEmptyReductions();
for(Phase.Node n : s) n.queueReductions();
- while(pendingReduct.size()>0)
- //pendingReduct.iterator().next().go();
- pendingReduct.removeFirst().go();
}
/** perform all shift operations, adding promoted nodes to <tt>next</tt> */
queueReductions(n2);
}
+ private HashSet<Node> queued = new HashSet<Node>();
/** FIXME */
public void queueReductions(Node n2) {
- newReduct(this, n2, null);
+ if (queued.contains(n2)) return;
+ queued.add(n2);
+ Node n = this;
+ for(Parser.Table.Reduction r : token==null ? n.state.getEofReductions() : n.state.getReductions(token)) {
+
+ // UGLY HACK
+ // The problem here is that a "reduction of length 1"
+ // performed twice with different values of n2 needs
+ // to only create a *single* new result, but must add
+ // multiple parents to the node holding that result.
+ // The current reducer doesn't differentiate between
+ // the next node of an n-pop reduction and the
+ // ultimate parent of the last pop, so we need to
+ // cache instances here as a way of avoiding
+ // recreating them.
+
+ // currently we have this weird problem where we
+ // have to do an individual reduct for each child
+ // when the reduction length is one (ie the
+ // children wind up being children of the newly
+ // created node rather than part of the popped
+ // sequence
+ if (r.numPop <= 0) continue;
+ if (r.numPop == 1) {
+ Forest ret = n.cache().get(r);
+ if (ret != null) r.reduce(n, n2, n.phase, ret);
+ else n.cache().put(r, r.reduce(n, n2, n.phase, null));
+ } else {
+ r.reduce(n, n2, Phase.this, null);
+ }
+ }
}
}
}
- public void newReduct(Node n, Node n2, Parser.Table.Reduction r) {
- new Reduct(n, n2, r)/*.go()*/;
- }
-
// Forest / Completed Reductions //////////////////////////////////////////////////////////////////////////////
/** a pending or completed reduction */
this.n = n;
this.n2 = n2;
this.r = r;
- if (reductions.contains(this)) { done = true; return; }
- reductions.add(this);
- pendingReduct.addFirst(this);
- pendingReductions++;
- //if (reducing) go();
+ //if (reductions.contains(this)) { done = true; return; }
+ //reductions.add(this);
+ //pendingReduct.addFirst(this);
+ //pendingReductions++;
+ go();
}
/** perform the reduction */
public void go() {
if (done) return;
done = true;
- pendingReduct.remove(this);
- pendingReductions--;
+ //pendingReduct.remove(this);
+ //pendingReductions--;
if (r==null) {
for(Parser.Table.Reduction r : token==null ? n.state.getEofReductions() : n.state.getReductions(token)) {
// children wind up being children of the newly
// created node rather than part of the popped
// sequence
-
+ if (r.numPop <= 0) continue;
if (r.numPop == 1) {
Forest ret = n.cache().get(r);
if (ret != null) r.reduce(n, n2, n.phase, ret);
else n.cache().put(r, r.reduce(n, n2, n.phase, null));
+ } else {
+ r.reduce(n, n2, Phase.this, null);
}
}
- for(Parser.Table.Reduction r : token==null ? n.state.getEofReductions() : n.state.getReductions(token)) {
- if (r.numPop <= 1) continue;
- r.reduce(n, n2, Phase.this, null);
- }
} else if (r.numPop != 1) {
r.reduce(n, n2, Phase.this, null);
}