projects
/
sbp.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
checkpoint
[sbp.git]
/
src
/
edu
/
berkeley
/
sbp
/
GSS.java
diff --git
a/src/edu/berkeley/sbp/GSS.java
b/src/edu/berkeley/sbp/GSS.java
index
0bc14ae
..
1d401a5
100644
(file)
--- a/
src/edu/berkeley/sbp/GSS.java
+++ b/
src/edu/berkeley/sbp/GSS.java
@@
-1,7
+1,5
@@
package edu.berkeley.sbp;
import edu.berkeley.sbp.*;
package edu.berkeley.sbp;
import edu.berkeley.sbp.*;
-import edu.berkeley.sbp.*;
-import edu.berkeley.sbp.*;
import edu.berkeley.sbp.util.*;
import java.io.*;
import java.util.*;
import edu.berkeley.sbp.util.*;
import java.io.*;
import java.util.*;
@@
-30,7
+28,7
@@
class GSS {
private Phase.Node[] reducing_list = null;
/** corresponds to a positions <i>between tokens</i> the input stream; same as Tomita's U_i's */
private Phase.Node[] reducing_list = null;
/** corresponds to a positions <i>between tokens</i> the input stream; same as Tomita's U_i's */
- public class Phase {
+ public class Phase implements Invokable<Parser.Table.State, Forest, GSS.Phase.Node> {
/** the token immediately after this phase */
public final Token token;
/** the token immediately after this phase */
public final Token token;
@@
-92,8
+90,7
@@
class GSS {
if (token==null) break;
int count = 0;
Parser.Table.Reduction r = null;
if (token==null) break;
int count = 0;
Parser.Table.Reduction r = null;
- for(Parser.Table.Reduction red : token==null ? state.getEofReductions() : state.getReductions(token)) { r = red; count++; }
- if (count==0) return; // BEWARE! this optimization is suspected to cause really nasty heisenbugs
+ if (!state.canReduce(token)) return;
//if (count > 1) break;
//if (r.numPop == 0) break;
//r.reduce(pending, parent, null, Phase.this, null);
//if (count > 1) break;
//if (r.numPop == 0) break;
//r.reduce(pending, parent, null, Phase.this, null);
@@
-126,8
+123,14
@@
class GSS {
}
}
}
}
+ public void invoke(Parser.Table.State st, Forest result, Node n) {
+ next.newNode(n, result, st, true, this);
+ }
+ private Phase next = null;
+
/** perform all shift operations, adding promoted nodes to <tt>next</tt> */
public void shift(Phase next, Forest result) {
/** perform all shift operations, adding promoted nodes to <tt>next</tt> */
public void shift(Phase next, Forest result) {
+ this.next = next;
closed = true;
Forest res = null;
boolean ok = false;
closed = true;
Forest res = null;
boolean ok = false;
@@
-141,11
+144,14
@@
class GSS {
}
if (!n.holder.valid()) continue;
if (token == null) continue;
}
if (!n.holder.valid()) continue;
if (token == null) continue;
+ n.state.invokeShifts(token, this, result, n);
+ /*
for(Parser.Table.State st : n.state.getShifts(token)) {
if (res == null) res = result;
next.newNode(n, res, st, true, this);
ok = true;
}
for(Parser.Table.State st : n.state.getShifts(token)) {
if (res == null) res = result;
next.newNode(n, res, st, true, this);
ok = true;
}
+ */
}
if (!ok && token != null) {
}
if (!ok && token != null) {
@@
-176,10
+182,11
@@
class GSS {
// GSS Nodes //////////////////////////////////////////////////////////////////////////////
/** a node in the GSS */
// GSS Nodes //////////////////////////////////////////////////////////////////////////////
/** a node in the GSS */
- public final class Node extends FastSet<Node> {
+ public final class Node extends FastSet<Node> implements Invokable<Parser.Table.Reduction, Node, Node> {
public void addParent(Node parent, boolean fromEmptyReduction) {
public void addParent(Node parent, boolean fromEmptyReduction) {
- parents().add(parent, true);
+ if (parents().contains(parent)) return;
+ parents().add(parent);
if (this!=parent && !fromEmptyReduction) queueReductions(parent);
}
if (this!=parent && !fromEmptyReduction) queueReductions(parent);
}
@@
-200,30
+207,50
@@
class GSS {
if (allqueued) return;
allqueued = true;
int where = parents().size();
if (allqueued) return;
allqueued = true;
int where = parents().size();
+ /*
for(Parser.Table.Reduction r : state.getReductions(token))
for(Parser.Table.Reduction r : state.getReductions(token))
- if (r.numPop >= 1)
+ if (r.numPop > 0)
r.reduce(this);
r.reduce(this);
+ */
+ state.invokeReductions(token, this, this, null);
}
public void queueReductions(Node n2) {
if (!allqueued) { queueReductions(); return; }
}
public void queueReductions(Node n2) {
if (!allqueued) { queueReductions(); return; }
+ /*
for(Parser.Table.Reduction r : state.getReductions(token))
if (r.numPop > 0)
r.reduce(this, n2);
for(Parser.Table.Reduction r : state.getReductions(token))
if (r.numPop > 0)
r.reduce(this, n2);
+ */
+ state.invokeReductions(token, this, this, n2);
}
}
-
+ public final void invoke(Parser.Table.Reduction r, Node n, Node n2) {
+ if (n==null) {
+ if (r.numPop==0) r.reduce(this);
+ return;
+ }
+ if (r.numPop==0) return;
+ if (n2==null) {
+ r.reduce(n);
+ } else {
+ r.reduce(n, n2);
+ }
+ }
public void queueEmptyReductions() {
public void queueEmptyReductions() {
- if (reducing)
- for(Parser.Table.Reduction r : token==null ? state.getEofReductions() : state.getReductions(token))
+ if (!reducing) return;
+ /*
+ for(Parser.Table.Reduction r : state.getReductions(token))
if (r.numPop==0)
r.reduce(this);
if (r.numPop==0)
r.reduce(this);
+ */
+ state.invokeReductions(token, this, null, null);
}
private Node(Node parent, Forest pending, Parser.Table.State state, Phase start) {
this.state = state;
if (pending != null) this.holder().merge(pending);
}
private Node(Node parent, Forest pending, Parser.Table.State state, Phase start) {
this.state = state;
if (pending != null) this.holder().merge(pending);
- if (parent != null) parents().add(parent, true);
+ if (parent != null) parents().add(parent);
if (Phase.this.hash.get(code(state, start)) != null) throw new Error("severe problem!");
Phase.this.hash.put(code(state, start), this);
Phase.this.numNodes++;
if (Phase.this.hash.get(code(state, start)) != null) throw new Error("severe problem!");
Phase.this.hash.put(code(state, start), this);
Phase.this.numNodes++;