1 package edu.berkeley.sbp;
2 import edu.berkeley.sbp.*;
3 import edu.berkeley.sbp.*;
4 import edu.berkeley.sbp.*;
5 import edu.berkeley.sbp.util.*;
8 import java.lang.reflect.*;
10 //////////////////////////////////////////////////////////////////////////////
13 // - fix public/package/private status
16 //////////////////////////////////////////////////////////////////////////////
17 // Optimizations to add
19 // ** NOTE: not all of these are appropriate for this class -- it is
20 // simply a list of optimizations not implemented. This
21 // class is meant to remain simple and easy to understand;
22 // optimizations which obscure that do not belong here (they
23 // should go into the compiled version instead)
25 // - most of our time is now spent creating and storing Reduct instances
26 // - we should be able to perform Reduct's immediately after creating them...
29 /** implements Tomita's Graph Structured Stack */
34 /** corresponds to a positions <i>between tokens</i> the input stream; same as Tomita's U_i's */
37 /** the token immediately after this phase */
38 public final Token token;
40 /** currently this is necessary only for the code() hack -- it doesn't actually correspond to the input */
41 private final int pos;
44 public Forest.Ref finalResult = null;
46 /** all reductions (pending and completed) */
47 private HashSet<Phase.Reduct> reductions = new HashSet<Phase.Reduct>(); /* ALLOC */
49 /** all nodes, keyed by the value returned by code() */
50 private HashMap<Long,Phase.Node> hash = new HashMap<Long,Phase.Node>(); /* ALLOC */
52 /** the number of pending reductions */
53 private int pendingReductions = 0;
54 private int totalReductions = 0;
55 //private HashSet<Reduct> pendingReduct = new HashSet<Reduct>();
56 private LinkedList<Reduct> pendingReduct = new LinkedList<Reduct>();
58 /** the number of nodes in this phase */
59 private int numNodes = 0;
61 boolean closed = false;
63 private Token.Location location;
64 public Phase(Phase previous, Token token, Token.Location location) {
65 this.pos = previous==null ? 0 : previous.pos+1;
67 this.location = location;
70 public boolean isDone() { return token == null; }
72 private String error = "generic syntax error";
73 public void checkFailure() throws Parser.Failed {
75 throw new Parser.Failed(error, getLocation());
78 public Token.Location getLocation() { return location; }
80 /** add a new node (merging with existing nodes if possible)
81 * @param parent the parent of the new node
82 * @param result the SPPF result corresponding to the new node
83 * @param state the state that the new node is in
84 * @param fromEmptyReduction true iff this node is being created as a result of a reduction of length zero (see GRMLR paper)
85 * @param start the earliest part of the input contributing to this node (used to make merging decisions)
87 public void newNode(Node parent, Forest pending, Parser.Table.State state, boolean fromEmptyReduction, Phase start) {
88 Node p = hash.get(code(state, start));
89 if (p != null) newNode2(p, parent, pending, state, fromEmptyReduction, start);
90 else newNode3(parent, pending, state, fromEmptyReduction, start);
92 private void newNode2(Node p, Node parent, Forest pending, Parser.Table.State state, boolean fromEmptyReduction, Phase start) {
93 p.holder.merge(pending);
94 if (p.parents.contains(parent)) return;
95 p.parents.add(parent, true);
96 if (p!=parent && !fromEmptyReduction) p.queueReductions(parent);
98 private void newNode3(Node parent, Forest pending, Parser.Table.State state, boolean fromEmptyReduction, Phase start) {
100 if (token != null && state.canShift(token)) break;
101 if (state.isAccepting()) break;
102 if (token==null) break;
104 Parser.Table.Reduction r = null;
105 for(Parser.Table.Reduction red : token==null ? state.getEofReductions() : state.getReductions(token)) { r = red; count++; }
106 //if (count==0) return; // BEWARE! this optimization is suspected to cause really nasty heisenbugs
107 //if (count > 1) break;
108 //if (r.numPop == 0) break;
109 //r.reduce(pending, parent, null, Phase.this, null);
113 Node n = new Node(parent, pending, state, start); // ALLOC
114 n.queueEmptyReductions();
115 if (!fromEmptyReduction) n.queueReductions();
119 boolean reducing = false;
120 /** perform all reduction operations */
121 public void reduce() {
123 HashSet<Phase.Node> s = new HashSet<Phase.Node>();
124 s.addAll(hash.values());
125 for(Phase.Node n : s) n.queueEmptyReductions();
126 for(Phase.Node n : s) n.queueReductions();
127 while(pendingReduct.size()>0)
128 //pendingReduct.iterator().next().go();
129 pendingReduct.removeFirst().go();
132 /** perform all shift operations, adding promoted nodes to <tt>next</tt> */
133 public void shift(Phase next, Forest result) {
137 for(Phase.Node n : hash.values()) {
138 if (n.holder==null) continue;
140 if (token == null && n.state.isAccepting()) {
142 if (finalResult==null) finalResult = new Forest.Ref();
143 finalResult.merge(n.holder);
145 if (!n.holder.valid()) continue;
146 if (token == null) continue;
147 for(Parser.Table.State st : n.state.getShifts(token)) {
148 if (res == null) res = result;
149 next.newNode(n, res, st, true, this);
154 if (!ok && token != null) {
155 StringBuffer error = new StringBuffer();
156 error.append("error: unable to shift token \"" + token + "\"\n");
157 error.append(" before: " +pendingReductions+ "\n");
158 error.append(" before: " +totalReductions+ "\n");
159 //for(Phase.Node n : hash.values()) {
160 //n.queueReductions();
161 //n.queueEmptyReductions();
163 error.append(" after: " +pendingReductions+ "\n");
164 error.append(" candidate states:\n");
165 for(Phase.Node n : hash.values()) {
166 //for(Sequence.Position p : n.state) error.append(" " + p + "\n");
167 //error.append(" --\n");
168 for(Parser.Table.Reduction r : n.state.getReductions(token)) error.append(" " + r + "\n");
169 //error.append(" ==\n");
171 next.error = error.toString();
174 // this massively improves GC performance
180 // GSS Nodes //////////////////////////////////////////////////////////////////////////////
182 //private HashMap<Parser.Table.Reduction,Forest> pcache = new HashMap<Parser.Table.Reduction,Forest>();
183 /** a node in the GSS */
184 public final class Node {
186 private Forest.Ref holder = null;
188 private HashMap<Parser.Table.Reduction,Forest> cache = null;
190 /** the set of nodes to which there is an edge starting at this node */
191 public final FastSet<Node> parents = new FastSet<Node>(); /* ALLOC */
193 /** what state this node is in */
194 public final Parser.Table.State state;
195 /** which Phase this Node belongs to (node that Node is also a non-static inner class of Phase) */
196 public final Phase phase = Phase.this;
198 public HashMap<Parser.Table.Reduction,Forest> cache() {
199 return cache==null ? (cache = new HashMap<Parser.Table.Reduction,Forest>()) : cache; }
200 public Forest.Ref holder() { return holder==null ? (holder = new Forest.Ref()) : holder; }
201 public Forest pending() { return Phase.this.closed ? holder().resolve() : holder; }
202 public FastSet<Node> parents() { return parents; }
205 public void queueReductions() {
206 for(Node n2 : parents)
211 public void queueReductions(Node n2) {
212 newReduct(this, n2, null);
217 public void queueEmptyReductions() {
219 for(Parser.Table.Reduction r : token==null ? state.getEofReductions() : state.getReductions(token)) {
221 //r.reduce(this, null, this.phase, r.zero());
222 Reduct red = new Reduct(this, null, r);
223 red.go(); /* ALLOC */
229 private Node(Node parent, Forest pending, Parser.Table.State state, Phase start) {
231 if (pending != null) this.holder().merge(pending);
232 if (parent != null) parents.add(parent, true);
233 if (Phase.this.hash.get(code(state, start)) != null) throw new Error("severe problem!");
234 Phase.this.hash.put(code(state, start), this);
235 Phase.this.numNodes++;
236 if (parent==null) holder().valid = true; // hack to make sure that the "base" node is always considered valid
240 public void newReduct(Node n, Node n2, Parser.Table.Reduction r) {
241 new Reduct(n, n2, r)/*.go()*/;
244 // Forest / Completed Reductions //////////////////////////////////////////////////////////////////////////////
246 /** a pending or completed reduction */
249 /** the node from which the reduction should begin */
250 public Node n = null;
252 /** the node on the other end of the edge to be reduced along (either: null, the second node of the reduction,
253 * or the parent of the result of a length-one reduction)
255 public Node n2 = null;
257 /** true iff the reduction has already been performed */
258 private boolean done = false;
260 /** the reduction to be applied */
261 public Parser.Table.Reduction r;
263 public Tree<String> result = null;
265 public Reduct(Node n, Node n2, Parser.Table.Reduction r) {
269 if (reductions.contains(this)) { done = true; return; }
270 reductions.add(this);
271 pendingReduct.addFirst(this);
275 /** perform the reduction */
279 pendingReduct.remove(this);
283 for(Parser.Table.Reduction r : token==null ? n.state.getEofReductions() : n.state.getReductions(token)) {
285 // currently we have this weird problem where we
286 // have to do an individual reduct for each child
287 // when the reduction length is one (ie the
288 // children wind up being children of the newly
289 // created node rather than part of the popped
292 if (r.numPop == 1) new Reduct(n, n2, r)/*.go()*/;
296 // FIXME: explain this
298 for(Parser.Table.Reduction r : token==null ? n.state.getEofReductions() : n.state.getReductions(token)) {
299 if (r.numPop <= 1) continue;
300 r.reduce(n, n2, Phase.this, null);
302 } else if (r.numPop==0) { r.reduce(n, n2, n.phase, r.zero());
303 } else if (r.numPop==1) {
305 // The problem here is that a "reduction of length 1"
306 // performed twice with different values of n2 needs
307 // to only create a *single* new result, but must add
308 // multiple parents to the node holding that result.
309 // The current reducer doesn't differentiate between
310 // the next node of an n-pop reduction and the
311 // ultimate parent of the last pop, so we need to
312 // cache instances here as a way of avoiding
315 Forest ret = n.cache().get(r);
316 if (ret != null) r.reduce(n, n2, n.phase, ret);
317 else n.cache().put(r, r.reduce(n, n2, n.phase, null));
320 r.reduce(n, n2, Phase.this, null);
324 // FIXME: this is a PITA
325 public int hashCode() { return n.hashCode() ^ (r==null ? 0 : r.hashCode()) ^ (n2==null ? 0 : n2.hashCode()); }
326 public boolean equals(Object o) {
327 if (o==null) return false;
328 if (o==this) return true;
329 if (!(o instanceof Reduct)) return false;
330 Reduct other = (Reduct)o;
331 return equal(r, other.r) && equal(n, other.n) && equal(n2, other.n2);
338 private static boolean equal(Object a, Object b) {
339 if (a==null && b==null) return true;
340 if (a==null || b==null) return false;
344 /** this is something of a hack right now */
345 private static long code(Parser.Table.State state, Phase start) {
346 return (((long)state.idx) << 32) | (start==null ? 0 : start.pos);
348 public boolean yak = false;