From: Jean Privat Date: Thu, 11 Dec 2014 20:47:34 +0000 (-0500) Subject: nitcc: remove warnings and improve doc X-Git-Tag: v0.7~56^2 X-Git-Url: http://nitlanguage.org?hp=74ef06299fdf541ea2c204c510295e5e7a14615a nitcc: remove warnings and improve doc Signed-off-by: Jean Privat --- diff --git a/contrib/nitcc/examples/minilang.nit b/contrib/nitcc/examples/minilang.nit index 4160f01..e316369 100644 --- a/contrib/nitcc/examples/minilang.nit +++ b/contrib/nitcc/examples/minilang.nit @@ -1,15 +1,23 @@ import minilang_test_parser +# An naive recursive stack-based interpreter of the minilang language. class Interpretor super Visitor + + # A stack of numeric values var stack = new Array[Int] + + # A stack of boolean values var bstack = new Array[Bool] + + # The current values assigned to each variable var vars = new HashMap[String, Int] redef fun visit(n) do n.accept_calculator(self) end redef class Node + # Execution of the node by the interpreter `v` fun accept_calculator(v: Interpretor) do visit_children(v) end diff --git a/contrib/nitcc/src/autom.nit b/contrib/nitcc/src/autom.nit index 8825b01..5ece999 100644 --- a/contrib/nitcc/src/autom.nit +++ b/contrib/nitcc/src/autom.nit @@ -23,18 +23,18 @@ class Automaton # The start state var start: State - # State that are accect states + # State that are accept states var accept = new Array[State] # All states var states = new Array[State] - # Tokens associated on accept states - # use `add_tag` to update + # Tokens associated on accept states. + # Use `add_tag` to update var tags = new HashMap[State, Set[Token]] - # Accept states associated on tokens - # use `add_tag` to update + # Accept states associated on tokens. + # Use `add_tag` to update var retrotags = new HashMap[Token, Set[State]] # Tag all accept states @@ -66,7 +66,7 @@ class Automaton assert retrotags[t].has(s) end - # Remove all occurences of a tag in an automaton + # Remove all occurrences of a tag in an automaton fun clear_tag(t: Token) do if not retrotags.has_key(t) then return @@ -78,7 +78,7 @@ class Automaton retrotags.keys.remove(t) end - # Remove tokens from conflicting state according the the inclusion of language + # Remove tokens from conflicting state according the inclusion of language. # REQUIRE: self isa DFA automaton fun solve_token_inclusion do @@ -101,8 +101,8 @@ class Automaton end end - # Initialize a new automaton for the empty language - # one state, no accept, no transition + # Initialize a new automaton for the empty language. + # One state, no accept, no transition. init empty do var state = new State @@ -110,8 +110,8 @@ class Automaton states.add state end - # Initialize a new automaton for the empty-string language - # one state, is accept, no transition + # Initialize a new automaton for the empty-string language. + # One state, is accept, no transition. init epsilon do var state = new State @@ -120,8 +120,8 @@ class Automaton states.add state end - # Initialize a new automation for the language that accepts only a single symbol - # Two state, the second is accept, one transition on `symbol` + # Initialize a new automation for the language that accepts only a single symbol. + # Two state, the second is accept, one transition on `symbol`. init atom(symbol: Int) do var s = new State @@ -148,8 +148,8 @@ class Automaton states.add a end - # Contatenate `other` to `self` - # other is modified and invalidated. + # Concatenate `other` to `self`. + # Other is modified and invalidated. fun concat(other: Automaton) do var s2 = other.start @@ -160,7 +160,7 @@ class Automaton states.add_all other.states end - # `self` become the alternation of `self` and `other` + # `self` become the alternation of `self` and `other`. # `other` is modified and invalidated. fun alternate(other: Automaton) do @@ -184,9 +184,9 @@ class Automaton states.add_all other.states end - # Return a new automaton that recognize `self` but not `other` - # For a theorical POV, this is the substraction of languages. - # Note: the implementation use `to_dfa` internally, so the theorical complexity is not cheap. + # Return a new automaton that recognize `self` but not `other`. + # For a theoretical POV, this is the subtraction of languages. + # Note: the implementation use `to_dfa` internally, so the theoretical complexity is not cheap. fun except(other: Automaton): Automaton do var ta = new Token("1") @@ -209,8 +209,8 @@ class Automaton return c end - # `self` absorbs all states, transisions, tags, and acceptations of `other` - # An epsilon transition is added between `self.start` and `other.start` + # `self` absorbs all states, transitions, tags, and acceptations of `other`. + # An epsilon transition is added between `self.start` and `other.start`. fun absorb(other: Automaton) do states.add_all other.states @@ -494,8 +494,8 @@ class Automaton f.close end - # Transform a NFA to a DFA - # note: the DFA is not miminized + # Transform a NFA to a DFA. + # note: the DFA is not minimized. fun to_dfa: Automaton do trim @@ -581,8 +581,8 @@ class Automaton return dfa end - # epsilon-closure on a state of states - # used by `to_dfa` + # Epsilon-closure on a state of states. + # Used by `to_dfa`. private fun eclosure(states: Collection[State]): Set[State] do var res = new ArraySet[State] @@ -601,8 +601,8 @@ class Automaton return res end - # trans on a set of states - # Used by `to_dfa` + # Trans on a set of states. + # Used by `to_dfa`. fun trans(states: Collection[State], symbol: Int): Set[State] do var res = new ArraySet[State] @@ -621,9 +621,9 @@ class Automaton return res end - # Generate the Nit source code of the lexer - # `filepath` is the name of the ouptit file - # `parser` is the name of the parser module (used to import the token classes) + # Generate the Nit source code of the lexer. + # `filepath` is the name of the output file. + # `parser` is the name of the parser module (used to import the token classes). fun gen_to_nit(filepath: String, name: String, parser: nullable String) do var gen = new DFAGenerator(filepath, name, self, parser) @@ -638,12 +638,9 @@ private class DFAGenerator var automaton: Automaton var parser: nullable String - var out: OStream - init(filepath: String, name: String, automaton: Automaton, parser: nullable String) do - self.filepath = filepath - self.name = name - self.automaton = automaton - self.parser = parser + var out: OStream is noinit + + init do self.out = new OFStream.open(filepath) end @@ -747,9 +744,9 @@ end # A state in a finite automaton class State # Outgoing transitions - var outs = new Array[Transition] - # Ingoing tyransitions + + # Ingoing transitions var ins = new Array[Transition] # Add a transitions to `to` on `symbol` (null means epsilon) @@ -761,6 +758,8 @@ class State return t end + # Get the first state following the transition `i`. + # Null if no transition for `i`. fun trans(i: Int): nullable State do for t in outs do @@ -778,7 +777,12 @@ end # A range of symbols on a transition class TSymbol + # The first symbol in the range var first: Int + + # The last symbol if any. + # + # `null` means infinity. var last: nullable Int redef fun to_s @@ -808,8 +812,8 @@ class Transition # The symbol on the transition (null means epsilon) var symbol: nullable TSymbol - # Remove the transition from the automaton - # Detash from `from` and `to` + # Remove the transition from the automaton. + # Detach from `from` and `to`. fun delete do from.outs.remove(self) diff --git a/contrib/nitcc/src/grammar.nit b/contrib/nitcc/src/grammar.nit index 620b0a0..eea312f 100644 --- a/contrib/nitcc/src/grammar.nit +++ b/contrib/nitcc/src/grammar.nit @@ -12,12 +12,12 @@ # See the License for the specific language governing permissions and # limitations under the License. -# A gramar describing a language +# A grammar describing a language class Gram - # The productions (non-terminal) of the conctete grammar + # The productions (non-terminal) of the concrete grammar var prods = new Array[Production] - # The additionnal abstract productions of the grammar + # The additional abstract productions of the grammar # TODO clean AST var ast_prods = new Array[Production] @@ -35,7 +35,7 @@ class Gram res.append("{p.name} =\n") end var last = null - if not p.alts.is_empty then p.alts.last + if not p.alts.is_empty then last = p.alts.last for a in p.alts do res.append("\t\{{a.name}:\} {a.elems.join(" ")}") if a.codes == null then a.make_codes @@ -55,7 +55,7 @@ class Gram return res.to_s end - # Inline (ie. remove from the conctete grammar) some production + # Inline (ie. remove from the concrete grammar) some production # REQUIRE: no circular production in `prods` fun inline(prods: Collection[Production]) do @@ -322,14 +322,14 @@ class Production # The alternative of the production var alts = new Array[Alternative] - # Additionnal alternatives in the AST + # Additional alternatives in the AST var ast_alts = new Array[Alternative] # Is self the accept production var accept = false # Is self transformed to a other production for the AST - # FIXME: cleaup AST + # FIXME: cleanup AST var spe: nullable Production = null is writable # Is self contains only a single alternative (then no need for a abstract production class in the AST) @@ -345,8 +345,10 @@ class Production # Is the production nullable var is_nullable = false + # The first tokens of the production var firsts = new HashSet[Item] + # The tokens that may follows the production (as in SLR) var afters = new HashSet[Item] @@ -432,7 +434,7 @@ class Alternative # Is the alternative unparsable? (ie not in the automaton) var phony = false is writable - # Imitialize codes with the elements + # Initialize codes with the elements fun make_codes do if codes != null then return @@ -445,20 +447,27 @@ class Alternative end end -# A step in the construction of the AST. used to modelize transformations +# A step in the construction of the AST. +# Used to model transformations interface Code end + # Get a element from the stack class CodePop super Code redef fun to_s do return "pop" end -# Allocate a new AST node for an alternative using the correct number of poped elements + +# Allocate a new AST node for an alternative using the correct number of popped elements class CodeNew super Code + + # The associated alternative var alt: Alternative + redef fun to_s do return "New {alt.name}/{alt.elems.length}" end + # Get null class CodeNull super Code @@ -477,7 +486,7 @@ abstract class Element # The mangled name of the element fun cname: String do return "N{name.to_cmangle}" - # the name of the class in the AST + # The name of the class in the AST fun acname: String do var res = acname_cache if res == null then @@ -486,15 +495,17 @@ abstract class Element end return res end + + # The name of the class in the AST fun acname=(s: String) do acname_cache = s end # A terminal element class Token super Element - # States of the LR automatio that shift on self + # States of the LR automaton that shift on self var shifts = new ArraySet[LRState] - # States of the LR automatio that reduce on self in the lookahead(1) + # States of the LR automaton that reduce on self in the lookahead(1) var reduces = new ArraySet[LRState] end @@ -872,13 +883,13 @@ end # A state in a LR automaton class LRState - # name of the automaton (short part from the start) + # Name of the automaton (short part from the start) var name: String - # malglen name + # Mangled name fun cname: String do return name.to_cmangle - # number + # Number var number: Int = -1 # Set of all items @@ -893,7 +904,7 @@ class LRState # Ingoing transitions var outs = new Array[LRTransition] - # trans function + # Trans function fun trans(e: Element): nullable LRState do for t in outs do if t.elem == e then return t.to @@ -915,7 +926,7 @@ class LRState return true end - # Recusively extends item outside the core + # Recursively extends item outside the core fun extends(i: Item) do var e = i.next @@ -940,7 +951,7 @@ class LRState var gotos = new ArraySet[Production] # Reduction guarded by tokens var guarded_reduce = new HashMap[Token, Set[Item]] - # Shitfs guarded by tokens + # Shifts guarded by tokens var guarded_shift = new HashMap[Token, Set[Item]] # Does the state need a guard to perform an action? @@ -949,7 +960,7 @@ class LRState # Is the state LR0? fun is_lr0: Bool do return reduces.length <= 1 and shifts.is_empty or reduces.is_empty - # compute guards and conflicts + # Compute guards and conflicts fun analysis do # Extends the core @@ -1031,7 +1042,7 @@ class LRState end end - # Return `i` and all other items of the state that expands, directly or undirectly, to `i` + # Return `i` and all other items of the state that expands, directly or indirectly, to `i` fun back_expand(i: Item): Set[Item] do var res = new ArraySet[Item] @@ -1056,7 +1067,7 @@ end class LRTransition # The origin state var from: LRState - # The testination state + # The destination state var to: LRState # The element labeling the transition var elem: Element @@ -1085,14 +1096,14 @@ class Item return b.to_s end - # The element thatr follow the dot, null if the fdot is at the end + # The element that follows the dot, null if the dot is at the end fun next: nullable Element do if pos >= alt.elems.length then return null return alt.elems[pos] end - # SLR loohahead + # SLR lookahead fun lookahead: Set[Token] do var res = new HashSet[Token] diff --git a/contrib/nitcc/src/nitcc_lexer0.nit b/contrib/nitcc/src/nitcc_lexer0.nit index 7fa876f..f3ff58d 100644 --- a/contrib/nitcc/src/nitcc_lexer0.nit +++ b/contrib/nitcc/src/nitcc_lexer0.nit @@ -12,24 +12,29 @@ # See the License for the specific language governing permissions and # limitations under the License. -# Ad-hoc hand-writen lexer for nitcc -# This avoid to commit (and relyon ) a generated lexer -# +# Ad-hoc hand-written lexer for nitcc +# This avoid to commit (and rely on) a generated lexer module nitcc_lexer0 # Required for the tokens definitions import nitcc_parser -# Hand-writen lexer of nitcc -# Used only for the boostrap of the tool. +# Hand-written lexer of nitcc. +# Used only for the bootstrap of the tool. class Lexer_nitcc + # The text to tokenize var text: String - var iter: Iterator[Char] = "".chars.iterator + # The iterator on text + private var iter: Iterator[Char] is noinit + + # The current position var pos = 0 - var tokens = new Array[NToken] + # The tokens currently produced + private var tokens = new Array[NToken] + # Tokenize and returns the tokens fun lex: Array[NToken] do iter = text.chars.iterator @@ -88,13 +93,13 @@ class Lexer_nitcc return tokens end - fun error(c: Char) + private fun error(c: Char) do print "pos {pos}: Lexer error on '{c}'." abort end - fun str + private fun str do var b = new FlatBuffer b.add('\'') @@ -121,7 +126,7 @@ class Lexer_nitcc abort end - fun id(c: Char) + private fun id(c: Char) do var b = new FlatBuffer b.add c @@ -138,7 +143,7 @@ class Lexer_nitcc tokens.add token end - fun kw(c: Char) + private fun kw(c: Char) do var b = new FlatBuffer b.add c @@ -155,7 +160,7 @@ class Lexer_nitcc tokens.add token end - fun trim + private fun trim do while iter.is_ok and iter.item <= ' ' do iter.next diff --git a/contrib/nitcc/src/nitcc_semantic.nit b/contrib/nitcc/src/nitcc_semantic.nit index 73baa05..af417d4 100644 --- a/contrib/nitcc/src/nitcc_semantic.nit +++ b/contrib/nitcc/src/nitcc_semantic.nit @@ -28,6 +28,7 @@ import re2nfa class CollectNameVisitor super Visitor + # All the productions var nprods = new Array[Nprod] # Symbol table to associate things (prods and exprs) with their name @@ -44,7 +45,7 @@ class CollectNameVisitor # The current production, used to initialize priorities var prod: nullable Production = null - # The current priority counter to name tranformed productions + # The current priority counter to name transformed productions var pricpt: Int = 0 # Run the semantic analysis of the grammar @@ -53,13 +54,13 @@ class CollectNameVisitor # First visit to collect names enter_visit(n) - # Second visit to use collectec names and build rhings + # Second visit to use collected names and build things var v2 = new CheckNameVisitor(self) v2.enter_visit(n) - # Inline all the ? + # Inline all the `?` gram.inline(v2.quesizes.values) - # Inlile all the prods sufixed by '_imline' #TODO use a real keyword + # Inline all the prods suffixed by '_inline' #TODO use a real keyword for p in gram.prods do if not p.name.has_suffix("_inline") then continue print "inline {p}" @@ -102,14 +103,14 @@ private class CheckNameVisitor # The collected element names, for the alternative var elems_names = new Array[nullable String] - # The collected elementname, for the nelem + # The collected element names, for the nelem var elemname: nullable String = null # Is the alternative transformed, for the alternative var trans = false # The current priority class - # Used the check, and tranform the grammar + # Used the check, and transform the grammar var pri: nullable Npriority = null # Known ignored tokens @@ -119,7 +120,7 @@ private class CheckNameVisitor var rejecteds = new Array[Element] # Pool of elements that are modified with + (reuse them!) - private var plusizes = new HashMap[Element, Production] + var plusizes = new HashMap[Element, Production] # Create a + version of an element fun plusize(e: Element): Production @@ -136,7 +137,7 @@ private class CheckNameVisitor end # Pool of elements that are modified with ? (reuse them!) - private var quesizes = new HashMap[Element, Production] + var quesizes = new HashMap[Element, Production] # Create a ? version of an element fun quesize(e: Element): Production @@ -155,14 +156,13 @@ private class CheckNameVisitor end # The current nexpr, used to track dependency on named expressions (see `Nexpr::precs`) - var nexpr: nullable Nexpr + var nexpr: nullable Nexpr = null # The current production, used to initialize alternatives - var prod: nullable Production + var prod: nullable Production = null # The main visitor, used to access the grammar of the symbol table var v1: CollectNameVisitor - init(v1: CollectNameVisitor) do self.v1 = v1 redef fun visit(n) do n.accept_check_name_visitor(self) end @@ -204,7 +204,7 @@ redef class Nexpr # The associated NFA (cached, see `build_nfa`) private var nfa: nullable Automaton - # Build the NFA, possibily building the NFA of required expressions + # Build the NFA, possibly building the NFA of required expressions # Print an error if there is a circular dependency # The result is cached fun build_nfa: Automaton do @@ -269,7 +269,7 @@ redef class Nign abort else if e isa Token then # The token was build and registered during the visit - # So, unregister then, the bit Ignred token will be build later + # So, unregister them, the big Ignored token will be build later v.v1.gram.tokens.remove(e) else abort @@ -298,13 +298,13 @@ redef class Nrej end redef class Nprod - # The associated main production - # ie the last priority class + # The associated main production. + # i.e. the last priority class. var prod: nullable Production - # The associated most-priority production - # ie the first priority class - # If there is no priority then `sub_prod == prod` + # The associated most-priority production. + # i.e. the first priority class. + # If there is no priority then `sub_prod == prod`. var sub_prod: nullable Production redef fun accept_collect_prod(v) do @@ -376,13 +376,14 @@ redef class Natrans end redef class Npriority + # It is the last priority group? var is_last = false # The associated production var prod: nullable Production - # The production in the with the next less priority class - # null is there is no priority or if the first priority class + # The production in the with the next less priority class. + # `null` if there is no priority or if the first priority class. var next: nullable Production redef fun accept_collect_prod(v) do @@ -409,7 +410,7 @@ redef class Npriority v.pri = self super - # Inject a new alternative that goes to the next less prioty class + # Inject a new alternative that goes to the next less priority class var alt = prod.new_alt2(prod.name + "_" + prod.alts.length.to_s, [next.as(not null)]) alt.trans = true alt.codes = [new CodePop] @@ -462,6 +463,8 @@ redef class Npriority_unary end redef class Alternative + # The short name of the alternative. + # Used for errors var short_name: nullable String end diff --git a/contrib/nitcc/src/re2nfa.nit b/contrib/nitcc/src/re2nfa.nit index 6bca660..d0d90ac 100644 --- a/contrib/nitcc/src/re2nfa.nit +++ b/contrib/nitcc/src/re2nfa.nit @@ -37,7 +37,6 @@ redef class Nstr redef fun make_rfa: Automaton do var a = new Automaton.epsilon - var val for c in self.value.chars do var b = new Automaton.atom(c.ascii) a.concat(b)