| @@ -78,6 +78,6 @@ def test_earley_equals_lalr(): | |||
| if __name__ == '__main__': | |||
| test_python_lib() | |||
| test_earley_equals_lalr() | |||
| # test_earley_equals_lalr() | |||
| # python_parser3.parse(_read(sys.argv[1]) + '\n') | |||
| @@ -1,3 +1,25 @@ | |||
| class Symbol(object): | |||
| is_term = NotImplemented | |||
| def __init__(self, name): | |||
| self.name = name | |||
| def __eq__(self, other): | |||
| assert isinstance(other, Symbol), other | |||
| return self.is_term == other.is_term and self.name == other.name | |||
| def __hash__(self): | |||
| return hash(self.name) | |||
| class Terminal(Symbol): | |||
| is_term = True | |||
| @property | |||
| def filter_out(self): | |||
| return self.name.startswith('_') | |||
| class NonTerminal(Symbol): | |||
| is_term = False | |||
| class Rule(object): | |||
| """ | |||
| @@ -3,7 +3,7 @@ | |||
| import re | |||
| from .utils import Str, classify | |||
| from .common import is_terminal, PatternStr, PatternRE, TokenDef | |||
| from .common import PatternStr, PatternRE, TokenDef | |||
| ###{standalone | |||
| class LexError(Exception): | |||
| @@ -234,7 +234,7 @@ class ContextualLexer: | |||
| lexer = lexer_by_tokens[key] | |||
| except KeyError: | |||
| accepts = set(accepts) | set(ignore) | set(always_accept) | |||
| state_tokens = [tokens_by_name[n] for n in accepts if is_terminal(n) and n!='$END'] | |||
| state_tokens = [tokens_by_name[n] for n in accepts if n.is_term and n.name!='$END'] | |||
| lexer = Lexer(state_tokens, ignore=ignore, user_callbacks=user_callbacks) | |||
| lexer_by_tokens[key] = lexer | |||
| @@ -12,7 +12,7 @@ from .parse_tree_builder import ParseTreeBuilder | |||
| from .parser_frontends import LALR | |||
| from .parsers.lalr_parser import UnexpectedToken | |||
| from .common import is_terminal, GrammarError, LexerConf, ParserConf, PatternStr, PatternRE, TokenDef | |||
| from .grammar import RuleOptions, Rule | |||
| from .grammar import RuleOptions, Rule, Terminal, NonTerminal | |||
| from .tree import Tree, Transformer, InlineTransformer, Visitor, SlottedTree as ST | |||
| @@ -523,7 +523,9 @@ class Grammar: | |||
| if alias and name.startswith('_'): | |||
| raise GrammarError("Rule %s is marked for expansion (it starts with an underscore) and isn't allowed to have aliases (alias=%s)" % (name, alias)) | |||
| rule = Rule(name, expansion, alias, options) | |||
| expansion = [Terminal(x) if is_terminal(x) else NonTerminal(x) for x in expansion] | |||
| rule = Rule(NonTerminal(name), expansion, alias, options) | |||
| compiled_rules.append(rule) | |||
| return tokens, compiled_rules, self.ignore | |||
| @@ -578,12 +580,16 @@ def options_from_rule(name, *x): | |||
| return name, expansions, RuleOptions(keep_all_tokens, expand1, priority=priority) | |||
| def symbols_from_strcase(expansion): | |||
| return [Terminal(x) if is_terminal(x) else NonTerminal(x) for x in expansion] | |||
| class GrammarLoader: | |||
| def __init__(self): | |||
| tokens = [TokenDef(name, PatternRE(value)) for name, value in TOKENS.items()] | |||
| rules = [options_from_rule(name, x) for name, x in RULES.items()] | |||
| rules = [Rule(r, x.split(), None, o) for r, xs, o in rules for x in xs] | |||
| rules = [options_from_rule(name, x) for name, x in RULES.items()] | |||
| rules = [Rule(NonTerminal(r), symbols_from_strcase(x.split()), None, o) for r, xs, o in rules for x in xs] | |||
| callback = ParseTreeBuilder(rules, ST).create_callback() | |||
| lexer_conf = LexerConf(tokens, ['WS', 'COMMENT']) | |||
| @@ -84,7 +84,7 @@ class ChildFilterLALR(ChildFilter): | |||
| return self.node_builder(filtered) | |||
| def _should_expand(sym): | |||
| return not is_terminal(sym) and sym.startswith('_') | |||
| return not sym.is_term and sym.name.startswith('_') | |||
| def maybe_create_child_filter(expansion, filter_out, ambiguous): | |||
| to_include = [(i, _should_expand(sym)) for i, sym in enumerate(expansion) if sym not in filter_out] | |||
| @@ -109,8 +109,8 @@ class ParseTreeBuilder: | |||
| def _init_builders(self, rules): | |||
| filter_out = {rule.origin for rule in rules if rule.options and rule.options.filter_out} | |||
| filter_out |= {sym for rule in rules for sym in rule.expansion if is_terminal(sym) and sym.startswith('_')} | |||
| assert all(x.startswith('_') for x in filter_out) | |||
| filter_out |= {sym for rule in rules for sym in rule.expansion if sym.is_term and sym.filter_out} | |||
| assert all(t.filter_out for t in filter_out) | |||
| for rule in rules: | |||
| options = rule.options | |||
| @@ -132,9 +132,9 @@ class ParseTreeBuilder: | |||
| callback = Callback() | |||
| for rule, wrapper_chain in self.rule_builders: | |||
| internal_callback_name = '_callback_%s_%s' % (rule.origin, '_'.join(rule.expansion)) | |||
| internal_callback_name = '_callback_%s_%s' % (rule.origin, '_'.join(x.name for x in rule.expansion)) | |||
| user_callback_name = rule.alias or rule.origin | |||
| user_callback_name = rule.alias or rule.origin.name | |||
| try: | |||
| f = transformer._get_func(user_callback_name) | |||
| except AttributeError: | |||
| @@ -1,7 +1,7 @@ | |||
| from ..utils import bfs, fzset, classify | |||
| from ..common import GrammarError, is_terminal | |||
| from ..grammar import Rule | |||
| from ..common import GrammarError | |||
| from ..grammar import Rule, Terminal, NonTerminal | |||
| class RulePtr(object): | |||
| @@ -67,7 +67,7 @@ def calculate_sets(rules): | |||
| FIRST = {} | |||
| FOLLOW = {} | |||
| for sym in symbols: | |||
| FIRST[sym]={sym} if is_terminal(sym) else set() | |||
| FIRST[sym]={sym} if sym.is_term else set() | |||
| FOLLOW[sym]=set() | |||
| # Calculate NULLABLE and FIRST | |||
| @@ -108,16 +108,16 @@ class GrammarAnalyzer(object): | |||
| def __init__(self, parser_conf, debug=False): | |||
| self.debug = debug | |||
| rules = parser_conf.rules + [Rule('$root', [parser_conf.start, '$END'])] | |||
| rules = parser_conf.rules + [Rule(NonTerminal('$root'), [NonTerminal(parser_conf.start), Terminal('$END')])] | |||
| self.rules_by_origin = classify(rules, lambda r: r.origin) | |||
| assert len(rules) == len(set(rules)) | |||
| for r in rules: | |||
| for sym in r.expansion: | |||
| if not (is_terminal(sym) or sym in self.rules_by_origin): | |||
| if not (sym.is_term or sym in self.rules_by_origin): | |||
| raise GrammarError("Using an undefined rule: %s" % sym) # TODO test validation | |||
| self.start_state = self.expand_rule('$root') | |||
| self.start_state = self.expand_rule(NonTerminal('$root')) | |||
| self.FIRST, self.FOLLOW, self.NULLABLE = calculate_sets(rules) | |||
| @@ -125,7 +125,7 @@ class GrammarAnalyzer(object): | |||
| "Returns all init_ptrs accessible by rule (recursive)" | |||
| init_ptrs = set() | |||
| def _expand_rule(rule): | |||
| assert not is_terminal(rule), rule | |||
| assert not rule.is_term, rule | |||
| for r in self.rules_by_origin[rule]: | |||
| init_ptr = RulePtr(r, 0) | |||
| @@ -133,7 +133,7 @@ class GrammarAnalyzer(object): | |||
| if r.expansion: # if not empty rule | |||
| new_r = init_ptr.next | |||
| if not is_terminal(new_r): | |||
| if not new_r.is_term: | |||
| yield new_r | |||
| for _ in bfs([rule], _expand_rule): | |||
| @@ -142,8 +142,8 @@ class GrammarAnalyzer(object): | |||
| return fzset(init_ptrs) | |||
| def _first(self, r): | |||
| if is_terminal(r): | |||
| if r.is_term: | |||
| return {r} | |||
| else: | |||
| return {rp.next for rp in self.expand_rule(r) if is_terminal(rp.next)} | |||
| return {rp.next for rp in self.expand_rule(r) if rp.next.is_term} | |||
| @@ -10,9 +10,9 @@ import logging | |||
| from collections import defaultdict | |||
| from ..utils import classify, classify_bool, bfs, fzset | |||
| from ..common import GrammarError, is_terminal | |||
| from ..common import GrammarError | |||
| from .grammar_analysis import GrammarAnalyzer | |||
| from .grammar_analysis import GrammarAnalyzer, Terminal | |||
| class Action: | |||
| def __init__(self, name): | |||
| @@ -70,12 +70,12 @@ class LALR_Analyzer(GrammarAnalyzer): | |||
| rps = {rp.advance(sym) for rp in rps} | |||
| for rp in set(rps): | |||
| if not rp.is_satisfied and not is_terminal(rp.next): | |||
| if not rp.is_satisfied and not rp.next.is_term: | |||
| rps |= self.expand_rule(rp.next) | |||
| new_state = fzset(rps) | |||
| lookahead[sym].append((Shift, new_state)) | |||
| if sym == '$END': | |||
| if sym == Terminal('$END'): | |||
| self.end_states.append( new_state ) | |||
| yield new_state | |||
| @@ -93,7 +93,7 @@ class LALR_Analyzer(GrammarAnalyzer): | |||
| if not len(v) == 1: | |||
| raise GrammarError("Collision in %s: %s" %(k, ', '.join(['\n * %s: %s' % x for x in v]))) | |||
| self.states[state] = {k:v[0] for k, v in lookahead.items()} | |||
| self.states[state] = {k.name:v[0] for k, v in lookahead.items()} | |||
| for _ in bfs([self.start_state], step): | |||
| pass | |||
| @@ -59,7 +59,7 @@ class _Parser: | |||
| value = self.callbacks[rule](s) | |||
| _action, new_state = get_action(rule.origin) | |||
| _action, new_state = get_action(rule.origin.name) | |||
| assert _action is Shift | |||
| state_stack.append(new_state) | |||
| value_stack.append(value) | |||