From 76e870969ca2ae306f21dbdb790c770dae6067f6 Mon Sep 17 00:00:00 2001 From: Chris Guidry Date: Sat, 13 Jan 2024 15:21:34 -0500 Subject: [PATCH] Bumping linters --- .pre-commit-config.yaml | 8 ++++---- src/measured/_parser.py | 17 ----------------- 2 files changed, 4 insertions(+), 21 deletions(-) diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 28b2fa0..17db7d7 100644 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -7,19 +7,19 @@ repos: - id: check-toml - id: check-yaml - repo: https://github.com/PyCQA/isort - rev: "5.12.0" + rev: "5.13.2" hooks: - id: isort - repo: https://github.com/psf/black - rev: "22.6.0" + rev: "23.12.1" hooks: - id: black-jupyter - repo: https://github.com/PyCQA/flake8 - rev: "4.0.1" + rev: "7.0.0" hooks: - id: flake8 - repo: https://github.com/pre-commit/mirrors-mypy - rev: "v1.0.0" + rev: "v1.8.0" hooks: - id: mypy additional_dependencies: diff --git a/src/measured/_parser.py b/src/measured/_parser.py index 30e2c13..8987fd7 100644 --- a/src/measured/_parser.py +++ b/src/measured/_parser.py @@ -139,7 +139,6 @@ def match_examples( self, (UnexpectedToken, UnexpectedEOF) ) and isinstance(ut, (UnexpectedToken, UnexpectedEOF)): if ut.token == self.token: ## - logger.debug("Exact Match at example [%s][%s]" % (i, j)) return label @@ -357,7 +356,6 @@ def classify(seq, key=None, value=None): def _deserialize(data, namespace, memo): if isinstance(data, dict): if "__type__" in data: ## - class_ = namespace[data["__type__"]] return class_.deserialize(data, memo) elif "@" in data: @@ -482,7 +480,6 @@ def get_regexp_width(expr): class Meta: - empty: bool line: int column: int @@ -642,7 +639,6 @@ def _apply_v_args(cls, visit_wrapper): assert mro[0] is cls libmembers = {name for _cls in mro[1:] for name, _ in getmembers(_cls)} for name, value in getmembers(cls): - ## if name.startswith("_") or ( @@ -776,7 +772,6 @@ def _call_userfunc(self, tree, new_children=None): class TransformerChain(Generic[_Leaf_T, _Return_T]): - transformers: "Tuple[Union[Transformer, TransformerChain], ...]" def __init__(self, *transformers: "Union[Transformer, TransformerChain]") -> None: @@ -797,7 +792,6 @@ def __mul__( class Transformer_InPlace(Transformer): # -- def _transform_tree(self, tree): ## - return self._call_userfunc(tree) def transform(self, tree: Tree[_Leaf_T]) -> _Return_T: @@ -1199,7 +1193,6 @@ def __eq__(self, other): class Pattern(Serialize, ABC): - value: str flags: Collection[str] raw: Optional[str] @@ -1309,7 +1302,6 @@ def __repr__(self): def user_repr(self) -> str: if self.name.startswith("__"): ## - return self.pattern.raw or self.name else: return self.name @@ -1522,7 +1514,6 @@ def _build_mres(self, terminals, max_size): try: mre = self.re_.compile(pattern, self.g_regex_flags) except AssertionError: ## - return self._build_mres(terminals, max_size // 2) mres.append((mre, {i: n for n, i in mre.groupindex.items()})) @@ -1604,7 +1595,6 @@ def make_lexer_state(self, text): class BasicLexer(Lexer): - terminals: Collection[TerminalDef] ignore_types: FrozenSet[str] newline_types: FrozenSet[str] @@ -1745,7 +1735,6 @@ def next_token(self, lex_state: LexerState, parser_state: Any = None) -> Token: class ContextualLexer(Lexer): - lexers: Dict[str, BasicLexer] root_lexer: BasicLexer @@ -2021,7 +2010,6 @@ def __call__(self, children): if filtered: filtered += children[i].children else: ## - filtered = children[i].children else: filtered.append(children[i]) @@ -2045,7 +2033,6 @@ def __call__(self, children): if filtered: filtered += children[i].children else: ## - filtered = children[i].children else: filtered.append(children[i]) @@ -2271,7 +2258,6 @@ def default_callback(data, children): default_callback = self.tree_class for rule, wrapper_chain in self.rule_builders: - user_callback_name = ( rule.alias or rule.options.template_source or rule.origin.name ) @@ -2631,7 +2617,6 @@ def __init__(self, lexer_conf, parser_conf, options, parser=None): ## if parser: ## - self.parser = parser else: create_parser = _parser_creators.get(parser_conf.parser_type) @@ -2704,7 +2689,6 @@ def parse_interactive(self, text=None, start=None): def _validate_frontend_args(parser, lexer) -> None: assert_config(parser, ("lalr", "earley", "cyk")) if not isinstance(lexer, type): ## - expected = { "lalr": ("basic", "contextual"), "earley": ("basic", "dynamic", "dynamic_complete"), @@ -3082,7 +3066,6 @@ def __init__(self, grammar: "Union[Grammar, str, IO[str]]", **options) -> None: self._load(cached_parser_data, **options) return except Exception: ## - logger.exception( "Failed to load Lark from cache: %r. We will try to carry on." % cache_fn