From 25a3e8ba59932ccdc1ccf6c9f083d532405ed09f Mon Sep 17 00:00:00 2001 From: Harmen Stoppels Date: Fri, 24 Jan 2025 15:53:42 +0100 Subject: [PATCH] Remove unused Tokenizer.full_match (#48650) --- lib/spack/spack/tokenize.py | 1 - 1 file changed, 1 deletion(-) diff --git a/lib/spack/spack/tokenize.py b/lib/spack/spack/tokenize.py index f42ecb2e95c..f5e9b37803e 100644 --- a/lib/spack/spack/tokenize.py +++ b/lib/spack/spack/tokenize.py @@ -50,7 +50,6 @@ class Tokenizer: def __init__(self, tokens: Type[TokenBase]): self.tokens = tokens self.regex = re.compile("|".join(f"(?P<{token}>{token.regex})" for token in tokens)) - self.full_match = True def tokenize(self, text: str) -> Generator[Token, None, None]: if not text: