From af37d16bae4f1c96d3936771d74424d1d54ddf74 Mon Sep 17 00:00:00 2001 From: Victor Uriarte Date: Sun, 29 May 2016 11:26:43 -0700 Subject: Fix flake8 styling --- sqlparse/engine/__init__.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) (limited to 'sqlparse/engine/__init__.py') diff --git a/sqlparse/engine/__init__.py b/sqlparse/engine/__init__.py index 4d7fe88..1268bd1 100644 --- a/sqlparse/engine/__init__.py +++ b/sqlparse/engine/__init__.py @@ -43,8 +43,8 @@ class FilterStack(object): for filter_ in self.preprocess: stream = filter_.process(self, stream) - if self.stmtprocess or self.postprocess or self.split_statements \ - or self._grouping: + if (self.stmtprocess or self.postprocess or + self.split_statements or self._grouping): splitter = StatementFilter() stream = splitter.process(self, stream) -- cgit v1.2.1 From 81e408a4f2c8281e0da7c93ed75da90df1a518e2 Mon Sep 17 00:00:00 2001 From: Victor Uriarte Date: Thu, 2 Jun 2016 13:53:26 -0700 Subject: Remove redundant functions --- sqlparse/engine/__init__.py | 16 +--------------- 1 file changed, 1 insertion(+), 15 deletions(-) (limited to 'sqlparse/engine/__init__.py') diff --git a/sqlparse/engine/__init__.py b/sqlparse/engine/__init__.py index 1268bd1..c6b8b0c 100644 --- a/sqlparse/engine/__init__.py +++ b/sqlparse/engine/__init__.py @@ -9,9 +9,6 @@ from sqlparse import lexer from sqlparse.engine import grouping from sqlparse.engine.filter import StatementFilter -# XXX remove this when cleanup is complete -Filter = object - class FilterStack(object): @@ -22,20 +19,9 @@ class FilterStack(object): self.split_statements = False self._grouping = False - def _flatten(self, stream): - for token in stream: - if token.is_group(): - for t in self._flatten(token.tokens): - yield t - else: - yield token - def enable_grouping(self): self._grouping = True - def full_analyze(self): - self.enable_grouping() - def run(self, sql, encoding=None): stream = lexer.tokenize(sql, encoding) # Process token stream @@ -71,7 +57,7 @@ class FilterStack(object): def _run2(stream): for stmt in stream: - stmt.tokens = list(self._flatten(stmt.tokens)) + stmt.tokens = list(stmt.flatten()) for filter_ in self.postprocess: stmt = filter_.process(self, stmt) yield stmt -- cgit v1.2.1