From f4876322e7fe36923d4eb89757a844b3d90be306 Mon Sep 17 00:00:00 2001 From: Andi Albrecht Date: Mon, 12 Nov 2012 06:43:32 +0100 Subject: PEP8'ified to code a bit. --- tests/test_filters.py | 13 +++++++---- tests/test_format.py | 2 +- tests/test_functions.py | 20 +++++++++------- tests/test_parse.py | 2 +- tests/test_pipeline.py | 47 ++++++++++++++++++------------------- tests/test_split.py | 3 ++- tests/test_tokenize.py | 61 +++++++++++++++++++++++++------------------------ 7 files changed, 79 insertions(+), 69 deletions(-) (limited to 'tests') diff --git a/tests/test_filters.py b/tests/test_filters.py index 26f0b52..d827454 100644 --- a/tests/test_filters.py +++ b/tests/test_filters.py @@ -6,7 +6,7 @@ Created on 24/03/2012 import unittest from sqlparse.filters import StripWhitespace, Tokens2Unicode -from sqlparse.lexer import tokenize +from sqlparse.lexer import tokenize class Test__StripWhitespace(unittest.TestCase): @@ -49,17 +49,20 @@ GROUP BY dir_entries.inode LIMIT 1""" def test_StripWhitespace1(self): - self.assertEqual(Tokens2Unicode(StripWhitespace(tokenize(self.sql))), + self.assertEqual( + Tokens2Unicode(StripWhitespace(tokenize(self.sql))), 'INSERT INTO dir_entries(type)VALUES(:type);INSERT INTO ' 'directories(inode)VALUES(:inode)LIMIT 1') def test_StripWhitespace2(self): - self.assertEqual(Tokens2Unicode(StripWhitespace(tokenize(self.sql2))), + self.assertEqual( + Tokens2Unicode(StripWhitespace(tokenize(self.sql2))), 'SELECT child_entry,asdf AS inode,creation FROM links WHERE ' 'parent_dir==:parent_dir AND name==:name LIMIT 1') def test_StripWhitespace3(self): - self.assertEqual(Tokens2Unicode(StripWhitespace(tokenize(self.sql3))), + self.assertEqual( + Tokens2Unicode(StripWhitespace(tokenize(self.sql3))), 'SELECT 0 AS st_dev,0 AS st_uid,0 AS st_gid,dir_entries.type AS ' 'st_mode,dir_entries.inode AS st_ino,COUNT(links.child_entry)AS ' 'st_nlink,:creation AS st_ctime,dir_entries.access AS st_atime,' @@ -72,4 +75,4 @@ LIMIT 1""" if __name__ == "__main__": #import sys;sys.argv = ['', 'Test.testName'] - unittest.main() \ No newline at end of file + unittest.main() diff --git a/tests/test_format.py b/tests/test_format.py index 9c764d7..c33ac93 100644 --- a/tests/test_format.py +++ b/tests/test_format.py @@ -91,7 +91,7 @@ class TestFormatReindent(TestCaseBase): self.assertRaises(SQLParseError, sqlparse.format, 'foo', reindent=True, indent_width='foo') self.assertRaises(SQLParseError, sqlparse.format, 'foo', - reindent=True, indent_width= -12) + reindent=True, indent_width=-12) def test_stmts(self): f = lambda sql: sqlparse.format(sql, reindent=True) diff --git a/tests/test_functions.py b/tests/test_functions.py index aa382ce..52e2ce7 100644 --- a/tests/test_functions.py +++ b/tests/test_functions.py @@ -6,12 +6,12 @@ Created on 13/02/2012 from unittest import main, TestCase from sqlparse.filters import IncludeStatement, Tokens2Unicode -from sqlparse.lexer import tokenize +from sqlparse.lexer import tokenize import sys sys.path.insert(0, '..') -from sqlparse.filters import compact +from sqlparse.filters import compact from sqlparse.functions import getcolumns, getlimit, IsType @@ -27,15 +27,17 @@ class Test_IncludeStatement(TestCase): def test_includeStatement(self): stream = tokenize(self.sql) - includeStatement = IncludeStatement('tests/files', raiseexceptions=True) + includeStatement = IncludeStatement('tests/files', + raiseexceptions=True) stream = includeStatement.process(None, stream) stream = compact(stream) result = Tokens2Unicode(stream) - self.assertEqual(result, - 'INSERT INTO dir_entries(type)VALUES(:type);INSERT INTO ' - 'directories(inode)VALUES(:inode)LIMIT 1') + self.assertEqual( + result, ( + 'INSERT INTO dir_entries(type)VALUES(:type);INSERT INTO ' + 'directories(inode)VALUES(:inode)LIMIT 1')) class Test_SQL(TestCase): @@ -96,7 +98,8 @@ class Test_Compact(Test_SQL): result = compact(stream) - self.assertEqual(Tokens2Unicode(result), + self.assertEqual( + Tokens2Unicode(result), 'SELECT child_entry,asdf AS inode,creation FROM links WHERE ' 'parent_dir==:parent_dir AND name==:name LIMIT 1') @@ -105,7 +108,8 @@ class Test_Compact(Test_SQL): result = compact(stream) - self.assertEqual(Tokens2Unicode(result), + self.assertEqual( + Tokens2Unicode(result), 'SELECT 0 AS st_dev,0 AS st_uid,0 AS st_gid,dir_entries.type AS ' 'st_mode,dir_entries.inode AS st_ino,COUNT(links.child_entry)AS ' 'st_nlink,:creation AS st_ctime,dir_entries.access AS st_atime,' diff --git a/tests/test_parse.py b/tests/test_parse.py index 05141a4..59b8e72 100644 --- a/tests/test_parse.py +++ b/tests/test_parse.py @@ -20,7 +20,7 @@ class SQLParseTest(TestCaseBase): def test_multistatement(self): sql1 = 'select * from foo;' sql2 = 'select * from bar;' - stmts = sqlparse.parse(sql1+sql2) + stmts = sqlparse.parse(sql1 + sql2) self.assertEqual(len(stmts), 2) self.assertEqual(str(stmts[0]), sql1) self.assertEqual(str(stmts[1]), sql2) diff --git a/tests/test_pipeline.py b/tests/test_pipeline.py index 56eed4a..3442a5b 100644 --- a/tests/test_pipeline.py +++ b/tests/test_pipeline.py @@ -4,6 +4,7 @@ from sqlparse.filters import ColumnsSelect from sqlparse.lexer import tokenize from sqlparse.pipeline import Pipeline + class Test(unittest.TestCase): def setUp(self): @@ -34,35 +35,35 @@ class Test(unittest.TestCase): def test_3(self): sql = """ - SELECT - 0 AS st_dev, - 0 AS st_uid, - 0 AS st_gid, +SELECT +0 AS st_dev, +0 AS st_uid, +0 AS st_gid, - dir_entries.type AS st_mode, - dir_entries.inode AS st_ino, - COUNT(links.child_entry) AS st_nlink, +dir_entries.type AS st_mode, +dir_entries.inode AS st_ino, +COUNT(links.child_entry) AS st_nlink, - :creation AS st_ctime, - dir_entries.access AS st_atime, - dir_entries.modification AS st_mtime, - -- :creation AS st_ctime, - -- CAST(STRFTIME('%s',dir_entries.access) AS INTEGER) AS st_atime, - -- CAST(STRFTIME('%s',dir_entries.modification) AS INTEGER) AS st_mtime, +:creation AS st_ctime, +dir_entries.access AS st_atime, +dir_entries.modification AS st_mtime, +-- :creation AS st_ctime, +-- CAST(STRFTIME('%s',dir_entries.access) AS INTEGER) AS st_atime, +-- CAST(STRFTIME('%s',dir_entries.modification) AS INTEGER) AS st_mtime, - COALESCE(files.size,0) AS st_size, -- Python-FUSE - COALESCE(files.size,0) AS size -- PyFilesystem +COALESCE(files.size,0) AS st_size, -- Python-FUSE +COALESCE(files.size,0) AS size -- PyFilesystem - FROM dir_entries - LEFT JOIN files - ON dir_entries.inode == files.inode - LEFT JOIN links - ON dir_entries.inode == links.child_entry +FROM dir_entries +LEFT JOIN files +ON dir_entries.inode == files.inode +LEFT JOIN links +ON dir_entries.inode == links.child_entry - WHERE dir_entries.inode == :inode +WHERE dir_entries.inode == :inode - GROUP BY dir_entries.inode - LIMIT 1""" +GROUP BY dir_entries.inode +LIMIT 1""" self.assertEqual([u'st_dev', u'st_uid', u'st_gid', u'st_mode', u'st_ino', u'st_nlink', u'st_ctime', u'st_atime', u'st_mtime', u'st_size', u'size'], diff --git a/tests/test_split.py b/tests/test_split.py index 321fca2..f0b6dda 100644 --- a/tests/test_split.py +++ b/tests/test_split.py @@ -111,7 +111,8 @@ class SQLSplitTest(TestCaseBase): def test_if_function(self): # see issue 33 # don't let IF as a function confuse the splitter - sql = ('CREATE TEMPORARY TABLE tmp SELECT IF(a=1, a, b) AS o FROM one; ' + sql = ('CREATE TEMPORARY TABLE tmp ' + 'SELECT IF(a=1, a, b) AS o FROM one; ' 'SELECT t FROM two') stmts = sqlparse.split(sql) self.assertEqual(len(stmts), 2) diff --git a/tests/test_tokenize.py b/tests/test_tokenize.py index 5b403f9..417aef6 100644 --- a/tests/test_tokenize.py +++ b/tests/test_tokenize.py @@ -12,9 +12,9 @@ from sqlparse.tokens import * class TestTokenize(unittest.TestCase): def test_simple(self): - sql = 'select * from foo;' - stream = lexer.tokenize(sql) - self.assert_(type(stream) is types.GeneratorType) + s = 'select * from foo;' + stream = lexer.tokenize(s) + self.assert_(isinstance(stream, types.GeneratorType)) tokens = list(stream) self.assertEqual(len(tokens), 8) self.assertEqual(len(tokens[0]), 2) @@ -22,60 +22,61 @@ class TestTokenize(unittest.TestCase): self.assertEqual(tokens[-1], (Punctuation, u';')) def test_backticks(self): - sql = '`foo`.`bar`' - tokens = list(lexer.tokenize(sql)) + s = '`foo`.`bar`' + tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 3) self.assertEqual(tokens[0], (Name, u'`foo`')) def test_linebreaks(self): # issue1 - sql = 'foo\nbar\n' - tokens = lexer.tokenize(sql) - self.assertEqual(''.join(str(x[1]) for x in tokens), sql) - sql = 'foo\rbar\r' - tokens = lexer.tokenize(sql) - self.assertEqual(''.join(str(x[1]) for x in tokens), sql) - sql = 'foo\r\nbar\r\n' - tokens = lexer.tokenize(sql) - self.assertEqual(''.join(str(x[1]) for x in tokens), sql) - sql = 'foo\r\nbar\n' - tokens = lexer.tokenize(sql) - self.assertEqual(''.join(str(x[1]) for x in tokens), sql) + s = 'foo\nbar\n' + tokens = lexer.tokenize(s) + self.assertEqual(''.join(str(x[1]) for x in tokens), s) + s = 'foo\rbar\r' + tokens = lexer.tokenize(s) + self.assertEqual(''.join(str(x[1]) for x in tokens), s) + s = 'foo\r\nbar\r\n' + tokens = lexer.tokenize(s) + self.assertEqual(''.join(str(x[1]) for x in tokens), s) + s = 'foo\r\nbar\n' + tokens = lexer.tokenize(s) + self.assertEqual(''.join(str(x[1]) for x in tokens), s) def test_inline_keywords(self): # issue 7 - sql = "create created_foo" - tokens = list(lexer.tokenize(sql)) + s = "create created_foo" + tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 3) self.assertEqual(tokens[0][0], Keyword.DDL) self.assertEqual(tokens[2][0], Name) self.assertEqual(tokens[2][1], u'created_foo') - sql = "enddate" - tokens = list(lexer.tokenize(sql)) + s = "enddate" + tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 1) self.assertEqual(tokens[0][0], Name) - sql = "join_col" - tokens = list(lexer.tokenize(sql)) + s = "join_col" + tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 1) self.assertEqual(tokens[0][0], Name) - sql = "left join_col" - tokens = list(lexer.tokenize(sql)) + s = "left join_col" + tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 3) self.assertEqual(tokens[2][0], Name) self.assertEqual(tokens[2][1], 'join_col') def test_negative_numbers(self): - sql = "values(-1)" - tokens = list(lexer.tokenize(sql)) + s = "values(-1)" + tokens = list(lexer.tokenize(s)) self.assertEqual(len(tokens), 4) self.assertEqual(tokens[2][0], Number.Integer) self.assertEqual(tokens[2][1], '-1') def test_tab_expansion(self): - sql = "\t" + s = "\t" lex = lexer.Lexer() lex.tabsize = 5 - tokens = list(lex.get_tokens(sql)) + tokens = list(lex.get_tokens(s)) self.assertEqual(tokens[0][1], " " * 5) + class TestToken(unittest.TestCase): def test_str(self): @@ -124,9 +125,9 @@ class TestTokenList(unittest.TestCase): self.assertEqual(x.token_matching(1, [lambda t: t.ttype is Keyword]), None) + class TestStream(unittest.TestCase): def test_simple(self): - import types from cStringIO import StringIO stream = StringIO("SELECT 1; SELECT 2;") -- cgit v1.2.1