diff options
| author | Mike Bayer <mike_mp@zzzcomputing.com> | 2013-06-03 13:13:16 -0400 |
|---|---|---|
| committer | Mike Bayer <mike_mp@zzzcomputing.com> | 2013-06-03 13:13:16 -0400 |
| commit | 74c98bf182a1cac1ca1837da69e1c0550beaaab5 (patch) | |
| tree | 4815d198d2aa4a6497330fb5d81e53bf4acfbb2d /test | |
| parent | ff399ac75074916045410cedae72489cb60e8b50 (diff) | |
| parent | c2a158c137ee07a146f02e5ee89ec42e486c6a37 (diff) | |
| download | sqlalchemy-74c98bf182a1cac1ca1837da69e1c0550beaaab5.tar.gz | |
Merge branch 'master' into ticket_1068
Diffstat (limited to 'test')
94 files changed, 1554 insertions, 2633 deletions
diff --git a/test/aaa_profiling/test_compiler.py b/test/aaa_profiling/test_compiler.py index 1b7798d06..8304296da 100644 --- a/test/aaa_profiling/test_compiler.py +++ b/test/aaa_profiling/test_compiler.py @@ -29,7 +29,7 @@ class CompileTest(fixtures.TestBase, AssertsExecutionResults): for c in t.c: c.type._type_affinity from sqlalchemy import types - for t in types._type_map.values(): + for t in list(types._type_map.values()): t._type_affinity cls.dialect = default.DefaultDialect() diff --git a/test/aaa_profiling/test_memusage.py b/test/aaa_profiling/test_memusage.py index 57bddc859..20c6f0a65 100644 --- a/test/aaa_profiling/test_memusage.py +++ b/test/aaa_profiling/test_memusage.py @@ -47,7 +47,7 @@ def profile_memory(times=50): gc_collect() samples[x] = len(get_objects_skipping_sqlite_issue()) - print "sample gc sizes:", samples + print("sample gc sizes:", samples) assert len(_sessions) == 0 diff --git a/test/aaa_profiling/test_orm.py b/test/aaa_profiling/test_orm.py index b9eeb8361..199b96e5f 100644 --- a/test/aaa_profiling/test_orm.py +++ b/test/aaa_profiling/test_orm.py @@ -144,7 +144,7 @@ class LoadManyToOneFromIdentityTest(fixtures.MappedTest): child.insert().execute([ {'id':i, 'data':'c%d' % i} - for i in xrange(1, 251) + for i in range(1, 251) ]) parent.insert().execute([ { @@ -152,7 +152,7 @@ class LoadManyToOneFromIdentityTest(fixtures.MappedTest): 'data':'p%dc%d' % (i, (i % 250) + 1), 'child_id':(i % 250) + 1 } - for i in xrange(1, 1000) + for i in range(1, 1000) ]) def test_many_to_one_load_no_identity(self): @@ -234,11 +234,11 @@ class MergeBackrefsTest(fixtures.MappedTest): s = Session() s.add_all([ A(id=i, - bs=[B(id=(i * 5) + j) for j in xrange(1, 5)], + bs=[B(id=(i * 5) + j) for j in range(1, 5)], c=C(id=i), - ds=[D(id=(i * 5) + j) for j in xrange(1, 5)] + ds=[D(id=(i * 5) + j) for j in range(1, 5)] ) - for i in xrange(1, 5) + for i in range(1, 5) ]) s.commit() @@ -249,11 +249,11 @@ class MergeBackrefsTest(fixtures.MappedTest): s = Session() for a in [ A(id=i, - bs=[B(id=(i * 5) + j) for j in xrange(1, 5)], + bs=[B(id=(i * 5) + j) for j in range(1, 5)], c=C(id=i), - ds=[D(id=(i * 5) + j) for j in xrange(1, 5)] + ds=[D(id=(i * 5) + j) for j in range(1, 5)] ) - for i in xrange(1, 5) + for i in range(1, 5) ]: s.merge(a) diff --git a/test/aaa_profiling/test_resultset.py b/test/aaa_profiling/test_resultset.py index 0146d1b08..27e60410d 100644 --- a/test/aaa_profiling/test_resultset.py +++ b/test/aaa_profiling/test_resultset.py @@ -2,6 +2,7 @@ from sqlalchemy import * from sqlalchemy.testing import fixtures, AssertsExecutionResults, profiling from sqlalchemy import testing from sqlalchemy.testing import eq_ +from sqlalchemy.util import u NUM_FIELDS = 10 NUM_RECORDS = 1000 @@ -19,10 +20,10 @@ class ResultSetTest(fixtures.TestBase, AssertsExecutionResults): def setup(self): metadata.create_all() - t.insert().execute([dict(('field%d' % fnum, u'value%d' % fnum) + t.insert().execute([dict(('field%d' % fnum, u('value%d' % fnum)) for fnum in range(NUM_FIELDS)) for r_num in range(NUM_RECORDS)]) - t2.insert().execute([dict(('field%d' % fnum, u'value%d' % fnum) + t2.insert().execute([dict(('field%d' % fnum, u('value%d' % fnum)) for fnum in range(NUM_FIELDS)) for r_num in range(NUM_RECORDS)]) @@ -88,7 +89,7 @@ class RowProxyTest(fixtures.TestBase): keymap = {} for index, (keyobjs, processor, values) in \ - enumerate(zip(keys, processors, row)): + enumerate(list(zip(keys, processors, row))): for key in keyobjs: keymap[key] = (processor, key, index) keymap[index] = (processor, key, index) diff --git a/test/aaa_profiling/test_zoomark.py b/test/aaa_profiling/test_zoomark.py index 5fc9ffc2a..145f3c594 100644 --- a/test/aaa_profiling/test_zoomark.py +++ b/test/aaa_profiling/test_zoomark.py @@ -1,3 +1,5 @@ +from __future__ import unicode_literals + """Benchmark for SQLAlchemy. An adaptation of Robert Brewers' ZooMark speed tests. """ @@ -72,68 +74,68 @@ class ZooMarkTest(fixtures.TestBase): Zoo = metadata.tables['Zoo'] Animal = metadata.tables['Animal'] engine = metadata.bind - wap = engine.execute(Zoo.insert(), Name=u'Wild Animal Park', + wap = engine.execute(Zoo.insert(), Name='Wild Animal Park', Founded=datetime.date(2000, 1, 1), Opens=datetime.time(8, 15, 59), LastEscape= datetime.datetime(2004, 7, 29, 5, 6, 7), Admission=4.95).inserted_primary_key[0] - sdz = engine.execute(Zoo.insert(), Name=u'San Diego Zoo', + sdz = engine.execute(Zoo.insert(), Name='San Diego Zoo', Founded=datetime.date(1935, 9, 13), Opens=datetime.time(9, 0, 0), Admission=0).inserted_primary_key[0] - engine.execute(Zoo.insert(inline=True), Name=u'Montr\xe9al Biod\xf4me', + engine.execute(Zoo.insert(inline=True), Name='Montr\xe9al Biod\xf4me', Founded=datetime.date(1992, 6, 19), Opens=datetime.time(9, 0, 0), Admission=11.75) - seaworld = engine.execute(Zoo.insert(), Name=u'Sea_World', + seaworld = engine.execute(Zoo.insert(), Name='Sea_World', Admission=60).inserted_primary_key[0] # Let's add a crazy futuristic Zoo to test large date values. - lp = engine.execute(Zoo.insert(), Name=u'Luna Park', + lp = engine.execute(Zoo.insert(), Name='Luna Park', Founded=datetime.date(2072, 7, 17), Opens=datetime.time(0, 0, 0), Admission=134.95).inserted_primary_key[0] # Animals - leopardid = engine.execute(Animal.insert(), Species=u'Leopard', + leopardid = engine.execute(Animal.insert(), Species='Leopard', Lifespan=73.5).inserted_primary_key[0] engine.execute(Animal.update(Animal.c.ID == leopardid), ZooID=wap, LastEscape=datetime.datetime( 2004, 12, 21, 8, 15, 0, 999907,) ) - lion = engine.execute(Animal.insert(), Species=u'Lion', + lion = engine.execute(Animal.insert(), Species='Lion', ZooID=wap).inserted_primary_key[0] - engine.execute(Animal.insert(), Species=u'Slug', Legs=1, Lifespan=.75) - tiger = engine.execute(Animal.insert(), Species=u'Tiger', + engine.execute(Animal.insert(), Species='Slug', Legs=1, Lifespan=.75) + tiger = engine.execute(Animal.insert(), Species='Tiger', ZooID=sdz).inserted_primary_key[0] # Override Legs.default with itself just to make sure it works. - engine.execute(Animal.insert(inline=True), Species=u'Bear', Legs=4) - engine.execute(Animal.insert(inline=True), Species=u'Ostrich', Legs=2, + engine.execute(Animal.insert(inline=True), Species='Bear', Legs=4) + engine.execute(Animal.insert(inline=True), Species='Ostrich', Legs=2, Lifespan=103.2) - engine.execute(Animal.insert(inline=True), Species=u'Centipede', + engine.execute(Animal.insert(inline=True), Species='Centipede', Legs=100) - emp = engine.execute(Animal.insert(), Species=u'Emperor Penguin', + emp = engine.execute(Animal.insert(), Species='Emperor Penguin', Legs=2, ZooID=seaworld).inserted_primary_key[0] - adelie = engine.execute(Animal.insert(), Species=u'Adelie Penguin', + adelie = engine.execute(Animal.insert(), Species='Adelie Penguin', Legs=2, ZooID=seaworld).inserted_primary_key[0] - engine.execute(Animal.insert(inline=True), Species=u'Millipede', + engine.execute(Animal.insert(inline=True), Species='Millipede', Legs=1000000, ZooID=sdz) # Add a mother and child to test relationships - bai_yun = engine.execute(Animal.insert(), Species=u'Ape', - Name=u'Bai Yun', Legs=2).inserted_primary_key[0] - engine.execute(Animal.insert(inline=True), Species=u'Ape', - Name=u'Hua Mei', Legs=2, MotherID=bai_yun) + bai_yun = engine.execute(Animal.insert(), Species='Ape', + Name='Bai Yun', Legs=2).inserted_primary_key[0] + engine.execute(Animal.insert(inline=True), Species='Ape', + Name='Hua Mei', Legs=2, MotherID=bai_yun) def test_baseline_2_insert(self): Animal = metadata.tables['Animal'] i = Animal.insert(inline=True) - for x in xrange(ITERATIONS): - tick = i.execute(Species=u'Tick', Name=u'Tick %d' % x, + for x in range(ITERATIONS): + tick = i.execute(Species='Tick', Name='Tick %d' % x, Legs=8) def test_baseline_3_properties(self): @@ -146,28 +148,28 @@ class ZooMarkTest(fixtures.TestBase): return list(engine.execute(select).first()) - for x in xrange(ITERATIONS): + for x in range(ITERATIONS): # Zoos WAP = fullobject(Zoo.select(Zoo.c.Name - == u'Wild Animal Park')) + == 'Wild Animal Park')) SDZ = fullobject(Zoo.select(Zoo.c.Founded == datetime.date(1935, 9, 13))) Biodome = fullobject(Zoo.select(Zoo.c.Name - == u'Montr\xe9al Biod\xf4me')) + == 'Montr\xe9al Biod\xf4me')) seaworld = fullobject(Zoo.select(Zoo.c.Admission == float(60))) # Animals leopard = fullobject(Animal.select(Animal.c.Species - == u'Leopard')) + == 'Leopard')) ostrich = fullobject(Animal.select(Animal.c.Species - == u'Ostrich')) + == 'Ostrich')) millipede = fullobject(Animal.select(Animal.c.Legs == 1000000)) - ticks = fullobject(Animal.select(Animal.c.Species == u'Tick' + ticks = fullobject(Animal.select(Animal.c.Species == 'Tick' )) def test_baseline_4_expressions(self): @@ -180,7 +182,7 @@ class ZooMarkTest(fixtures.TestBase): return [list(row) for row in engine.execute(select).fetchall()] - for x in xrange(ITERATIONS): + for x in range(ITERATIONS): assert len(fulltable(Zoo.select())) == 5 assert len(fulltable(Animal.select())) == ITERATIONS + 12 assert len(fulltable(Animal.select(Animal.c.Legs == 4))) \ @@ -194,9 +196,9 @@ class ZooMarkTest(fixtures.TestBase): assert len(fulltable(Animal.select(Animal.c.Lifespan > 70))) == 2 assert len(fulltable(Animal.select(Animal.c.Species. - startswith(u'L')))) == 2 + startswith('L')))) == 2 assert len(fulltable(Animal.select(Animal.c.Species. - endswith(u'pede')))) == 2 + endswith('pede')))) == 2 assert len(fulltable(Animal.select(Animal.c.LastEscape != None))) == 1 assert len(fulltable(Animal.select(None @@ -204,10 +206,10 @@ class ZooMarkTest(fixtures.TestBase): # In operator (containedby) - assert len(fulltable(Animal.select(Animal.c.Species.like(u'%pede%' + assert len(fulltable(Animal.select(Animal.c.Species.like('%pede%' )))) == 2 - assert len(fulltable(Animal.select(Animal.c.Species.in_([u'Lion' - , u'Tiger', u'Bear'])))) == 3 + assert len(fulltable(Animal.select(Animal.c.Species.in_(['Lion' + , 'Tiger', 'Bear'])))) == 3 # Try In with cell references class thing(object): @@ -215,20 +217,20 @@ class ZooMarkTest(fixtures.TestBase): pet, pet2 = thing(), thing() - pet.Name, pet2.Name = u'Slug', u'Ostrich' + pet.Name, pet2.Name = 'Slug', 'Ostrich' assert len(fulltable(Animal.select(Animal.c.Species.in_([pet.Name, pet2.Name])))) == 2 # logic and other functions - assert len(fulltable(Animal.select(Animal.c.Species.like(u'Slug' + assert len(fulltable(Animal.select(Animal.c.Species.like('Slug' )))) == 1 - assert len(fulltable(Animal.select(Animal.c.Species.like(u'%pede%' + assert len(fulltable(Animal.select(Animal.c.Species.like('%pede%' )))) == 2 - name = u'Lion' + name = 'Lion' assert len(fulltable(Animal.select(func.length(Animal.c.Species) == len(name)))) == ITERATIONS + 3 - assert len(fulltable(Animal.select(Animal.c.Species.like(u'%i%' + assert len(fulltable(Animal.select(Animal.c.Species.like('%i%' )))) == ITERATIONS + 7 # Test now(), today(), year(), month(), day() @@ -250,7 +252,7 @@ class ZooMarkTest(fixtures.TestBase): Zoo = metadata.tables['Zoo'] engine = metadata.bind - for x in xrange(ITERATIONS): + for x in range(ITERATIONS): # views @@ -274,7 +276,7 @@ class ZooMarkTest(fixtures.TestBase): for species, lifespan in engine.execute(select([Animal.c.Species, Animal.c.Lifespan])).fetchall(): assert lifespan == expected[species] - expected = [u'Montr\xe9al Biod\xf4me', 'Wild Animal Park'] + expected = ['Montr\xe9al Biod\xf4me', 'Wild Animal Park'] e = select([Zoo.c.Name], and_(Zoo.c.Founded != None, Zoo.c.Founded <= func.current_timestamp(), Zoo.c.Founded >= datetime.date(1990, 1, 1))) @@ -290,21 +292,21 @@ class ZooMarkTest(fixtures.TestBase): def test_baseline_6_editing(self): Zoo = metadata.tables['Zoo'] engine = metadata.bind - for x in xrange(ITERATIONS): + for x in range(ITERATIONS): # Edit - SDZ = engine.execute(Zoo.select(Zoo.c.Name == u'San Diego Zoo' + SDZ = engine.execute(Zoo.select(Zoo.c.Name == 'San Diego Zoo' )).first() engine.execute(Zoo.update(Zoo.c.ID == SDZ['ID' - ]), Name=u'The San Diego Zoo', + ]), Name='The San Diego Zoo', Founded=datetime.date(1900, 1, 1), Opens=datetime.time(7, 30, 0), Admission='35.00') # Test edits - SDZ = engine.execute(Zoo.select(Zoo.c.Name == u'The San Diego Zoo' + SDZ = engine.execute(Zoo.select(Zoo.c.Name == 'The San Diego Zoo' )).first() assert SDZ['Founded'] == datetime.date(1900, 1, 1), \ SDZ['Founded'] @@ -312,14 +314,14 @@ class ZooMarkTest(fixtures.TestBase): # Change it back engine.execute(Zoo.update(Zoo.c.ID == SDZ['ID' - ]), Name=u'San Diego Zoo', + ]), Name='San Diego Zoo', Founded=datetime.date(1935, 9, 13), Opens=datetime.time(9, 0, 0), Admission='0') # Test re-edits - SDZ = engine.execute(Zoo.select(Zoo.c.Name == u'San Diego Zoo' + SDZ = engine.execute(Zoo.select(Zoo.c.Name == 'San Diego Zoo' )).first() assert SDZ['Founded'] == datetime.date(1935, 9, 13) @@ -333,14 +335,14 @@ class ZooMarkTest(fixtures.TestBase): return [list(row) for row in engine.execute(select).fetchall()] - for x in xrange(ITERATIONS): + for x in range(ITERATIONS): za = fulltable(select([Zoo.c.ID] + list(Animal.c), - Zoo.c.Name == u'San Diego Zoo', + Zoo.c.Name == 'San Diego Zoo', from_obj=[join(Zoo, Animal)])) - SDZ = Zoo.select(Zoo.c.Name == u'San Diego Zoo') + SDZ = Zoo.select(Zoo.c.Name == 'San Diego Zoo') e = fulltable(select([Zoo.c.ID, Animal.c.ID], - and_(Zoo.c.Name == u'San Diego Zoo', - Animal.c.Species == u'Leopard'), + and_(Zoo.c.Name == 'San Diego Zoo', + Animal.c.Species == 'Leopard'), from_obj=[join(Zoo, Animal)])) # Now try the same query with INNER, LEFT, and RIGHT JOINs. diff --git a/test/aaa_profiling/test_zoomark_orm.py b/test/aaa_profiling/test_zoomark_orm.py index f8c5477f7..ddcad681a 100644 --- a/test/aaa_profiling/test_zoomark_orm.py +++ b/test/aaa_profiling/test_zoomark_orm.py @@ -1,3 +1,5 @@ +from __future__ import unicode_literals + """Benchmark for SQLAlchemy. An adaptation of Robert Brewers' ZooMark speed tests. """ @@ -75,14 +77,14 @@ class ZooMarkTest(fixtures.TestBase): class Zoo(object): def __init__(self, **kwargs): - for k, v in kwargs.iteritems(): + for k, v in kwargs.items(): setattr(self, k, v) class Animal(object): def __init__(self, **kwargs): - for k, v in kwargs.iteritems(): + for k, v in kwargs.items(): setattr(self, k, v) @@ -90,93 +92,93 @@ class ZooMarkTest(fixtures.TestBase): mapper(Animal, animal) def test_baseline_1a_populate(self): - wap = Zoo(Name=u'Wild Animal Park', Founded=datetime.date(2000, + wap = Zoo(Name='Wild Animal Park', Founded=datetime.date(2000, 1, 1), Opens=datetime.time(8, 15, 59), LastEscape=datetime.datetime( 2004, 7, 29, 5, 6, 7, ), Admission=4.95) session.add(wap) - sdz = Zoo(Name=u'San Diego Zoo', Founded=datetime.date(1835, 9, + sdz = Zoo(Name='San Diego Zoo', Founded=datetime.date(1835, 9, 13), Opens=datetime.time(9, 0, 0), Admission=0) session.add(sdz) - bio = Zoo(Name=u'Montr\xe9al Biod\xf4me', + bio = Zoo(Name='Montr\xe9al Biod\xf4me', Founded=datetime.date(1992, 6, 19), Opens=datetime.time(9, 0, 0), Admission=11.75) session.add(bio) - seaworld = Zoo(Name=u'Sea_World', Admission=60) + seaworld = Zoo(Name='Sea_World', Admission=60) session.add(seaworld) # Let's add a crazy futuristic Zoo to test large date values. - lp = Zoo(Name=u'Luna Park', Founded=datetime.date(2072, 7, 17), + lp = Zoo(Name='Luna Park', Founded=datetime.date(2072, 7, 17), Opens=datetime.time(0, 0, 0), Admission=134.95) session.add(lp) session.flush() # Animals - leopard = Animal(Species=u'Leopard', Lifespan=73.5) + leopard = Animal(Species='Leopard', Lifespan=73.5) session.add(leopard) leopard.ZooID = wap.ID leopard.LastEscape = \ datetime.datetime(2004, 12, 21, 8, 15, 0, 999907, ) - session.add(Animal(Species=u'Lion', ZooID=wap.ID)) - session.add(Animal(Species=u'Slug', Legs=1, Lifespan=.75)) - session.add(Animal(Species=u'Tiger', ZooID=sdz.ID)) + session.add(Animal(Species='Lion', ZooID=wap.ID)) + session.add(Animal(Species='Slug', Legs=1, Lifespan=.75)) + session.add(Animal(Species='Tiger', ZooID=sdz.ID)) # Override Legs.default with itself just to make sure it works. - session.add(Animal(Species=u'Bear', Legs=4)) - session.add(Animal(Species=u'Ostrich', Legs=2, Lifespan=103.2)) - session.add(Animal(Species=u'Centipede', Legs=100)) - session.add(Animal(Species=u'Emperor Penguin', Legs=2, + session.add(Animal(Species='Bear', Legs=4)) + session.add(Animal(Species='Ostrich', Legs=2, Lifespan=103.2)) + session.add(Animal(Species='Centipede', Legs=100)) + session.add(Animal(Species='Emperor Penguin', Legs=2, ZooID=seaworld.ID)) - session.add(Animal(Species=u'Adelie Penguin', Legs=2, + session.add(Animal(Species='Adelie Penguin', Legs=2, ZooID=seaworld.ID)) - session.add(Animal(Species=u'Millipede', Legs=1000000, + session.add(Animal(Species='Millipede', Legs=1000000, ZooID=sdz.ID)) # Add a mother and child to test relationships - bai_yun = Animal(Species=u'Ape', Nameu=u'Bai Yun', Legs=2) + bai_yun = Animal(Species='Ape', Nameu='Bai Yun', Legs=2) session.add(bai_yun) - session.add(Animal(Species=u'Ape', Name=u'Hua Mei', Legs=2, + session.add(Animal(Species='Ape', Name='Hua Mei', Legs=2, MotherID=bai_yun.ID)) session.flush() session.commit() def test_baseline_2_insert(self): - for x in xrange(ITERATIONS): - session.add(Animal(Species=u'Tick', Name=u'Tick %d' % x, + for x in range(ITERATIONS): + session.add(Animal(Species='Tick', Name='Tick %d' % x, Legs=8)) session.flush() def test_baseline_3_properties(self): - for x in xrange(ITERATIONS): + for x in range(ITERATIONS): # Zoos WAP = list(session.query(Zoo).filter(Zoo.Name - == u'Wild Animal Park')) + == 'Wild Animal Park')) SDZ = list(session.query(Zoo).filter(Zoo.Founded == datetime.date(1835, 9, 13))) Biodome = list(session.query(Zoo).filter(Zoo.Name - == u'Montr\xe9al Biod\xf4me')) + == 'Montr\xe9al Biod\xf4me')) seaworld = list(session.query(Zoo).filter(Zoo.Admission == float(60))) # Animals leopard = list(session.query(Animal).filter(Animal.Species - == u'Leopard')) + == 'Leopard')) ostrich = list(session.query(Animal).filter(Animal.Species - == u'Ostrich')) + == 'Ostrich')) millipede = list(session.query(Animal).filter(Animal.Legs == 1000000)) ticks = list(session.query(Animal).filter(Animal.Species - == u'Tick')) + == 'Tick')) def test_baseline_4_expressions(self): - for x in xrange(ITERATIONS): + for x in range(ITERATIONS): assert len(list(session.query(Zoo))) == 5 assert len(list(session.query(Animal))) == ITERATIONS + 12 assert len(list(session.query(Animal).filter(Animal.Legs @@ -190,9 +192,9 @@ class ZooMarkTest(fixtures.TestBase): assert len(list(session.query(Animal).filter(Animal.Lifespan > 70))) == 2 assert len(list(session.query(Animal). - filter(Animal.Species.like(u'L%')))) == 2 + filter(Animal.Species.like('L%')))) == 2 assert len(list(session.query(Animal). - filter(Animal.Species.like(u'%pede')))) == 2 + filter(Animal.Species.like('%pede')))) == 2 assert len(list(session.query(Animal).filter(Animal.LastEscape != None))) == 1 assert len(list(session.query(Animal).filter(Animal.LastEscape @@ -201,29 +203,29 @@ class ZooMarkTest(fixtures.TestBase): # In operator (containedby) assert len(list(session.query(Animal).filter( - Animal.Species.like(u'%pede%')))) == 2 + Animal.Species.like('%pede%')))) == 2 assert len(list(session.query(Animal). - filter(Animal.Species.in_((u'Lion' - , u'Tiger', u'Bear'))))) == 3 + filter(Animal.Species.in_(('Lion' + , 'Tiger', 'Bear'))))) == 3 # Try In with cell references class thing(object): pass pet, pet2 = thing(), thing() - pet.Name, pet2.Name = u'Slug', u'Ostrich' + pet.Name, pet2.Name = 'Slug', 'Ostrich' assert len(list(session.query(Animal). filter(Animal.Species.in_((pet.Name, pet2.Name))))) == 2 # logic and other functions - name = u'Lion' + name = 'Lion' assert len(list(session.query(Animal). filter(func.length(Animal.Species) == len(name)))) == ITERATIONS + 3 assert len(list(session.query(Animal). - filter(Animal.Species.like(u'%i%' + filter(Animal.Species.like('%i%' )))) == ITERATIONS + 7 # Test now(), today(), year(), month(), day() @@ -246,7 +248,7 @@ class ZooMarkTest(fixtures.TestBase): # TODO: convert to ORM engine = metadata.bind - for x in xrange(ITERATIONS): + for x in range(ITERATIONS): # views @@ -270,7 +272,7 @@ class ZooMarkTest(fixtures.TestBase): for species, lifespan in engine.execute(select([Animal.c.Species, Animal.c.Lifespan])).fetchall(): assert lifespan == expected[species] - expected = [u'Montr\xe9al Biod\xf4me', 'Wild Animal Park'] + expected = ['Montr\xe9al Biod\xf4me', 'Wild Animal Park'] e = select([Zoo.c.Name], and_(Zoo.c.Founded != None, Zoo.c.Founded <= func.current_timestamp(), Zoo.c.Founded >= datetime.date(1990, 1, 1))) @@ -284,13 +286,13 @@ class ZooMarkTest(fixtures.TestBase): legs.sort() def test_baseline_6_editing(self): - for x in xrange(ITERATIONS): + for x in range(ITERATIONS): # Edit - SDZ = session.query(Zoo).filter(Zoo.Name == u'San Diego Zoo' + SDZ = session.query(Zoo).filter(Zoo.Name == 'San Diego Zoo' ).one() - SDZ.Name = u'The San Diego Zoo' + SDZ.Name = 'The San Diego Zoo' SDZ.Founded = datetime.date(1900, 1, 1) SDZ.Opens = datetime.time(7, 30, 0) SDZ.Admission = 35.00 @@ -298,19 +300,19 @@ class ZooMarkTest(fixtures.TestBase): # Test edits SDZ = session.query(Zoo).filter(Zoo.Name - == u'The San Diego Zoo').one() + == 'The San Diego Zoo').one() assert SDZ.Founded == datetime.date(1900, 1, 1), SDZ.Founded # Change it back - SDZ.Name = u'San Diego Zoo' + SDZ.Name = 'San Diego Zoo' SDZ.Founded = datetime.date(1835, 9, 13) SDZ.Opens = datetime.time(9, 0, 0) SDZ.Admission = 0 # Test re-edits - SDZ = session.query(Zoo).filter(Zoo.Name == u'San Diego Zoo' + SDZ = session.query(Zoo).filter(Zoo.Name == 'San Diego Zoo' ).one() assert SDZ.Founded == datetime.date(1835, 9, 13), \ SDZ.Founded diff --git a/test/base/test_dependency.py b/test/base/test_dependency.py index f4a0a4c8b..b16516f15 100644 --- a/test/base/test_dependency.py +++ b/test/base/test_dependency.py @@ -82,7 +82,7 @@ class DependencySortTest(fixtures.TestBase): try: list(topological.sort(tuples, allitems)) assert False - except exc.CircularDependencyError, err: + except exc.CircularDependencyError as err: eq_(err.cycles, set(['node1', 'node3', 'node2', 'node5', 'node4'])) eq_(err.edges, set([('node3', 'node1'), ('node4', 'node1'), @@ -105,7 +105,7 @@ class DependencySortTest(fixtures.TestBase): try: list(topological.sort(tuples, allitems)) assert False - except exc.CircularDependencyError, err: + except exc.CircularDependencyError as err: eq_(err.cycles, set(['node1', 'node3', 'node2'])) eq_(err.edges, set([('node3', 'node1'), ('node2', 'node3'), ('node3', 'node2'), ('node1', 'node2'), @@ -271,7 +271,7 @@ class DependencySortTest(fixtures.TestBase): ('node4', 'node17'), ('node2', 'node20'), ('node19', 'node10'), ('node8', 'node4'), ('node11', 'node3'), ('node6', 'node1') ] - allnodes = ['node%d' % i for i in xrange(1, 21)] + allnodes = ['node%d' % i for i in range(1, 21)] eq_( topological.find_cycles(tuples, allnodes), set(['node11', 'node10', 'node13', 'node15', 'node14', 'node17', diff --git a/test/base/test_events.py b/test/base/test_events.py index 4efb30aba..7cfb5fa7d 100644 --- a/test/base/test_events.py +++ b/test/base/test_events.py @@ -190,7 +190,7 @@ class ClsLevelListenTest(fixtures.TestBase): def test_lis_subcalss_lis(self): @event.listens_for(self.TargetOne, "event_one") def handler1(x, y): - print 'handler1' + print('handler1') class SubTarget(self.TargetOne): pass @@ -207,7 +207,7 @@ class ClsLevelListenTest(fixtures.TestBase): def test_lis_multisub_lis(self): @event.listens_for(self.TargetOne, "event_one") def handler1(x, y): - print 'handler1' + print('handler1') class SubTarget(self.TargetOne): pass diff --git a/test/base/test_except.py b/test/base/test_except.py index a8c7de201..f2428c22d 100644 --- a/test/base/test_except.py +++ b/test/base/test_except.py @@ -5,14 +5,15 @@ from sqlalchemy import exc as sa_exceptions from sqlalchemy.testing import fixtures from sqlalchemy.testing import eq_ -# Py3K -#StandardError = BaseException -# Py2K -from exceptions import StandardError, KeyboardInterrupt, SystemExit +# start Py3K +Exception = BaseException +# end Py3K +# start Py2K +#from exceptions import StandardError, KeyboardInterrupt, SystemExit # end Py2K -class Error(StandardError): +class Error(Exception): """This class will be old-style on <= 2.4 and new-style on >= 2.5.""" @@ -47,7 +48,7 @@ class WrapTest(fixtures.TestBase): try: raise sa_exceptions.DBAPIError.instance('this is a message' , None, OperationalError(), DatabaseError) - except sa_exceptions.DBAPIError, exc: + except sa_exceptions.DBAPIError as exc: assert str(exc) \ == "(OperationalError) 'this is a message' None" @@ -58,7 +59,7 @@ class WrapTest(fixtures.TestBase): {'a': 1, 'b': 2, 'c': 3, 'd': 4, 'e': 5, 'f': 6, 'g': 7, 'h': 8, 'i': 9, 'j': 10, 'k': 11, }, OperationalError(), DatabaseError) - except sa_exceptions.DBAPIError, exc: + except sa_exceptions.DBAPIError as exc: assert str(exc).startswith("(OperationalError) 'this is a " "message' {") @@ -67,7 +68,7 @@ class WrapTest(fixtures.TestBase): raise sa_exceptions.DBAPIError.instance('this is a message', [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11,], OperationalError(), DatabaseError) - except sa_exceptions.DBAPIError, exc: + except sa_exceptions.DBAPIError as exc: assert str(exc).startswith("(OperationalError) 'this is a " "message' [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11]") @@ -77,7 +78,7 @@ class WrapTest(fixtures.TestBase): [{1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: 1}, {1:1}, {1: 1}, {1: 1},], OperationalError(), DatabaseError) - except sa_exceptions.DBAPIError, exc: + except sa_exceptions.DBAPIError as exc: eq_(str(exc) , "(OperationalError) 'this is a message' [{1: 1}, "\ "{1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: "\ @@ -87,7 +88,7 @@ class WrapTest(fixtures.TestBase): {1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: 1}, {1:1}, {1: 1}, {1: 1}, {1: 1}, ], OperationalError(), DatabaseError) - except sa_exceptions.DBAPIError, exc: + except sa_exceptions.DBAPIError as exc: eq_(str(exc) , "(OperationalError) 'this is a message' [{1: 1}, " "{1: 1}, {1: 1}, {1: 1}, {1: 1}, {1: 1}, " @@ -100,7 +101,7 @@ class WrapTest(fixtures.TestBase): (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), ], OperationalError(), DatabaseError) - except sa_exceptions.DBAPIError, exc: + except sa_exceptions.DBAPIError as exc: eq_(str(exc), "(OperationalError) 'this is a message' [(1,), "\ "(1,), (1,), (1,), (1,), (1,), (1,), (1,), (1,), (1,)]") @@ -109,7 +110,7 @@ class WrapTest(fixtures.TestBase): (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), (1, ), ], OperationalError(), DatabaseError) - except sa_exceptions.DBAPIError, exc: + except sa_exceptions.DBAPIError as exc: eq_(str(exc), "(OperationalError) 'this is a message' [(1,), " "(1,), (1,), (1,), (1,), (1,), (1,), (1,) " @@ -121,7 +122,7 @@ class WrapTest(fixtures.TestBase): try: raise sa_exceptions.DBAPIError.instance('', [], ProgrammingError(), DatabaseError) - except sa_exceptions.DBAPIError, e: + except sa_exceptions.DBAPIError as e: self.assert_(True) self.assert_('Error in str() of DB-API' in e.args[0]) @@ -129,7 +130,7 @@ class WrapTest(fixtures.TestBase): try: raise sa_exceptions.DBAPIError.instance('', [], OutOfSpec(), DatabaseError) - except sa_exceptions.DBAPIError, e: + except sa_exceptions.DBAPIError as e: self.assert_(e.__class__ is sa_exceptions.DBAPIError) except OutOfSpec: self.assert_(False) @@ -137,7 +138,7 @@ class WrapTest(fixtures.TestBase): try: raise sa_exceptions.DBAPIError.instance('', [], sa_exceptions.ArgumentError(), DatabaseError) - except sa_exceptions.DBAPIError, e: + except sa_exceptions.DBAPIError as e: self.assert_(e.__class__ is sa_exceptions.DBAPIError) except sa_exceptions.ArgumentError: self.assert_(False) diff --git a/test/base/test_utils.py b/test/base/test_utils.py index b28d26e71..194ef1222 100644 --- a/test/base/test_utils.py +++ b/test/base/test_utils.py @@ -16,7 +16,7 @@ class KeyedTupleTest(): eq_(len(keyed_tuple), 0) eq_(keyed_tuple.__dict__, {'_labels': []}) - eq_(keyed_tuple.keys(), []) + eq_(list(keyed_tuple.keys()), []) eq_(keyed_tuple._fields, ()) eq_(keyed_tuple._asdict(), {}) @@ -27,7 +27,7 @@ class KeyedTupleTest(): eq_(len(keyed_tuple), 2) eq_(keyed_tuple.__dict__, {'_labels': []}) - eq_(keyed_tuple.keys(), []) + eq_(list(keyed_tuple.keys()), []) eq_(keyed_tuple._fields, ()) eq_(keyed_tuple._asdict(), {}) @@ -37,7 +37,7 @@ class KeyedTupleTest(): def test_basic_creation(self): keyed_tuple = util.KeyedTuple([1, 2], ['a', 'b']) eq_(str(keyed_tuple), '(1, 2)') - eq_(keyed_tuple.keys(), ['a', 'b']) + eq_(list(keyed_tuple.keys()), ['a', 'b']) eq_(keyed_tuple._fields, ('a', 'b')) eq_(keyed_tuple._asdict(), {'a': 1, 'b': 2}) @@ -66,7 +66,7 @@ class KeyedTupleTest(): # TODO: consider not allowing None labels expected = {'a': 1, None: 2, 'b': 3, '_labels': ['a', None, 'b']} eq_(keyed_tuple.__dict__, expected) - eq_(keyed_tuple.keys(), ['a', 'b']) + eq_(list(keyed_tuple.keys()), ['a', 'b']) eq_(keyed_tuple._fields, ('a', 'b')) eq_(keyed_tuple._asdict(), {'a': 1, 'b': 3}) @@ -86,7 +86,7 @@ class KeyedTupleTest(): # TODO: consider not allowing duplicate labels expected = {'a': 1, 'b': 3, '_labels': ['a', 'b', 'b']} eq_(keyed_tuple.__dict__, expected) - eq_(keyed_tuple.keys(), ['a', 'b', 'b']) + eq_(list(keyed_tuple.keys()), ['a', 'b', 'b']) eq_(keyed_tuple._fields, ('a', 'b', 'b')) eq_(keyed_tuple._asdict(), {'a': 1, 'b': 3}) @@ -125,12 +125,12 @@ class OrderedDictTest(fixtures.TestBase): o['snack'] = 'attack' o['c'] = 3 - eq_(o.keys(), ['a', 'b', 'snack', 'c']) - eq_(o.values(), [1, 2, 'attack', 3]) + eq_(list(o.keys()), ['a', 'b', 'snack', 'c']) + eq_(list(o.values()), [1, 2, 'attack', 3]) o.pop('snack') - eq_(o.keys(), ['a', 'b', 'c']) - eq_(o.values(), [1, 2, 3]) + eq_(list(o.keys()), ['a', 'b', 'c']) + eq_(list(o.values()), [1, 2, 3]) try: o.pop('eep') @@ -146,40 +146,40 @@ class OrderedDictTest(fixtures.TestBase): except TypeError: pass - eq_(o.keys(), ['a', 'b', 'c']) - eq_(o.values(), [1, 2, 3]) + eq_(list(o.keys()), ['a', 'b', 'c']) + eq_(list(o.values()), [1, 2, 3]) o2 = util.OrderedDict(d=4) o2['e'] = 5 - eq_(o2.keys(), ['d', 'e']) - eq_(o2.values(), [4, 5]) + eq_(list(o2.keys()), ['d', 'e']) + eq_(list(o2.values()), [4, 5]) o.update(o2) - eq_(o.keys(), ['a', 'b', 'c', 'd', 'e']) - eq_(o.values(), [1, 2, 3, 4, 5]) + eq_(list(o.keys()), ['a', 'b', 'c', 'd', 'e']) + eq_(list(o.values()), [1, 2, 3, 4, 5]) o.setdefault('c', 'zzz') o.setdefault('f', 6) - eq_(o.keys(), ['a', 'b', 'c', 'd', 'e', 'f']) - eq_(o.values(), [1, 2, 3, 4, 5, 6]) + eq_(list(o.keys()), ['a', 'b', 'c', 'd', 'e', 'f']) + eq_(list(o.values()), [1, 2, 3, 4, 5, 6]) def test_odict_constructor(self): o = util.OrderedDict([('name', 'jbe'), ('fullname', 'jonathan' ), ('password', '')]) - eq_(o.keys(), ['name', 'fullname', 'password']) + eq_(list(o.keys()), ['name', 'fullname', 'password']) def test_odict_copy(self): o = util.OrderedDict() o["zzz"] = 1 o["aaa"] = 2 - eq_(o.keys(), ['zzz', 'aaa']) + eq_(list(o.keys()), ['zzz', 'aaa']) o2 = o.copy() - eq_(o2.keys(), o.keys()) + eq_(list(o2.keys()), list(o.keys())) o3 = copy.copy(o) - eq_(o3.keys(), o.keys()) + eq_(list(o3.keys()), list(o.keys())) class OrderedSetTest(fixtures.TestBase): @@ -198,7 +198,7 @@ class FrozenDictTest(fixtures.TestBase): def test_serialize(self): d = util.immutabledict({1: 2, 3: 4}) for loads, dumps in picklers(): - print loads(dumps(d)) + print(loads(dumps(d))) class MemoizedAttrTest(fixtures.TestBase): @@ -252,7 +252,7 @@ class ColumnCollectionTest(fixtures.TestBase): try: cc['col1'] in cc assert False - except exc.ArgumentError, e: + except exc.ArgumentError as e: eq_(str(e), "__contains__ requires a string argument") def test_compare(self): @@ -414,14 +414,14 @@ class IdentitySetTest(fixtures.TestBase): for type_ in (object, ImmutableSubclass): data = [type_(), type_()] ids = util.IdentitySet() - for i in range(2) + range(2): + for i in list(range(2)) + list(range(2)): ids.add(data[i]) self.assert_eq(ids, data) for type_ in (EqOverride, HashOverride, HashEqOverride): data = [type_(1), type_(1), type_(2)] ids = util.IdentitySet() - for i in range(3) + range(3): + for i in list(range(3)) + list(range(3)): ids.add(data[i]) self.assert_eq(ids, data) @@ -843,13 +843,12 @@ class IdentitySetTest(fixtures.TestBase): return super_, sub_, twin1, twin2, unique1, unique2 def _assert_unorderable_types(self, callable_): - # Py3K - #assert_raises_message( - # TypeError, 'unorderable types', callable_) - # Py2K - assert_raises_message( - TypeError, 'cannot compare sets using cmp()', callable_) - # end Py2K + if util.py3k: + assert_raises_message( + TypeError, 'unorderable types', callable_) + else: + assert_raises_message( + TypeError, 'cannot compare sets using cmp()', callable_) def test_basic_sanity(self): IdentitySet = util.IdentitySet @@ -969,23 +968,23 @@ class DictlikeIteritemsTest(fixtures.TestBase): d = subdict(a=1, b=2, c=3) self._ok(d) - # Py2K - def test_UserDict(self): - import UserDict - d = UserDict.UserDict(a=1, b=2, c=3) - self._ok(d) - # end Py2K +# start Py2K +# def test_UserDict(self): +# import UserDict +# d = UserDict.UserDict(a=1, b=2, c=3) +# self._ok(d) +# end Py2K def test_object(self): self._notok(object()) - # Py2K - def test_duck_1(self): - class duck1(object): - def iteritems(duck): - return iter(self.baseline) - self._ok(duck1()) - # end Py2K +# start Py2K +# def test_duck_1(self): +# class duck1(object): +# def iteritems(duck): +# return iter(self.baseline) +# self._ok(duck1()) +# end Py2K def test_duck_2(self): class duck2(object): @@ -993,16 +992,16 @@ class DictlikeIteritemsTest(fixtures.TestBase): return list(self.baseline) self._ok(duck2()) - # Py2K - def test_duck_3(self): - class duck3(object): - def iterkeys(duck): - return iter(['a', 'b', 'c']) - - def __getitem__(duck, key): - return dict(a=1, b=2, c=3).get(key) - self._ok(duck3()) - # end Py2K +# start Py2K +# def test_duck_3(self): +# class duck3(object): +# def iterkeys(duck): +# return iter(['a', 'b', 'c']) +# +# def __getitem__(duck, key): +# return dict(a=1, b=2, c=3).get(key) +# self._ok(duck3()) +# end Py2K def test_duck_4(self): class duck4(object): @@ -1029,9 +1028,9 @@ class DictlikeIteritemsTest(fixtures.TestBase): class DuckTypeCollectionTest(fixtures.TestBase): def test_sets(self): - # Py2K - import sets - # end Py2K +# start Py2K +# import sets +# end Py2K class SetLike(object): def add(self): @@ -1041,9 +1040,9 @@ class DuckTypeCollectionTest(fixtures.TestBase): __emulates__ = set for type_ in (set, - # Py2K - sets.Set, - # end Py2K +# start Py2K +# sets.Set, +# end Py2K SetLike, ForcedSet): eq_(util.duck_type_collection(type_), set) @@ -1051,9 +1050,9 @@ class DuckTypeCollectionTest(fixtures.TestBase): eq_(util.duck_type_collection(instance), set) for type_ in (frozenset, - # Py2K - sets.ImmutableSet - # end Py2K +# start Py2K +# sets.ImmutableSet +# end Py2K ): is_(util.duck_type_collection(type_), None) instance = type_() @@ -1195,7 +1194,7 @@ class SymbolTest(fixtures.TestBase): sym3 = util.pickle.loads(s) for protocol in 0, 1, 2: - print protocol + print(protocol) serial = util.pickle.dumps(sym1) rt = util.pickle.loads(serial) assert rt is sym1 @@ -1570,21 +1569,21 @@ class TestClassHierarchy(fixtures.TestBase): eq_(set(util.class_hierarchy(A)), set((A, B, C, object))) eq_(set(util.class_hierarchy(B)), set((A, B, C, object))) - # Py2K - def test_oldstyle_mixin(self): - class A(object): - pass - - class Mixin: - pass - - class B(A, Mixin): - pass - - eq_(set(util.class_hierarchy(B)), set((A, B, object))) - eq_(set(util.class_hierarchy(Mixin)), set()) - eq_(set(util.class_hierarchy(A)), set((A, B, object))) - # end Py2K +# start Py2K +# def test_oldstyle_mixin(self): +# class A(object): +# pass +# +# class Mixin: +# pass +# +# class B(A, Mixin): +# pass +# +# eq_(set(util.class_hierarchy(B)), set((A, B, object))) +# eq_(set(util.class_hierarchy(Mixin)), set()) +# eq_(set(util.class_hierarchy(A)), set((A, B, object))) +# end Py2K class TestClassProperty(fixtures.TestBase): diff --git a/test/dialect/test_firebird.py b/test/dialect/test_firebird.py index 5a80a3776..6019dc8f9 100644 --- a/test/dialect/test_firebird.py +++ b/test/dialect/test_firebird.py @@ -28,7 +28,7 @@ class DomainReflectionTest(fixtures.TestBase, AssertsExecutionResults): ) con.execute('CREATE DOMAIN img_domain AS BLOB SUB_TYPE ' 'BINARY') - except ProgrammingError, e: + except ProgrammingError as e: if not 'attempt to store duplicate value' in str(e): raise e con.execute('''CREATE GENERATOR gen_testtable_id''') diff --git a/test/dialect/test_mssql.py b/test/dialect/test_mssql.py index 7b8f53390..4a99ef5f7 100644 --- a/test/dialect/test_mssql.py +++ b/test/dialect/test_mssql.py @@ -16,8 +16,8 @@ from sqlalchemy import testing from sqlalchemy.testing import emits_warning_on, assert_raises_message import decimal from sqlalchemy.engine.reflection import Inspector -from sqlalchemy.util.compat import b -from sqlalchemy import sql +from sqlalchemy.util import b, u, ue +from sqlalchemy import sql, util class CompileTest(fixtures.TestBase, AssertsCompiledSQL): @@ -414,7 +414,7 @@ class CompileTest(fixtures.TestBase, AssertsCompiledSQL): self.assert_compile( s, "SELECT TOP 10 t.x, t.y FROM t WHERE t.x = :x_1 ORDER BY t.y", - checkparams={u'x_1': 5} + checkparams={'x_1': 5} ) def test_limit_zero_using_top(self): @@ -425,7 +425,7 @@ class CompileTest(fixtures.TestBase, AssertsCompiledSQL): self.assert_compile( s, "SELECT TOP 0 t.x, t.y FROM t WHERE t.x = :x_1 ORDER BY t.y", - checkparams={u'x_1': 5} + checkparams={'x_1': 5} ) def test_offset_using_window(self): @@ -435,14 +435,14 @@ class CompileTest(fixtures.TestBase, AssertsCompiledSQL): # test that the select is not altered with subsequent compile # calls - for i in xrange(2): + for i in range(2): self.assert_compile( s, "SELECT anon_1.x, anon_1.y FROM (SELECT t.x AS x, t.y " "AS y, ROW_NUMBER() OVER (ORDER BY t.y) AS " "mssql_rn FROM t WHERE t.x = :x_1) AS " "anon_1 WHERE mssql_rn > :mssql_rn_1", - checkparams={u'mssql_rn_1': 20, u'x_1': 5} + checkparams={'mssql_rn_1': 20, 'x_1': 5} ) def test_limit_offset_using_window(self): @@ -458,7 +458,7 @@ class CompileTest(fixtures.TestBase, AssertsCompiledSQL): "FROM t " "WHERE t.x = :x_1) AS anon_1 " "WHERE mssql_rn > :mssql_rn_1 AND mssql_rn <= :mssql_rn_2", - checkparams={u'mssql_rn_1': 20, u'mssql_rn_2': 30, u'x_1': 5} + checkparams={'mssql_rn_1': 20, 'mssql_rn_2': 30, 'x_1': 5} ) def test_limit_offset_with_correlated_order_by(self): @@ -479,7 +479,7 @@ class CompileTest(fixtures.TestBase, AssertsCompiledSQL): "FROM t1 " "WHERE t1.x = :x_1) AS anon_1 " "WHERE mssql_rn > :mssql_rn_1 AND mssql_rn <= :mssql_rn_2", - checkparams={u'mssql_rn_1': 20, u'mssql_rn_2': 30, u'x_1': 5} + checkparams={'mssql_rn_1': 20, 'mssql_rn_2': 30, 'x_1': 5} ) def test_limit_zero_offset_using_window(self): @@ -493,7 +493,7 @@ class CompileTest(fixtures.TestBase, AssertsCompiledSQL): s, "SELECT TOP 0 t.x, t.y FROM t " "WHERE t.x = :x_1 ORDER BY t.y", - checkparams={u'x_1': 5} + checkparams={'x_1': 5} ) def test_sequence_start_0(self): @@ -862,11 +862,11 @@ class QueryUnicodeTest(fixtures.TestBase): # encode in UTF-8 (sting object) because this is the default # dialect encoding - con.execute(u"insert into unitest_table values ('bien u\ - umang\xc3\xa9')".encode('UTF-8')) + con.execute(ue("insert into unitest_table values ('bien u\ + umang\xc3\xa9')").encode('UTF-8')) try: r = t1.select().execute().first() - assert isinstance(r[1], unicode), \ + assert isinstance(r[1], util.text_type), \ '%s is %s instead of unicode, working on %s' % (r[1], type(r[1]), meta.bind) finally: @@ -1718,7 +1718,7 @@ class TypeRoundTripTest(fixtures.TestBase, AssertsExecutionResults, ComparesTabl )] for value in test_items: float_table.insert().execute(floatcol=value) - except Exception, e: + except Exception as e: raise e @@ -1964,8 +1964,8 @@ class MonkeyPatchedBinaryTest(fixtures.TestBase): def test_unicode(self): module = __import__('pymssql') - result = module.Binary(u'foo') - eq_(result, u'foo') + result = module.Binary('foo') + eq_(result, 'foo') def test_bytes(self): module = __import__('pymssql') @@ -2084,7 +2084,7 @@ class InfoCoerceUnicodeTest(fixtures.TestBase): dialect = mssql.dialect() value = CoerceUnicode().bind_processor(dialect)('a string') - assert isinstance(value, unicode) + assert isinstance(value, util.text_type) class ReflectHugeViewTest(fixtures.TestBase): __only_on__ = 'mssql' @@ -2096,13 +2096,13 @@ class ReflectHugeViewTest(fixtures.TestBase): t = Table('base_table', self.metadata, *[ Column("long_named_column_number_%d" % i, Integer) - for i in xrange(self.col_num) + for i in range(self.col_num) ] ) self.view_str = view_str = \ "CREATE VIEW huge_named_view AS SELECT %s FROM base_table" % ( ",".join("long_named_column_number_%d" % i - for i in xrange(self.col_num)) + for i in range(self.col_num)) ) assert len(view_str) > 4000 diff --git a/test/dialect/test_mysql.py b/test/dialect/test_mysql.py index f89b0b229..728098d3a 100644 --- a/test/dialect/test_mysql.py +++ b/test/dialect/test_mysql.py @@ -5,6 +5,7 @@ from sqlalchemy.testing import eq_, assert_raises, assert_raises_message from sqlalchemy import * from sqlalchemy import sql, exc, schema, types as sqltypes +from sqlalchemy.util import u from sqlalchemy.dialects.mysql import base as mysql from sqlalchemy.engine.url import make_url from sqlalchemy.testing import fixtures, AssertsCompiledSQL, AssertsExecutionResults @@ -372,9 +373,9 @@ class TypesTest(fixtures.TestBase, AssertsExecutionResults, AssertsCompiledSQL): try: self.assert_(list(row) == expected) except: - print "Storing %s" % store - print "Expected %s" % expected - print "Found %s" % list(row) + print("Storing %s" % store) + print("Expected %s" % expected) + print("Found %s" % list(row)) raise table.delete().execute().close() @@ -684,17 +685,17 @@ class EnumTest(fixtures.TestBase, AssertsExecutionResults, AssertsCompiledSQL): metadata = MetaData(unicode_engine) t1 = Table('table', metadata, Column('id', Integer, primary_key=True), - Column('value', Enum(u'réveillé', u'drôle', u'S’il')), - Column('value2', mysql.ENUM(u'réveillé', u'drôle', u'S’il')) + Column('value', Enum(u('réveillé'), u('drôle'), u('S’il'))), + Column('value2', mysql.ENUM(u('réveillé'), u('drôle'), u('S’il'))) ) metadata.create_all() try: - t1.insert().execute(value=u'drôle', value2=u'drôle') - t1.insert().execute(value=u'réveillé', value2=u'réveillé') - t1.insert().execute(value=u'S’il', value2=u'S’il') + t1.insert().execute(value=u('drôle'), value2=u('drôle')) + t1.insert().execute(value=u('réveillé'), value2=u('réveillé')) + t1.insert().execute(value=u('S’il'), value2=u('S’il')) eq_(t1.select().order_by(t1.c.id).execute().fetchall(), - [(1, u'drôle', u'drôle'), (2, u'réveillé', u'réveillé'), - (3, u'S’il', u'S’il')] + [(1, u('drôle'), u('drôle')), (2, u('réveillé'), u('réveillé')), + (3, u('S’il'), u('S’il'))] ) # test reflection of the enum labels @@ -706,10 +707,10 @@ class EnumTest(fixtures.TestBase, AssertsExecutionResults, AssertsCompiledSQL): # latin-1 stuff forcing its way in ? assert t2.c.value.type.enums[0:2] == \ - (u'réveillé', u'drôle') # u'S’il') # eh ? + (u('réveillé'), u('drôle')) # u'S’il') # eh ? assert t2.c.value2.type.enums[0:2] == \ - (u'réveillé', u'drôle') # u'S’il') # eh ? + (u('réveillé'), u('drôle')) # u'S’il') # eh ? finally: metadata.drop_all() @@ -1370,7 +1371,7 @@ class SQLModeDetectionTest(fixtures.TestBase): def _options(self, modes): def connect(con, record): cursor = con.cursor() - print "DOING THiS:", "set sql_mode='%s'" % (",".join(modes)) + print("DOING THiS:", "set sql_mode='%s'" % (",".join(modes))) cursor.execute("set sql_mode='%s'" % (",".join(modes))) e = engines.testing_engine(options={ 'pool_events':[ diff --git a/test/dialect/test_oracle.py b/test/dialect/test_oracle.py index def4654f0..d26ca6c03 100644 --- a/test/dialect/test_oracle.py +++ b/test/dialect/test_oracle.py @@ -1,5 +1,5 @@ # coding: utf-8 -from __future__ import with_statement + from sqlalchemy.testing import eq_ from sqlalchemy import * @@ -7,6 +7,8 @@ from sqlalchemy import types as sqltypes, exc, schema from sqlalchemy.sql import table, column from sqlalchemy.testing import fixtures, AssertsExecutionResults, AssertsCompiledSQL from sqlalchemy import testing +from sqlalchemy.util import u, b +from sqlalchemy import util from sqlalchemy.testing import assert_raises, assert_raises_message from sqlalchemy.testing.engines import testing_engine from sqlalchemy.dialects.oracle import cx_oracle, base as oracle @@ -817,7 +819,7 @@ class TwoPhaseTest(fixtures.TablesTest): ) def test_twophase_prepare_false(self): conn = self._connection() - for i in xrange(2): + for i in range(2): trans = conn.begin_twophase() conn.execute("select 1 from dual") trans.prepare() @@ -827,7 +829,7 @@ class TwoPhaseTest(fixtures.TablesTest): def test_twophase_prepare_true(self): conn = self._connection() - for i in xrange(2): + for i in range(2): trans = conn.begin_twophase() conn.execute("insert into datatable (id, data) " "values (%s, 'somedata')" % i) @@ -880,7 +882,7 @@ class DialectTypesTest(fixtures.TestBase, AssertsCompiledSQL): b = bindparam("foo", "hello world!") assert b.type.dialect_impl(dialect).get_dbapi_type(dbapi) == 'STRING' - b = bindparam("foo", u"hello world!") + b = bindparam("foo", "hello world!") assert b.type.dialect_impl(dialect).get_dbapi_type(dbapi) == 'STRING' def test_long(self): @@ -1277,10 +1279,10 @@ class TypesTest(fixtures.TestBase): Column('data', oracle.RAW(35)) ) metadata.create_all() - testing.db.execute(raw_table.insert(), id=1, data="ABCDEF") + testing.db.execute(raw_table.insert(), id=1, data=b("ABCDEF")) eq_( testing.db.execute(raw_table.select()).first(), - (1, "ABCDEF") + (1, b("ABCDEF")) ) @testing.provide_metadata @@ -1301,11 +1303,11 @@ class TypesTest(fixtures.TestBase): t2.c.data.type.dialect_impl(testing.db.dialect), cx_oracle._OracleNVarChar) - data = u'm’a réveillé.' + data = u('m’a réveillé.') t2.insert().execute(data=data) res = t2.select().execute().first()['data'] eq_(res, data) - assert isinstance(res, unicode) + assert isinstance(res, util.text_type) def test_char_length(self): @@ -1367,10 +1369,10 @@ class TypesTest(fixtures.TestBase): try: engine.execute(t.insert(), id=1, data='this is text', - bindata='this is binary') + bindata=b('this is binary')) row = engine.execute(t.select()).first() eq_(row['data'].read(), 'this is text') - eq_(row['bindata'].read(), 'this is binary') + eq_(row['bindata'].read(), b('this is binary')) finally: t.drop(engine) @@ -1455,7 +1457,8 @@ class BufferedColumnTest(fixtures.TestBase, AssertsCompiledSQL): stream = os.path.join( os.path.dirname(__file__), "..", 'binary_data_one.dat') - stream = file(stream).read(12000) + with open(stream, "rb") as file_: + stream = file_.read(12000) for i in range(1, 11): binary_table.insert().execute(id=i, data=stream) @@ -1651,28 +1654,28 @@ class UnicodeSchemaTest(fixtures.TestBase): metadata.create_all() table.insert().execute( - {'_underscorecolumn': u'’é'}, + {'_underscorecolumn': u('’é')}, ) result = testing.db.execute( - table.select().where(table.c._underscorecolumn==u'’é') + table.select().where(table.c._underscorecolumn==u('’é')) ).scalar() - eq_(result, u'’é') + eq_(result, u('’é')) @testing.provide_metadata def test_quoted_column_unicode(self): metadata = self.metadata table=Table("atable", metadata, - Column(u"méil", Unicode(255), primary_key=True), + Column(u("méil"), Unicode(255), primary_key=True), ) metadata.create_all() table.insert().execute( - {u'méil': u'’é'}, + {u('méil'): u('’é')}, ) result = testing.db.execute( - table.select().where(table.c[u'méil']==u'’é') + table.select().where(table.c[u('méil')] == u('’é')) ).scalar() - eq_(result, u'’é') + eq_(result, u('’é')) class DBLinkReflectionTest(fixtures.TestBase): @@ -1712,5 +1715,5 @@ class DBLinkReflectionTest(fixtures.TestBase): t = Table('test_table_syn', m, autoload=True, autoload_with=testing.db, oracle_resolve_synonyms=True) - eq_(t.c.keys(), ['id', 'data']) + eq_(list(t.c.keys()), ['id', 'data']) eq_(list(t.primary_key), [t.c.id]) diff --git a/test/dialect/test_postgresql.py b/test/dialect/test_postgresql.py index 4fd5bc9c1..00e5c07ab 100644 --- a/test/dialect/test_postgresql.py +++ b/test/dialect/test_postgresql.py @@ -1,6 +1,6 @@ # coding: utf-8 -from __future__ import with_statement + from sqlalchemy.testing.assertions import eq_, assert_raises, \ assert_raises_message, is_, AssertsExecutionResults, \ @@ -555,28 +555,29 @@ class EnumTest(fixtures.TestBase, AssertsExecutionResults, AssertsCompiledSQL): @testing.fails_on('postgresql+pg8000', 'zxjdbc fails on ENUM: column "XXX" is of type ' 'XXX but expression is of type text') + @testing.provide_metadata def test_unicode_labels(self): - metadata = MetaData(testing.db) + metadata = self.metadata t1 = Table('table', metadata, Column('id', Integer, primary_key=True), Column('value', - Enum(u'réveillé', u'drôle', u'S’il', + Enum(util.u('réveillé'), util.u('drôle'), util.u('S’il'), name='onetwothreetype')) ) - metadata.create_all() - try: - t1.insert().execute(value=u'drôle') - t1.insert().execute(value=u'réveillé') - t1.insert().execute(value=u'S’il') - eq_(t1.select().order_by(t1.c.id).execute().fetchall(), - [(1, u'drôle'), (2, u'réveillé'), (3, u'S’il')] - ) - m2 = MetaData(testing.db) - t2 = Table('table', m2, autoload=True) - assert t2.c.value.type.enums == (u'réveillé', u'drôle', u'S’il') - finally: - metadata.drop_all() + t1.insert().execute(value=util.u('drôle')) + t1.insert().execute(value=util.u('réveillé')) + t1.insert().execute(value=util.u('S’il')) + eq_(t1.select().order_by(t1.c.id).execute().fetchall(), + [(1, util.u('drôle')), (2, util.u('réveillé')), + (3, util.u('S’il'))] + ) + m2 = MetaData(testing.db) + t2 = Table('table', m2, autoload=True) + eq_( + t2.c.value.type.enums, + (util.u('réveillé'), util.u('drôle'), util.u('S’il')) + ) def test_non_native_type(self): metadata = MetaData() @@ -1239,7 +1240,7 @@ class DomainReflectionTest(fixtures.TestBase, AssertsExecutionResults): : try: con.execute(ddl) - except exc.DBAPIError, e: + except exc.DBAPIError as e: if not 'already exists' in str(e): raise e con.execute('CREATE TABLE testtable (question integer, answer ' @@ -1477,7 +1478,7 @@ class ReflectionTest(fixtures.TestBase): meta1.create_all() meta2 = MetaData(testing.db) subject = Table('subject', meta2, autoload=True) - eq_(subject.primary_key.columns.keys(), [u'p2', u'p1']) + eq_(subject.primary_key.columns.keys(), ['p2', 'p1']) @testing.provide_metadata def test_pg_weirdchar_reflection(self): @@ -1750,7 +1751,7 @@ class ReflectionTest(fixtures.TestBase): conn.execute("ALTER TABLE t RENAME COLUMN x to y") ind = testing.db.dialect.get_indexes(conn, "t", None) - eq_(ind, [{'unique': False, 'column_names': [u'y'], 'name': u'idx1'}]) + eq_(ind, [{'unique': False, 'column_names': ['y'], 'name': 'idx1'}]) conn.close() class CustomTypeReflectionTest(fixtures.TestBase): @@ -2175,18 +2176,18 @@ class ArrayTest(fixtures.TablesTest, AssertsExecutionResults): def test_insert_array(self): arrtable = self.tables.arrtable - arrtable.insert().execute(intarr=[1, 2, 3], strarr=[u'abc', - u'def']) + arrtable.insert().execute(intarr=[1, 2, 3], strarr=[util.u('abc'), + util.u('def')]) results = arrtable.select().execute().fetchall() eq_(len(results), 1) eq_(results[0]['intarr'], [1, 2, 3]) - eq_(results[0]['strarr'], ['abc', 'def']) + eq_(results[0]['strarr'], [util.u('abc'), util.u('def')]) def test_array_where(self): arrtable = self.tables.arrtable - arrtable.insert().execute(intarr=[1, 2, 3], strarr=[u'abc', - u'def']) - arrtable.insert().execute(intarr=[4, 5, 6], strarr=u'ABC') + arrtable.insert().execute(intarr=[1, 2, 3], strarr=[util.u('abc'), + util.u('def')]) + arrtable.insert().execute(intarr=[4, 5, 6], strarr=util.u('ABC')) results = arrtable.select().where(arrtable.c.intarr == [1, 2, 3]).execute().fetchall() eq_(len(results), 1) @@ -2195,7 +2196,7 @@ class ArrayTest(fixtures.TablesTest, AssertsExecutionResults): def test_array_concat(self): arrtable = self.tables.arrtable arrtable.insert().execute(intarr=[1, 2, 3], - strarr=[u'abc', u'def']) + strarr=[util.u('abc'), util.u('def')]) results = select([arrtable.c.intarr + [4, 5, 6]]).execute().fetchall() eq_(len(results), 1) @@ -2204,15 +2205,15 @@ class ArrayTest(fixtures.TablesTest, AssertsExecutionResults): def test_array_subtype_resultprocessor(self): arrtable = self.tables.arrtable arrtable.insert().execute(intarr=[4, 5, 6], - strarr=[[u'm\xe4\xe4'], [u'm\xf6\xf6' - ]]) - arrtable.insert().execute(intarr=[1, 2, 3], strarr=[u'm\xe4\xe4' - , u'm\xf6\xf6']) + strarr=[[util.ue('m\xe4\xe4')], [ + util.ue('m\xf6\xf6')]]) + arrtable.insert().execute(intarr=[1, 2, 3], strarr=[ + util.ue('m\xe4\xe4'), util.ue('m\xf6\xf6')]) results = \ arrtable.select(order_by=[arrtable.c.intarr]).execute().fetchall() eq_(len(results), 2) - eq_(results[0]['strarr'], [u'm\xe4\xe4', u'm\xf6\xf6']) - eq_(results[1]['strarr'], [[u'm\xe4\xe4'], [u'm\xf6\xf6']]) + eq_(results[0]['strarr'], [util.ue('m\xe4\xe4'), util.ue('m\xf6\xf6')]) + eq_(results[1]['strarr'], [[util.ue('m\xe4\xe4')], [util.ue('m\xf6\xf6')]]) def test_array_literal(self): eq_( @@ -2264,7 +2265,7 @@ class ArrayTest(fixtures.TablesTest, AssertsExecutionResults): testing.db.execute( arrtable.insert(), intarr=[4, 5, 6], - strarr=[u'abc', u'def'] + strarr=[util.u('abc'), util.u('def')] ) eq_( testing.db.scalar(select([arrtable.c.intarr[2:3]])), @@ -2900,8 +2901,8 @@ class HStoreTest(fixtures.TestBase): dialect, None) assert_raises_message( ValueError, - r'''After '\[\.\.\.\], "key1"=>"value1", ', could not parse ''' - '''residual at position 36: 'crapcrapcrap, "key3"\[\.\.\.\]''', + r'''After u?'\[\.\.\.\], "key1"=>"value1", ', could not parse ''' + '''residual at position 36: u?'crapcrapcrap, "key3"\[\.\.\.\]''', proc, '"key2"=>"value2", "key1"=>"value1", ' 'crapcrapcrap, "key3"=>"value3"' @@ -3206,3 +3207,28 @@ class HStoreRoundTripTest(fixtures.TablesTest): def test_fixed_round_trip_native(self): engine = testing.db self._test_fixed_round_trip(engine) + + def _test_unicode_round_trip(self, engine): + s = select([ + hstore( + array([u'réveillé', u'drôle', u'S’il']), + array([u'réveillé', u'drôle', u'S’il']) + ) + ]) + eq_( + engine.scalar(s), + { + u'réveillé': u'réveillé', + u'drôle': u'drôle', + u'S’il': u'S’il' + } + ) + + def test_unicode_round_trip_python(self): + engine = self._non_native_engine() + self._test_unicode_round_trip(engine) + + @testing.only_on("postgresql+psycopg2") + def test_unicode_round_trip_native(self): + engine = testing.db + self._test_unicode_round_trip(engine) diff --git a/test/dialect/test_sqlite.py b/test/dialect/test_sqlite.py index 97962a54a..4ede13ff6 100644 --- a/test/dialect/test_sqlite.py +++ b/test/dialect/test_sqlite.py @@ -1,10 +1,17 @@ +#!coding: utf-8 + """SQLite-specific tests.""" from sqlalchemy.testing import eq_, assert_raises, \ assert_raises_message import datetime -from sqlalchemy import * -from sqlalchemy import exc, sql, schema, pool, types as sqltypes +from sqlalchemy import Table, String, select, Text, CHAR, bindparam, Column,\ + Unicode, Date, MetaData, UnicodeText, Time, Integer, TIMESTAMP, \ + Boolean, func, NUMERIC, DateTime, extract, ForeignKey, text, Numeric,\ + DefaultClause, and_, DECIMAL, TypeDecorator, create_engine, Float, \ + INTEGER, UniqueConstraint, DATETIME, DATE, TIME, BOOLEAN +from sqlalchemy.util import u, ue +from sqlalchemy import exc, sql, schema, pool, types as sqltypes, util from sqlalchemy.dialects.sqlite import base as sqlite, \ pysqlite as pysqlite_dialect from sqlalchemy.engine.url import make_url @@ -84,7 +91,7 @@ class TestTypes(fixtures.TestBase, AssertsExecutionResults): eq_(row, (1, datetime.date(2010, 5, 10), datetime.datetime( 2010, 5, 10, 12, 15, 25, ))) r = engine.execute(func.current_date()).scalar() - assert isinstance(r, basestring) + assert isinstance(r, util.string_types) finally: t.drop(engine) engine.dispose() @@ -104,8 +111,8 @@ class TestTypes(fixtures.TestBase, AssertsExecutionResults): UnicodeText(), ): bindproc = t.dialect_impl(dialect).bind_processor(dialect) - assert not bindproc or isinstance(bindproc(u'some string'), - unicode) + assert not bindproc or \ + isinstance(bindproc(util.u('some string')), util.text_type) @testing.provide_metadata def test_type_reflection(self): @@ -485,6 +492,20 @@ class DialectTest(fixtures.TestBase, AssertsExecutionResults): 'constrained_columns': ['tid'] }]) + @testing.provide_metadata + def test_description_encoding(self): + # amazingly, pysqlite seems to still deliver cursor.description + # as encoded bytes in py2k + + t = Table('x', self.metadata, + Column(u('méil'), Integer, primary_key=True), + Column(ue('\u6e2c\u8a66'), Integer), + ) + self.metadata.create_all(testing.db) + + result = testing.db.execute(t.select()) + assert u('méil') in result.keys() + assert ue('\u6e2c\u8a66') in result.keys() def test_attached_as_schema(self): cx = testing.db.connect() @@ -566,7 +587,7 @@ class DialectTest(fixtures.TestBase, AssertsExecutionResults): eq_(inspector.get_indexes('foo'), []) eq_(inspector.get_indexes('foo', include_auto_indexes=True), [{'unique': 1, 'name' - : u'sqlite_autoindex_foo_1', 'column_names': [u'bar']}]) + : 'sqlite_autoindex_foo_1', 'column_names': ['bar']}]) finally: meta.drop_all() diff --git a/test/dialect/test_sybase.py b/test/dialect/test_sybase.py index 025d49aae..1318a282b 100644 --- a/test/dialect/test_sybase.py +++ b/test/dialect/test_sybase.py @@ -19,7 +19,7 @@ class CompileTest(fixtures.TestBase, AssertsCompiledSQL): 'year': 'year', } - for field, subst in mapping.items(): + for field, subst in list(mapping.items()): self.assert_compile( select([extract(field, t.c.col1)]), 'SELECT DATEPART("%s", t.col1) AS anon_1 FROM t' % subst) diff --git a/test/engine/test_bind.py b/test/engine/test_bind.py index f76350fcc..973cf4d84 100644 --- a/test/engine/test_bind.py +++ b/test/engine/test_bind.py @@ -1,6 +1,6 @@ """tests the "bind" attribute/argument across schema and SQL, including the deprecated versions of these arguments""" -from __future__ import with_statement + from sqlalchemy.testing import eq_, assert_raises from sqlalchemy import engine, exc from sqlalchemy import MetaData, ThreadLocalMetaData @@ -61,7 +61,7 @@ class BindTest(fixtures.TestBase): try: meth() assert False - except exc.UnboundExecutionError, e: + except exc.UnboundExecutionError as e: eq_(str(e), "The MetaData is not bound to an Engine or " "Connection. Execution can not proceed without a " @@ -82,7 +82,7 @@ class BindTest(fixtures.TestBase): try: meth() assert False - except exc.UnboundExecutionError, e: + except exc.UnboundExecutionError as e: eq_( str(e), "The Table 'test_table' " diff --git a/test/engine/test_ddlemit.py b/test/engine/test_ddlemit.py index 3dbd5756a..deaf09cf7 100644 --- a/test/engine/test_ddlemit.py +++ b/test/engine/test_ddlemit.py @@ -47,7 +47,7 @@ class EmitDDLTest(fixtures.TestBase): return (m, ) + tuple( Table('t%d' % i, m, Column('x', Integer)) - for i in xrange(1, 6) + for i in range(1, 6) ) def _table_seq_fixture(self): diff --git a/test/engine/test_ddlevents.py b/test/engine/test_ddlevents.py index 71379ec7e..6cc652baf 100644 --- a/test/engine/test_ddlevents.py +++ b/test/engine/test_ddlevents.py @@ -1,4 +1,4 @@ -from __future__ import with_statement + from sqlalchemy.testing import assert_raises, assert_raises_message from sqlalchemy.schema import DDL, CheckConstraint, AddConstraint, \ DropConstraint diff --git a/test/engine/test_execute.py b/test/engine/test_execute.py index 203d7bd71..1c577730b 100644 --- a/test/engine/test_execute.py +++ b/test/engine/test_execute.py @@ -1,4 +1,4 @@ -from __future__ import with_statement + from sqlalchemy.testing import eq_, assert_raises, assert_raises_message, \ config, is_ @@ -12,13 +12,13 @@ from sqlalchemy.testing.schema import Table, Column import sqlalchemy as tsa from sqlalchemy import testing from sqlalchemy.testing import engines +from sqlalchemy import util from sqlalchemy.testing.engines import testing_engine import logging.handlers from sqlalchemy.dialects.oracle.zxjdbc import ReturningParam from sqlalchemy.engine import result as _result, default from sqlalchemy.engine.base import Connection, Engine from sqlalchemy.testing import fixtures -import StringIO users, metadata, users_autoinc = None, None, None class ExecuteTest(fixtures.TestBase): @@ -256,7 +256,7 @@ class ExecuteTest(fixtures.TestBase): try: cursor = raw.cursor() cursor.execute("SELECTINCORRECT") - except testing.db.dialect.dbapi.DatabaseError, orig: + except testing.db.dialect.dbapi.DatabaseError as orig: # py3k has "orig" in local scope... the_orig = orig finally: @@ -622,7 +622,7 @@ class LogParamsTest(fixtures.TestBase): def test_log_large_dict(self): self.eng.execute( "INSERT INTO foo (data) values (:data)", - [{"data":str(i)} for i in xrange(100)] + [{"data":str(i)} for i in range(100)] ) eq_( self.buf.buffer[1].message, @@ -635,7 +635,7 @@ class LogParamsTest(fixtures.TestBase): def test_log_large_list(self): self.eng.execute( "INSERT INTO foo (data) values (?)", - [(str(i), ) for i in xrange(100)] + [(str(i), ) for i in range(100)] ) eq_( self.buf.buffer[1].message, @@ -654,7 +654,7 @@ class LogParamsTest(fixtures.TestBase): "100 total bound parameter sets ... {'data': '98'}, {'data': '99'}\]", lambda: self.eng.execute( "INSERT INTO nonexistent (data) values (:data)", - [{"data":str(i)} for i in xrange(100)] + [{"data":str(i)} for i in range(100)] ) ) @@ -668,7 +668,7 @@ class LogParamsTest(fixtures.TestBase): "\('98',\), \('99',\)\]", lambda: self.eng.execute( "INSERT INTO nonexistent (data) values (?)", - [(str(i), ) for i in xrange(100)] + [(str(i), ) for i in range(100)] ) ) @@ -834,9 +834,9 @@ class EchoTest(fixtures.TestBase): class MockStrategyTest(fixtures.TestBase): def _engine_fixture(self): - buf = StringIO.StringIO() + buf = util.StringIO() def dump(sql, *multiparams, **params): - buf.write(unicode(sql.compile(dialect=engine.dialect))) + buf.write(util.text_type(sql.compile(dialect=engine.dialect))) engine = create_engine('postgresql://', strategy='mock', executor=dump) return engine, buf @@ -939,7 +939,6 @@ class ResultProxyTest(fixtures.TestBase): def test_row_c_sequence_check(self): import csv import collections - from StringIO import StringIO metadata = MetaData() metadata.bind = 'sqlite://' @@ -952,7 +951,7 @@ class ResultProxyTest(fixtures.TestBase): users.insert().execute(name='Test') row = users.select().execute().fetchone() - s = StringIO() + s = util.StringIO() writer = csv.writer(s) # csv performs PySequenceCheck call writer.writerow(row) @@ -1026,7 +1025,7 @@ class AlternateResultProxyTest(fixtures.TestBase): ) m.create_all(engine) engine.execute(t.insert(), [ - {'x':i, 'y':"t_%d" % i} for i in xrange(1, 12) + {'x':i, 'y':"t_%d" % i} for i in range(1, 12) ]) def _test_proxy(self, cls): @@ -1039,13 +1038,13 @@ class AlternateResultProxyTest(fixtures.TestBase): assert isinstance(r, cls) for i in range(5): rows.append(r.fetchone()) - eq_(rows, [(i, "t_%d" % i) for i in xrange(1, 6)]) + eq_(rows, [(i, "t_%d" % i) for i in range(1, 6)]) rows = r.fetchmany(3) - eq_(rows, [(i, "t_%d" % i) for i in xrange(6, 9)]) + eq_(rows, [(i, "t_%d" % i) for i in range(6, 9)]) rows = r.fetchall() - eq_(rows, [(i, "t_%d" % i) for i in xrange(9, 12)]) + eq_(rows, [(i, "t_%d" % i) for i in range(9, 12)]) r = self.engine.execute(select([self.table])) rows = r.fetchmany(None) @@ -1059,7 +1058,7 @@ class AlternateResultProxyTest(fixtures.TestBase): r = self.engine.execute(select([self.table]).limit(5)) rows = r.fetchmany(6) - eq_(rows, [(i, "t_%d" % i) for i in xrange(1, 6)]) + eq_(rows, [(i, "t_%d" % i) for i in range(1, 6)]) def test_plain(self): self._test_proxy(_result.ResultProxy) @@ -1184,7 +1183,7 @@ class EngineEventsTest(fixtures.TestBase): try: conn.execute("SELECT FOO FROM I_DONT_EXIST") assert False - except tsa.exc.DBAPIError, e: + except tsa.exc.DBAPIError as e: assert canary[0][2] is e.orig assert canary[0][0] == "SELECT FOO FROM I_DONT_EXIST" diff --git a/test/engine/test_parseconnect.py b/test/engine/test_parseconnect.py index a00a942cb..73bdc76c4 100644 --- a/test/engine/test_parseconnect.py +++ b/test/engine/test_parseconnect.py @@ -1,6 +1,5 @@ -from sqlalchemy.testing import assert_raises, assert_raises_message, eq_ -import ConfigParser -import StringIO +from sqlalchemy.testing import assert_raises, eq_ +from sqlalchemy.util.compat import configparser, StringIO import sqlalchemy.engine.url as url from sqlalchemy import create_engine, engine_from_config, exc, pool from sqlalchemy.engine.util import _coerce_config @@ -103,8 +102,8 @@ pool_size=2 pool_threadlocal=1 pool_timeout=10 """ - ini = ConfigParser.ConfigParser() - ini.readfp(StringIO.StringIO(raw)) + ini = configparser.ConfigParser() + ini.readfp(StringIO(raw)) expected = { 'url': 'postgresql://scott:tiger@somehost/test?fooz=somevalue', @@ -234,7 +233,7 @@ pool_timeout=10 : True}, convert_unicode=True) try: e.connect() - except tsa.exc.DBAPIError, de: + except tsa.exc.DBAPIError as de: assert not de.connection_invalidated def test_ensure_dialect_does_is_disconnect_no_conn(self): @@ -266,7 +265,7 @@ pool_timeout=10 try: create_engine('sqlite://', module=ThrowOnConnect()).connect() assert False - except tsa.exc.DBAPIError, de: + except tsa.exc.DBAPIError as de: assert de.connection_invalidated def test_urlattr(self): diff --git a/test/engine/test_pool.py b/test/engine/test_pool.py index 9551086f5..260f461d0 100644 --- a/test/engine/test_pool.py +++ b/test/engine/test_pool.py @@ -527,23 +527,23 @@ class DeprecatedPoolListenerTest(PoolTestBase): self.assert_((item in innerself.checked_out) == in_cout) self.assert_((item in innerself.checked_in) == in_cin) def inst_connect(self, con, record): - print "connect(%s, %s)" % (con, record) + print("connect(%s, %s)" % (con, record)) assert con is not None assert record is not None self.connected.append(con) def inst_first_connect(self, con, record): - print "first_connect(%s, %s)" % (con, record) + print("first_connect(%s, %s)" % (con, record)) assert con is not None assert record is not None self.first_connected.append(con) def inst_checkout(self, con, record, proxy): - print "checkout(%s, %s, %s)" % (con, record, proxy) + print("checkout(%s, %s, %s)" % (con, record, proxy)) assert con is not None assert record is not None assert proxy is not None self.checked_out.append(con) def inst_checkin(self, con, record): - print "checkin(%s, %s)" % (con, record) + print("checkin(%s, %s)" % (con, record)) # con can be None if invalidated assert record is not None self.checked_in.append(con) @@ -740,8 +740,8 @@ class QueuePoolTest(PoolTestBase): def status(pool): tup = pool.size(), pool.checkedin(), pool.overflow(), \ pool.checkedout() - print 'Pool size: %d Connections in pool: %d Current '\ - 'Overflow: %d Current Checked out connections: %d' % tup + print('Pool size: %d Connections in pool: %d Current '\ + 'Overflow: %d Current Checked out connections: %d' % tup) return tup c1 = p.connect() @@ -814,7 +814,7 @@ class QueuePoolTest(PoolTestBase): max_overflow=1, use_threadlocal=False, timeout=3) timeouts = [] def checkout(): - for x in xrange(1): + for x in range(1): now = time.time() try: c1 = p.connect() @@ -825,7 +825,7 @@ class QueuePoolTest(PoolTestBase): c1.close() threads = [] - for i in xrange(10): + for i in range(10): th = threading.Thread(target=checkout) th.start() threads.append(th) @@ -862,7 +862,7 @@ class QueuePoolTest(PoolTestBase): except tsa.exc.TimeoutError: pass threads = [] - for i in xrange(thread_count): + for i in range(thread_count): th = threading.Thread(target=whammy) th.start() threads.append(th) @@ -1009,8 +1009,8 @@ class QueuePoolTest(PoolTestBase): strong_refs.add(c.connection) return c - for j in xrange(5): - conns = [_conn() for i in xrange(4)] + for j in range(5): + conns = [_conn() for i in range(4)] for c in conns: c.close() @@ -1154,7 +1154,7 @@ class SingletonThreadPoolTest(PoolTestBase): return p.connect() def checkout(): - for x in xrange(10): + for x in range(10): c = _conn() assert c c.cursor() @@ -1162,7 +1162,7 @@ class SingletonThreadPoolTest(PoolTestBase): time.sleep(.1) threads = [] - for i in xrange(10): + for i in range(10): th = threading.Thread(target=checkout) th.start() threads.append(th) diff --git a/test/engine/test_processors.py b/test/engine/test_processors.py index bc9af7305..b1c482f09 100644 --- a/test/engine/test_processors.py +++ b/test/engine/test_processors.py @@ -53,7 +53,7 @@ class PyDateProcessorTest(_DateProcessorTest): cls.module = type("util", (object,), dict( (k, staticmethod(v)) - for k, v in processors.py_fallback().items() + for k, v in list(processors.py_fallback().items()) ) ) @@ -156,7 +156,7 @@ class PyDistillArgsTest(_DistillArgsTest): cls.module = type("util", (object,), dict( (k, staticmethod(v)) - for k, v in util.py_fallback().items() + for k, v in list(util.py_fallback().items()) ) ) diff --git a/test/engine/test_reconnect.py b/test/engine/test_reconnect.py index 567647f20..ee3ff1459 100644 --- a/test/engine/test_reconnect.py +++ b/test/engine/test_reconnect.py @@ -7,7 +7,7 @@ import sqlalchemy as tsa from sqlalchemy import testing from sqlalchemy.testing import engines from sqlalchemy.testing.util import gc_collect -from sqlalchemy import exc +from sqlalchemy import exc, util from sqlalchemy.testing import fixtures from sqlalchemy.testing.engines import testing_engine @@ -173,7 +173,7 @@ class MockReconnectTest(fixtures.TestBase): try: trans.commit() assert False - except tsa.exc.InvalidRequestError, e: + except tsa.exc.InvalidRequestError as e: assert str(e) \ == "Can't reconnect until invalid transaction is "\ "rolled back" @@ -370,7 +370,7 @@ class RealReconnectTest(fixtures.TestBase): try: conn.execute(select([1])) assert False - except tsa.exc.DBAPIError, e: + except tsa.exc.DBAPIError as e: if not e.connection_invalidated: raise @@ -386,7 +386,7 @@ class RealReconnectTest(fixtures.TestBase): try: conn.execute(select([1])) assert False - except tsa.exc.DBAPIError, e: + except tsa.exc.DBAPIError as e: if not e.connection_invalidated: raise assert conn.invalidated @@ -407,7 +407,7 @@ class RealReconnectTest(fixtures.TestBase): try: c1.execute(select([1])) assert False - except tsa.exc.DBAPIError, e: + except tsa.exc.DBAPIError as e: assert e.connection_invalidated p2 = engine.pool @@ -415,7 +415,7 @@ class RealReconnectTest(fixtures.TestBase): try: c2.execute(select([1])) assert False - except tsa.exc.DBAPIError, e: + except tsa.exc.DBAPIError as e: assert e.connection_invalidated # pool isn't replaced @@ -464,6 +464,9 @@ class RealReconnectTest(fixtures.TestBase): conn.invalidate() conn.invalidate() + @testing.skip_if( + [lambda: util.py3k, "oracle+cx_oracle"], + "Crashes on py3k+cx_oracle") def test_explode_in_initializer(self): engine = engines.testing_engine() def broken_initialize(connection): @@ -503,7 +506,7 @@ class RealReconnectTest(fixtures.TestBase): try: conn.execute(select([1])) assert False - except tsa.exc.DBAPIError, e: + except tsa.exc.DBAPIError as e: if not e.connection_invalidated: raise assert not conn.closed @@ -523,7 +526,7 @@ class RealReconnectTest(fixtures.TestBase): try: conn.execute(select([1])) assert False - except tsa.exc.DBAPIError, e: + except tsa.exc.DBAPIError as e: if not e.connection_invalidated: raise @@ -542,7 +545,7 @@ class RealReconnectTest(fixtures.TestBase): try: conn.execute(select([1])) assert False - except tsa.exc.DBAPIError, e: + except tsa.exc.DBAPIError as e: if not e.connection_invalidated: raise assert not conn.closed @@ -558,7 +561,7 @@ class RealReconnectTest(fixtures.TestBase): try: trans.commit() assert False - except tsa.exc.InvalidRequestError, e: + except tsa.exc.InvalidRequestError as e: assert str(e) \ == "Can't reconnect until invalid transaction is "\ "rolled back" @@ -627,13 +630,13 @@ class InvalidateDuringResultTest(fixtures.TestBase): def test_invalidate_on_results(self): conn = engine.connect() result = conn.execute('select * from sometable') - for x in xrange(20): + for x in range(20): result.fetchone() engine.test_shutdown() try: - print 'ghost result: %r' % result.fetchone() + print('ghost result: %r' % result.fetchone()) assert False - except tsa.exc.DBAPIError, e: + except tsa.exc.DBAPIError as e: if not e.connection_invalidated: raise assert conn.invalidated diff --git a/test/engine/test_reflection.py b/test/engine/test_reflection.py index a2f6b3796..a562ef73b 100644 --- a/test/engine/test_reflection.py +++ b/test/engine/test_reflection.py @@ -7,6 +7,8 @@ from sqlalchemy.testing import ComparesTables, \ from sqlalchemy.testing.schema import Table, Column from sqlalchemy.testing import eq_, assert_raises, assert_raises_message from sqlalchemy import testing +from sqlalchemy.util import ue + metadata, users = None, None @@ -808,7 +810,7 @@ class ReflectionTest(fixtures.TestBase, ComparesTables): try: m4.reflect(only=['rt_a', 'rt_f']) self.assert_(False) - except sa.exc.InvalidRequestError, e: + except sa.exc.InvalidRequestError as e: self.assert_(e.args[0].endswith('(rt_f)')) m5 = MetaData(testing.db) @@ -830,7 +832,7 @@ class ReflectionTest(fixtures.TestBase, ComparesTables): ) if existing: - print "Other tables present in database, skipping some checks." + print("Other tables present in database, skipping some checks.") else: baseline.drop_all() m9 = MetaData(testing.db) @@ -1039,19 +1041,19 @@ class UnicodeReflectionTest(fixtures.TestBase): cls.metadata = metadata = MetaData() no_multibyte_period = set([ - (u'plain', u'col_plain', u'ix_plain') + ('plain', 'col_plain', 'ix_plain') ]) no_has_table = [ - (u'no_has_table_1', u'col_Unit\u00e9ble', u'ix_Unit\u00e9ble'), - (u'no_has_table_2', u'col_\u6e2c\u8a66', u'ix_\u6e2c\u8a66'), + ('no_has_table_1', ue('col_Unit\u00e9ble'), ue('ix_Unit\u00e9ble')), + ('no_has_table_2', ue('col_\u6e2c\u8a66'), ue('ix_\u6e2c\u8a66')), ] no_case_sensitivity = [ - (u'\u6e2c\u8a66', u'col_\u6e2c\u8a66', u'ix_\u6e2c\u8a66'), - (u'unit\u00e9ble', u'col_unit\u00e9ble', u'ix_unit\u00e9ble'), + (ue('\u6e2c\u8a66'), ue('col_\u6e2c\u8a66'), ue('ix_\u6e2c\u8a66')), + (ue('unit\u00e9ble'), ue('col_unit\u00e9ble'), ue('ix_unit\u00e9ble')), ] full = [ - (u'Unit\u00e9ble', u'col_Unit\u00e9ble', u'ix_Unit\u00e9ble'), - (u'\u6e2c\u8a66', u'col_\u6e2c\u8a66', u'ix_\u6e2c\u8a66'), + (ue('Unit\u00e9ble'), ue('col_Unit\u00e9ble'), ue('ix_Unit\u00e9ble')), + (ue('\u6e2c\u8a66'), ue('col_\u6e2c\u8a66'), ue('ix_\u6e2c\u8a66')), ] # as you can see, our options for this kind of thing diff --git a/test/engine/test_transaction.py b/test/engine/test_transaction.py index 5558ff778..2d7f39253 100644 --- a/test/engine/test_transaction.py +++ b/test/engine/test_transaction.py @@ -74,8 +74,8 @@ class TransactionTest(fixtures.TestBase): connection.execute(users.insert(), user_id=1, user_name='user3') transaction.commit() assert False - except Exception , e: - print "Exception: ", e + except Exception as e: + print("Exception: ", e) transaction.rollback() result = connection.execute("select * from query_users") @@ -121,10 +121,10 @@ class TransactionTest(fixtures.TestBase): trans2.rollback() raise transaction.rollback() - except Exception, e: + except Exception as e: transaction.rollback() raise - except Exception, e: + except Exception as e: try: assert str(e) == 'uh oh' # and not "This transaction is # inactive" @@ -167,7 +167,7 @@ class TransactionTest(fixtures.TestBase): connection.execute(users.insert(), user_id=2, user_name='user2') try: connection.execute(users.insert(), user_id=2, user_name='user2.5') - except Exception, e: + except Exception as e: trans.__exit__(*sys.exc_info()) assert not trans.is_active @@ -1019,7 +1019,7 @@ class ForUpdateTest(fixtures.TestBase): con = testing.db.connect() sel = counters.select(for_update=update_style, whereclause=counters.c.counter_id == 1) - for i in xrange(count): + for i in range(count): trans = con.begin() try: existing = con.execute(sel).first() @@ -1033,7 +1033,7 @@ class ForUpdateTest(fixtures.TestBase): raise AssertionError('Got %s post-update, expected ' '%s' % (readback['counter_value'], incr)) trans.commit() - except Exception, e: + except Exception as e: trans.rollback() errors.append(e) break @@ -1057,7 +1057,7 @@ class ForUpdateTest(fixtures.TestBase): db.execute(counters.insert(), counter_id=1, counter_value=0) iterations, thread_count = 10, 5 threads, errors = [], [] - for i in xrange(thread_count): + for i in range(thread_count): thrd = threading.Thread(target=self.increment, args=(iterations, ), kwargs={'errors': errors, @@ -1088,7 +1088,7 @@ class ForUpdateTest(fixtures.TestBase): rows = con.execute(sel).fetchall() time.sleep(0.25) trans.commit() - except Exception, e: + except Exception as e: trans.rollback() errors.append(e) con.close() @@ -1105,7 +1105,7 @@ class ForUpdateTest(fixtures.TestBase): db.execute(counters.insert(), counter_id=cid + 1, counter_value=0) errors, threads = [], [] - for i in xrange(thread_count): + for i in range(thread_count): thrd = threading.Thread(target=self.overlap, args=(groups.pop(0), errors, update_style)) diff --git a/test/ext/declarative/test_basic.py b/test/ext/declarative/test_basic.py index 0fe54a154..ad2970b70 100644 --- a/test/ext/declarative/test_basic.py +++ b/test/ext/declarative/test_basic.py @@ -13,7 +13,7 @@ from sqlalchemy.orm import relationship, create_session, class_mapper, \ deferred, column_property, composite,\ Session from sqlalchemy.testing import eq_ -from sqlalchemy.util import classproperty +from sqlalchemy.util import classproperty, with_metaclass from sqlalchemy.ext.declarative import declared_attr, AbstractConcreteBase, \ ConcreteBase, synonym_for from sqlalchemy.testing import fixtures @@ -127,14 +127,13 @@ class DeclarativeTest(DeclarativeTestBase): class BrokenMeta(type): def __getattribute__(self, attr): if attr == 'xyzzy': - raise AttributeError, 'xyzzy' + raise AttributeError('xyzzy') else: return object.__getattribute__(self,attr) # even though this class has an xyzzy attribute, getattr(cls,"xyzzy") # fails - class BrokenParent(object): - __metaclass__ = BrokenMeta + class BrokenParent(with_metaclass(BrokenMeta)): xyzzy = "magic" # _as_declarative() inspects obj.__class__.__bases__ @@ -275,7 +274,7 @@ class DeclarativeTest(DeclarativeTestBase): __tablename__ = 'foo' id = Column(Integer, primary_key=True) - _user_id = Column(Integer) + _user_id = Column(Integer) rel = relationship('User', uselist=False, foreign_keys=[User.id], @@ -1458,12 +1457,12 @@ def _produce_test(inline, stringbased): @classmethod def insert_data(cls): - params = [dict(zip(('id', 'name'), column_values)) + params = [dict(list(zip(('id', 'name'), column_values))) for column_values in [(7, 'jack'), (8, 'ed'), (9, 'fred'), (10, 'chuck')]] User.__table__.insert().execute(params) - Address.__table__.insert().execute([dict(zip(('id', - 'user_id', 'email'), column_values)) + Address.__table__.insert().execute([dict(list(zip(('id', + 'user_id', 'email'), column_values))) for column_values in [(1, 7, 'jack@bean.com'), (2, 8, 'ed@wood.com'), (3, 8, 'ed@bettyboop.com'), (4, 8, 'ed@lala.com'), (5, 9, 'fred@fred.com')]]) @@ -1492,6 +1491,6 @@ def _produce_test(inline, stringbased): for inline in True, False: for stringbased in True, False: testclass = _produce_test(inline, stringbased) - exec '%s = testclass' % testclass.__name__ + exec('%s = testclass' % testclass.__name__) del testclass diff --git a/test/ext/declarative/test_clsregistry.py b/test/ext/declarative/test_clsregistry.py index 1d09f158e..e78a1abbe 100644 --- a/test/ext/declarative/test_clsregistry.py +++ b/test/ext/declarative/test_clsregistry.py @@ -124,7 +124,7 @@ class ClsRegistryTest(fixtures.TestBase): base = weakref.WeakValueDictionary() - for i in xrange(3): + for i in range(3): f1 = MockClass(base, "foo.bar.Foo") f2 = MockClass(base, "foo.alt.Foo") clsregistry.add_class("Foo", f1) diff --git a/test/ext/declarative/test_inheritance.py b/test/ext/declarative/test_inheritance.py index 1a3a4aba5..01bf3f3f6 100644 --- a/test/ext/declarative/test_inheritance.py +++ b/test/ext/declarative/test_inheritance.py @@ -573,7 +573,7 @@ class DeclarativeInheritanceTest(DeclarativeTestBase): primary_language = Column(String(50)) assert Person.__table__.c.golf_swing is not None - assert not Person.__table__.c.has_key('primary_language') + assert 'primary_language' not in Person.__table__.c assert Engineer.__table__.c.primary_language is not None assert Engineer.primary_language is not None assert Manager.golf_swing is not None diff --git a/test/ext/declarative/test_mixin.py b/test/ext/declarative/test_mixin.py index fb674f27a..d8ec484e1 100644 --- a/test/ext/declarative/test_mixin.py +++ b/test/ext/declarative/test_mixin.py @@ -685,7 +685,7 @@ class DeclarativeMixinTest(DeclarativeTestBase): __mapper_args__ = dict(polymorphic_identity='specific') assert Specific.__table__ is Generic.__table__ - eq_(Generic.__table__.c.keys(), ['id', 'type', 'value']) + eq_(list(Generic.__table__.c.keys()), ['id', 'type', 'value']) assert class_mapper(Specific).polymorphic_on \ is Generic.__table__.c.type eq_(class_mapper(Specific).polymorphic_identity, 'specific') @@ -714,9 +714,9 @@ class DeclarativeMixinTest(DeclarativeTestBase): eq_(Generic.__table__.name, 'generic') eq_(Specific.__table__.name, 'specific') - eq_(Generic.__table__.c.keys(), ['timestamp', 'id', + eq_(list(Generic.__table__.c.keys()), ['timestamp', 'id', 'python_type']) - eq_(Specific.__table__.c.keys(), ['id']) + eq_(list(Specific.__table__.c.keys()), ['id']) eq_(Generic.__table__.kwargs, {'mysql_engine': 'InnoDB'}) eq_(Specific.__table__.kwargs, {'mysql_engine': 'InnoDB'}) @@ -749,12 +749,12 @@ class DeclarativeMixinTest(DeclarativeTestBase): primary_key=True) eq_(BaseType.__table__.name, 'basetype') - eq_(BaseType.__table__.c.keys(), ['timestamp', 'type', 'id', + eq_(list(BaseType.__table__.c.keys()), ['timestamp', 'type', 'id', 'value']) eq_(BaseType.__table__.kwargs, {'mysql_engine': 'InnoDB'}) assert Single.__table__ is BaseType.__table__ eq_(Joined.__table__.name, 'joined') - eq_(Joined.__table__.c.keys(), ['id']) + eq_(list(Joined.__table__.c.keys()), ['id']) eq_(Joined.__table__.kwargs, {'mysql_engine': 'InnoDB'}) def test_col_copy_vs_declared_attr_joined_propagation(self): @@ -839,7 +839,7 @@ class DeclarativeMixinTest(DeclarativeTestBase): __mapper_args__ = dict(polymorphic_identity='specific') eq_(BaseType.__table__.name, 'basetype') - eq_(BaseType.__table__.c.keys(), ['type', 'id', 'value']) + eq_(list(BaseType.__table__.c.keys()), ['type', 'id', 'value']) assert Specific.__table__ is BaseType.__table__ assert class_mapper(Specific).polymorphic_on \ is BaseType.__table__.c.type @@ -870,9 +870,9 @@ class DeclarativeMixinTest(DeclarativeTestBase): primary_key=True) eq_(BaseType.__table__.name, 'basetype') - eq_(BaseType.__table__.c.keys(), ['type', 'id', 'value']) + eq_(list(BaseType.__table__.c.keys()), ['type', 'id', 'value']) eq_(Specific.__table__.name, 'specific') - eq_(Specific.__table__.c.keys(), ['id']) + eq_(list(Specific.__table__.c.keys()), ['id']) def test_single_back_propagate(self): @@ -891,7 +891,7 @@ class DeclarativeMixinTest(DeclarativeTestBase): __mapper_args__ = dict(polymorphic_identity='specific') - eq_(BaseType.__table__.c.keys(), ['type', 'id', 'timestamp']) + eq_(list(BaseType.__table__.c.keys()), ['type', 'id', 'timestamp']) def test_table_in_model_and_same_column_in_mixin(self): @@ -987,7 +987,7 @@ class DeclarativeMixinTest(DeclarativeTestBase): id = Column(Integer, primary_key=True) __tablename__ = 'model' - eq_(Model.__table__.c.keys(), ['col1', 'col3', 'col2', 'col4', + eq_(list(Model.__table__.c.keys()), ['col1', 'col3', 'col2', 'col4', 'id']) def test_honor_class_mro_one(self): @@ -1078,8 +1078,8 @@ class DeclarativeMixinTest(DeclarativeTestBase): filter_class = FilterB id = Column(Integer(), primary_key=True) - TypeA(filters=[u'foo']) - TypeB(filters=[u'foo']) + TypeA(filters=['foo']) + TypeB(filters=['foo']) class DeclarativeMixinPropertyTest(DeclarativeTestBase): diff --git a/test/ext/test_associationproxy.py b/test/ext/test_associationproxy.py index c15b98bf2..a5fcc45cc 100644 --- a/test/ext/test_associationproxy.py +++ b/test/ext/test_associationproxy.py @@ -288,7 +288,7 @@ class CustomDictTest(DictTest): ch = Child('a', 'regular') p1._children.append(ch) - self.assert_(ch in p1._children.values()) + self.assert_(ch in list(p1._children.values())) self.assert_(len(p1._children) == 1) self.assert_(p1.children) @@ -300,7 +300,7 @@ class CustomDictTest(DictTest): p1.children['b'] = 'proxied' - self.assert_('proxied' in p1.children.values()) + self.assert_('proxied' in list(p1.children.values())) self.assert_('b' in p1.children) self.assert_('proxied' not in p1._children) self.assert_(len(p1.children) == 2) @@ -526,9 +526,9 @@ class SetTest(_CollectionOperations): try: self.assert_(p.children == control) except: - print 'Test %s.%s(%s):' % (set(base), op, other) - print 'want', repr(control) - print 'got', repr(p.children) + print('Test %s.%s(%s):' % (set(base), op, other)) + print('want', repr(control)) + print('got', repr(p.children)) raise p = self.roundtrip(p) @@ -536,9 +536,9 @@ class SetTest(_CollectionOperations): try: self.assert_(p.children == control) except: - print 'Test %s.%s(%s):' % (base, op, other) - print 'want', repr(control) - print 'got', repr(p.children) + print('Test %s.%s(%s):' % (base, op, other)) + print('want', repr(control)) + print('got', repr(p.children)) raise # in-place mutations @@ -553,15 +553,15 @@ class SetTest(_CollectionOperations): p.children = base[:] control = set(base[:]) - exec "p.children %s other" % op - exec "control %s other" % op + exec("p.children %s other" % op) + exec("control %s other" % op) try: self.assert_(p.children == control) except: - print 'Test %s %s %s:' % (set(base), op, other) - print 'want', repr(control) - print 'got', repr(p.children) + print('Test %s %s %s:' % (set(base), op, other)) + print('want', repr(control)) + print('got', repr(p.children)) raise p = self.roundtrip(p) @@ -569,9 +569,9 @@ class SetTest(_CollectionOperations): try: self.assert_(p.children == control) except: - print 'Test %s %s %s:' % (base, op, other) - print 'want', repr(control) - print 'got', repr(p.children) + print('Test %s %s %s:' % (base, op, other)) + print('want', repr(control)) + print('got', repr(p.children)) raise diff --git a/test/ext/test_serializer.py b/test/ext/test_serializer.py index 74e033b4f..84fff1304 100644 --- a/test/ext/test_serializer.py +++ b/test/ext/test_serializer.py @@ -47,12 +47,12 @@ class SerializeTest(fixtures.MappedTest): @classmethod def insert_data(cls): - params = [dict(zip(('id', 'name'), column_values)) + params = [dict(list(zip(('id', 'name'), column_values))) for column_values in [(7, 'jack'), (8, 'ed'), (9, 'fred'), (10, 'chuck')]] users.insert().execute(params) - addresses.insert().execute([dict(zip(('id', 'user_id', 'email' - ), column_values)) + addresses.insert().execute([dict(list(zip(('id', 'user_id', 'email' + ), column_values))) for column_values in [(1, 7, 'jack@bean.com'), (2, 8, 'ed@wood.com'), (3, 8, @@ -85,8 +85,8 @@ class SerializeTest(fixtures.MappedTest): users.metadata, None) eq_(str(expr), str(re_expr)) assert re_expr.bind is testing.db - eq_(re_expr.execute().fetchall(), [(7, u'jack'), (8, u'ed'), - (8, u'ed'), (8, u'ed'), (9, u'fred')]) + eq_(re_expr.execute().fetchall(), [(7, 'jack'), (8, 'ed'), + (8, 'ed'), (8, 'ed'), (9, 'fred')]) def test_query_one(self): q = Session.query(User).\ @@ -114,7 +114,7 @@ class SerializeTest(fixtures.MappedTest): Address(email='ed@lala.com'), Address(email='ed@bettyboop.com')]) - @testing.skip_if(lambda: util.pypy, "problems with pypy pickle reported") + @testing.requires.non_broken_pickle def test_query_two(self): q = \ Session.query(User).join(User.addresses).\ @@ -122,9 +122,9 @@ class SerializeTest(fixtures.MappedTest): q2 = serializer.loads(serializer.dumps(q, -1), users.metadata, Session) eq_(q2.all(), [User(name='fred')]) - eq_(list(q2.values(User.id, User.name)), [(9, u'fred')]) + eq_(list(q2.values(User.id, User.name)), [(9, 'fred')]) - @testing.skip_if(lambda: util.pypy, "problems with pypy pickle reported") + @testing.requires.non_broken_pickle def test_query_three(self): ua = aliased(User) q = \ @@ -136,9 +136,9 @@ class SerializeTest(fixtures.MappedTest): # try to pull out the aliased entity here... ua_2 = q2._entities[0].entity_zero.entity - eq_(list(q2.values(ua_2.id, ua_2.name)), [(9, u'fred')]) + eq_(list(q2.values(ua_2.id, ua_2.name)), [(9, 'fred')]) - @testing.skip_if(lambda: util.pypy, "problems with pypy pickle reported") + @testing.requires.non_broken_pickle def test_orm_join(self): from sqlalchemy.orm.util import join @@ -165,8 +165,7 @@ class SerializeTest(fixtures.MappedTest): eq_(list(q2.all()), [(u7, u8), (u7, u9), (u7, u10), (u8, u9), (u8, u10)]) - @testing.skip_if(lambda: util.pypy, "pickle sometimes has " - "problems here, sometimes not") + @testing.requires.non_broken_pickle def test_any(self): r = User.addresses.any(Address.email == 'x') ser = serializer.dumps(r, -1) diff --git a/test/orm/inheritance/test_assorted_poly.py b/test/orm/inheritance/test_assorted_poly.py index e8e6ba82a..d05a22f39 100644 --- a/test/orm/inheritance/test_assorted_poly.py +++ b/test/orm/inheritance/test_assorted_poly.py @@ -18,7 +18,7 @@ from sqlalchemy.testing.schema import Table, Column class AttrSettable(object): def __init__(self, **kwargs): - [setattr(self, k, v) for k, v in kwargs.iteritems()] + [setattr(self, k, v) for k, v in kwargs.items()] def __repr__(self): return self.__class__.__name__ + "(%s)" % (hex(id(self))) @@ -386,7 +386,7 @@ class RelationshipTest4(fixtures.MappedTest): # class definitions class Person(object): def __init__(self, **kwargs): - for key, value in kwargs.iteritems(): + for key, value in kwargs.items(): setattr(self, key, value) def __repr__(self): return "Ordinary person %s" % self.name @@ -400,7 +400,7 @@ class RelationshipTest4(fixtures.MappedTest): (self.name, self.longer_status) class Car(object): def __init__(self, **kwargs): - for key, value in kwargs.iteritems(): + for key, value in kwargs.items(): setattr(self, key, value) def __repr__(self): return "Car number %d" % self.car_id @@ -512,7 +512,7 @@ class RelationshipTest5(fixtures.MappedTest): class Person(object): def __init__(self, **kwargs): - for key, value in kwargs.iteritems(): + for key, value in kwargs.items(): setattr(self, key, value) def __repr__(self): return "Ordinary person %s" % self.name @@ -526,7 +526,7 @@ class RelationshipTest5(fixtures.MappedTest): (self.name, self.longer_status) class Car(object): def __init__(self, **kwargs): - for key, value in kwargs.iteritems(): + for key, value in kwargs.items(): setattr(self, key, value) def __repr__(self): return "Car number %d" % self.car_id @@ -646,7 +646,7 @@ class RelationshipTest7(fixtures.MappedTest): class PersistentObject(object): def __init__(self, **kwargs): - for key, value in kwargs.iteritems(): + for key, value in kwargs.items(): setattr(self, key, value) class Status(PersistentObject): @@ -838,7 +838,7 @@ class GenerativeTest(fixtures.TestBase, AssertsExecutionResults): # class definitions class PersistentObject(object): def __init__(self, **kwargs): - for key, value in kwargs.iteritems(): + for key, value in kwargs.items(): setattr(self, key, value) class Status(PersistentObject): def __repr__(self): @@ -968,7 +968,7 @@ class MultiLevelTest(fixtures.MappedTest): def test_threelevels(self): class Employee( object): def set( me, **kargs): - for k,v in kargs.iteritems(): setattr( me, k, v) + for k,v in kargs.items(): setattr( me, k, v) return me def __str__(me): return str(me.__class__.__name__)+':'+str(me.name) diff --git a/test/orm/inheritance/test_basic.py b/test/orm/inheritance/test_basic.py index bbfa54383..afd63f2b4 100644 --- a/test/orm/inheritance/test_basic.py +++ b/test/orm/inheritance/test_basic.py @@ -1355,7 +1355,7 @@ class SyncCompileTest(fixtures.MappedTest): def _do_test(self, j1, j2): class A(object): def __init__(self, **kwargs): - for key, value in kwargs.items(): + for key, value in list(kwargs.items()): setattr(self, key, value) class B(A): @@ -1830,7 +1830,7 @@ class OptimizedLoadTest(fixtures.MappedTest): "SELECT sub.counter AS sub_counter, base.counter AS base_counter, " "sub.counter2 AS sub_counter2 FROM base JOIN sub ON " "base.id = sub.id WHERE base.id = :param_1", - lambda ctx:{u'param_1': s1.id} + lambda ctx:{'param_1': s1.id} ), ) @@ -1910,7 +1910,7 @@ class OptimizedLoadTest(fixtures.MappedTest): "SELECT subsub.counter2 AS subsub_counter2, " "sub.counter2 AS sub_counter2 FROM subsub, sub " "WHERE :param_1 = sub.id AND sub.id = subsub.id", - lambda ctx:{u'param_1': s1.id} + lambda ctx:{'param_1': s1.id} ), ) @@ -2409,9 +2409,9 @@ class NameConflictTest(fixtures.MappedTest): polymorphic_identity='foo') sess = create_session() f = Foo() - f.content_type = u'bar' + f.content_type = 'bar' sess.add(f) sess.flush() f_id = f.id sess.expunge_all() - assert sess.query(Content).get(f_id).content_type == u'bar' + assert sess.query(Content).get(f_id).content_type == 'bar' diff --git a/test/orm/inheritance/test_concrete.py b/test/orm/inheritance/test_concrete.py index 40c3ab31b..573913f74 100644 --- a/test/orm/inheritance/test_concrete.py +++ b/test/orm/inheritance/test_concrete.py @@ -711,21 +711,21 @@ class ColKeysTest(fixtures.MappedTest): refugees_table = Table('refugee', metadata, Column('refugee_fid' , Integer, primary_key=True, test_needs_autoincrement=True), - Column('refugee_name', Unicode(30), + Column('refugee_name', String(30), key='name')) offices_table = Table('office', metadata, Column('office_fid', Integer, primary_key=True, test_needs_autoincrement=True), - Column('office_name', Unicode(30), + Column('office_name', String(30), key='name')) @classmethod def insert_data(cls): refugees_table.insert().execute(dict(refugee_fid=1, - name=u'refugee1'), dict(refugee_fid=2, name=u'refugee2' + name='refugee1'), dict(refugee_fid=2, name='refugee2' )) offices_table.insert().execute(dict(office_fid=1, - name=u'office1'), dict(office_fid=2, name=u'office2')) + name='office1'), dict(office_fid=2, name='office2')) def test_keys(self): pjoin = polymorphic_union({'refugee': refugees_table, 'office' diff --git a/test/orm/inheritance/test_magazine.py b/test/orm/inheritance/test_magazine.py index ae5aa8c8d..b459a01ae 100644 --- a/test/orm/inheritance/test_magazine.py +++ b/test/orm/inheritance/test_magazine.py @@ -8,7 +8,7 @@ from sqlalchemy.testing.schema import Table, Column class BaseObject(object): def __init__(self, *args, **kwargs): - for key, value in kwargs.iteritems(): + for key, value in kwargs.items(): setattr(self, key, value) class Publication(BaseObject): pass @@ -200,15 +200,15 @@ def _generate_round_trip_test(use_unions=False, use_joins=False): session.flush() - print [x for x in session] + print([x for x in session]) session.expunge_all() session.flush() session.expunge_all() p = session.query(Publication).filter(Publication.name=="Test").one() - print p.issues[0].locations[0].magazine.pages - print [page, page2, page3] + print(p.issues[0].locations[0].magazine.pages) + print([page, page2, page3]) assert repr(p.issues[0].locations[0].magazine.pages) == repr([page, page2, page3]), repr(p.issues[0].locations[0].magazine.pages) test_roundtrip = function_named( diff --git a/test/orm/inheritance/test_manytomany.py b/test/orm/inheritance/test_manytomany.py index e4df5d133..31c4ba40a 100644 --- a/test/orm/inheritance/test_manytomany.py +++ b/test/orm/inheritance/test_manytomany.py @@ -42,7 +42,7 @@ class InheritTest(fixtures.MappedTest): def testbasic(self): class Principal(object): def __init__(self, **kwargs): - for key, value in kwargs.iteritems(): + for key, value in kwargs.items(): setattr(self, key, value) class User(Principal): @@ -94,8 +94,8 @@ class InheritTest2(fixtures.MappedTest): mapper(Foo, foo) mapper(Bar, bar, inherits=Foo) - print foo.join(bar).primary_key - print class_mapper(Bar).primary_key + print(foo.join(bar).primary_key) + print(class_mapper(Bar).primary_key) b = Bar('somedata') sess = create_session() sess.add(b) @@ -133,8 +133,8 @@ class InheritTest2(fixtures.MappedTest): sess.expunge_all() l = sess.query(Bar).all() - print l[0] - print l[0].foos + print(l[0]) + print(l[0].foos) self.assert_unordered_result(l, Bar, # {'id':1, 'data':'barfoo', 'bid':1, 'foos':(Foo, [{'id':2,'data':'subfoo1'}, {'id':3,'data':'subfoo2'}])}, {'id':b.id, 'data':'barfoo', 'foos':(Foo, [{'id':f1.id,'data':'subfoo1'}, {'id':f2.id,'data':'subfoo2'}])}, @@ -197,7 +197,7 @@ class InheritTest3(fixtures.MappedTest): compare = [repr(b)] + sorted([repr(o) for o in b.foos]) sess.expunge_all() l = sess.query(Bar).all() - print repr(l[0]) + repr(l[0].foos) + print(repr(l[0]) + repr(l[0].foos)) found = [repr(l[0])] + sorted([repr(o) for o in l[0].foos]) eq_(found, compare) @@ -239,11 +239,11 @@ class InheritTest3(fixtures.MappedTest): sess.expunge_all() l = sess.query(Blub).all() - print l + print(l) self.assert_(repr(l[0]) == compare) sess.expunge_all() x = sess.query(Blub).filter_by(id=blubid).one() - print x + print(x) self.assert_(repr(x) == compare) diff --git a/test/orm/inheritance/test_polymorphic_rel.py b/test/orm/inheritance/test_polymorphic_rel.py index 1b9acb787..8c1f22114 100644 --- a/test/orm/inheritance/test_polymorphic_rel.py +++ b/test/orm/inheritance/test_polymorphic_rel.py @@ -7,7 +7,7 @@ from sqlalchemy import exc as sa_exc from sqlalchemy import testing from sqlalchemy.testing import assert_raises, eq_ -from _poly_fixtures import Company, Person, Engineer, Manager, Boss, \ +from ._poly_fixtures import Company, Person, Engineer, Manager, Boss, \ Machine, Paperwork, _Polymorphic,\ _PolymorphicPolymorphic, _PolymorphicUnions, _PolymorphicJoins,\ _PolymorphicAliasedJoins @@ -118,24 +118,24 @@ class _PolymorphicTestBase(object): eq_(q.count(), 1) eq_(q.all(), [ ( - Company(company_id=1, name=u'MegaCorp, Inc.'), + Company(company_id=1, name='MegaCorp, Inc.'), Engineer( - status=u'regular engineer', - engineer_name=u'dilbert', - name=u'dilbert', + status='regular engineer', + engineer_name='dilbert', + name='dilbert', company_id=1, - primary_language=u'java', + primary_language='java', person_id=1, - type=u'engineer'), - Company(company_id=1, name=u'MegaCorp, Inc.'), + type='engineer'), + Company(company_id=1, name='MegaCorp, Inc.'), Engineer( - status=u'regular engineer', - engineer_name=u'wally', - name=u'wally', + status='regular engineer', + engineer_name='wally', + name='wally', company_id=1, - primary_language=u'c++', + primary_language='c++', person_id=2, - type=u'engineer') + type='engineer') ) ]) @@ -337,8 +337,8 @@ class _PolymorphicTestBase(object): sess = create_session() expected = [ - (u'dogbert',), - (u'pointy haired boss',)] + ('dogbert',), + ('pointy haired boss',)] eq_(sess.query(Manager.name) .order_by(Manager.name).all(), expected) @@ -346,9 +346,9 @@ class _PolymorphicTestBase(object): def test_join_from_columns_or_subclass_two(self): sess = create_session() expected = [ - (u'dogbert',), - (u'dogbert',), - (u'pointy haired boss',)] + ('dogbert',), + ('dogbert',), + ('pointy haired boss',)] eq_(sess.query(Manager.name) .join(Paperwork, Manager.paperwork) .order_by(Manager.name).all(), @@ -357,14 +357,14 @@ class _PolymorphicTestBase(object): def test_join_from_columns_or_subclass_three(self): sess = create_session() expected = [ - (u'dilbert',), - (u'dilbert',), - (u'dogbert',), - (u'dogbert',), - (u'pointy haired boss',), - (u'vlad',), - (u'wally',), - (u'wally',)] + ('dilbert',), + ('dilbert',), + ('dogbert',), + ('dogbert',), + ('pointy haired boss',), + ('vlad',), + ('wally',), + ('wally',)] eq_(sess.query(Person.name) .join(Paperwork, Person.paperwork) .order_by(Person.name).all(), @@ -375,14 +375,14 @@ class _PolymorphicTestBase(object): # Load Person.name, joining from Person -> paperwork, get all # the people. expected = [ - (u'dilbert',), - (u'dilbert',), - (u'dogbert',), - (u'dogbert',), - (u'pointy haired boss',), - (u'vlad',), - (u'wally',), - (u'wally',)] + ('dilbert',), + ('dilbert',), + ('dogbert',), + ('dogbert',), + ('pointy haired boss',), + ('vlad',), + ('wally',), + ('wally',)] eq_(sess.query(Person.name) .join(paperwork, Person.person_id == paperwork.c.person_id) @@ -393,9 +393,9 @@ class _PolymorphicTestBase(object): sess = create_session() # same, on manager. get only managers. expected = [ - (u'dogbert',), - (u'dogbert',), - (u'pointy haired boss',)] + ('dogbert',), + ('dogbert',), + ('pointy haired boss',)] eq_(sess.query(Manager.name) .join(paperwork, Manager.person_id == paperwork.c.person_id) @@ -418,14 +418,14 @@ class _PolymorphicTestBase(object): # here, it joins to the full result set. This is 0.6's # behavior and is more or less wrong. expected = [ - (u'dilbert',), - (u'dilbert',), - (u'dogbert',), - (u'dogbert',), - (u'pointy haired boss',), - (u'vlad',), - (u'wally',), - (u'wally',)] + ('dilbert',), + ('dilbert',), + ('dogbert',), + ('dogbert',), + ('pointy haired boss',), + ('vlad',), + ('wally',), + ('wally',)] eq_(sess.query(Person.name) .join(paperwork, Manager.person_id == paperwork.c.person_id) @@ -435,9 +435,9 @@ class _PolymorphicTestBase(object): # when a join is present and managers.person_id is available, # you get the managers. expected = [ - (u'dogbert',), - (u'dogbert',), - (u'pointy haired boss',)] + ('dogbert',), + ('dogbert',), + ('pointy haired boss',)] eq_(sess.query(Person.name) .join(paperwork, Manager.person_id == paperwork.c.person_id) @@ -454,9 +454,9 @@ class _PolymorphicTestBase(object): def test_join_from_columns_or_subclass_eight(self): sess = create_session() expected = [ - (u'dogbert',), - (u'dogbert',), - (u'pointy haired boss',)] + ('dogbert',), + ('dogbert',), + ('pointy haired boss',)] eq_(sess.query(Manager.name) .join(paperwork, Manager.person_id == paperwork.c.person_id) @@ -474,9 +474,9 @@ class _PolymorphicTestBase(object): def test_join_from_columns_or_subclass_ten(self): sess = create_session() expected = [ - (u'pointy haired boss', u'review #1'), - (u'dogbert', u'review #2'), - (u'dogbert', u'review #3')] + ('pointy haired boss', 'review #1'), + ('dogbert', 'review #2'), + ('dogbert', 'review #3')] eq_(sess.query(Manager.name, Paperwork.description) .join(Paperwork, Manager.person_id == Paperwork.person_id) @@ -486,9 +486,9 @@ class _PolymorphicTestBase(object): def test_join_from_columns_or_subclass_eleven(self): sess = create_session() expected = [ - (u'pointy haired boss',), - (u'dogbert',), - (u'dogbert',)] + ('pointy haired boss',), + ('dogbert',), + ('dogbert',)] malias = aliased(Manager) eq_(sess.query(malias.name) .join(paperwork, @@ -977,32 +977,32 @@ class _PolymorphicTestBase(object): expected = [ (Engineer( - status=u'regular engineer', - engineer_name=u'dilbert', - name=u'dilbert', + status='regular engineer', + engineer_name='dilbert', + name='dilbert', company_id=1, - primary_language=u'java', + primary_language='java', person_id=1, - type=u'engineer'), - u'MegaCorp, Inc.'), + type='engineer'), + 'MegaCorp, Inc.'), (Engineer( - status=u'regular engineer', - engineer_name=u'wally', - name=u'wally', + status='regular engineer', + engineer_name='wally', + name='wally', company_id=1, - primary_language=u'c++', + primary_language='c++', person_id=2, - type=u'engineer'), - u'MegaCorp, Inc.'), + type='engineer'), + 'MegaCorp, Inc.'), (Engineer( - status=u'elbonian engineer', - engineer_name=u'vlad', - name=u'vlad', + status='elbonian engineer', + engineer_name='vlad', + name='vlad', company_id=2, - primary_language=u'cobol', + primary_language='cobol', person_id=5, - type=u'engineer'), - u'Elbonia, Inc.')] + type='engineer'), + 'Elbonia, Inc.')] eq_(sess.query(Engineer, Company.name) .join(Company.employees) .filter(Person.type == 'engineer').all(), @@ -1011,9 +1011,9 @@ class _PolymorphicTestBase(object): def test_mixed_entities_two(self): sess = create_session() expected = [ - (u'java', u'MegaCorp, Inc.'), - (u'cobol', u'Elbonia, Inc.'), - (u'c++', u'MegaCorp, Inc.')] + ('java', 'MegaCorp, Inc.'), + ('cobol', 'Elbonia, Inc.'), + ('c++', 'MegaCorp, Inc.')] eq_(sess.query(Engineer.primary_language, Company.name) .join(Company.employees) .filter(Person.type == 'engineer') @@ -1025,19 +1025,19 @@ class _PolymorphicTestBase(object): palias = aliased(Person) expected = [( Engineer( - status=u'elbonian engineer', - engineer_name=u'vlad', - name=u'vlad', - primary_language=u'cobol'), - u'Elbonia, Inc.', + status='elbonian engineer', + engineer_name='vlad', + name='vlad', + primary_language='cobol'), + 'Elbonia, Inc.', Engineer( - status=u'regular engineer', - engineer_name=u'dilbert', - name=u'dilbert', + status='regular engineer', + engineer_name='dilbert', + name='dilbert', company_id=1, - primary_language=u'java', + primary_language='java', person_id=1, - type=u'engineer'))] + type='engineer'))] eq_(sess.query(Person, Company.name, palias) .join(Company.employees) .filter(Company.name == 'Elbonia, Inc.') @@ -1049,19 +1049,19 @@ class _PolymorphicTestBase(object): palias = aliased(Person) expected = [( Engineer( - status=u'regular engineer', - engineer_name=u'dilbert', - name=u'dilbert', + status='regular engineer', + engineer_name='dilbert', + name='dilbert', company_id=1, - primary_language=u'java', + primary_language='java', person_id=1, - type=u'engineer'), - u'Elbonia, Inc.', + type='engineer'), + 'Elbonia, Inc.', Engineer( - status=u'elbonian engineer', - engineer_name=u'vlad', - name=u'vlad', - primary_language=u'cobol'),)] + status='elbonian engineer', + engineer_name='vlad', + name='vlad', + primary_language='cobol'),)] eq_(sess.query(palias, Company.name, Person) .join(Company.employees) .filter(Company.name == 'Elbonia, Inc.') @@ -1071,7 +1071,7 @@ class _PolymorphicTestBase(object): def test_mixed_entities_five(self): sess = create_session() palias = aliased(Person) - expected = [(u'vlad', u'Elbonia, Inc.', u'dilbert')] + expected = [('vlad', 'Elbonia, Inc.', 'dilbert')] eq_(sess.query(Person.name, Company.name, palias.name) .join(Company.employees) .filter(Company.name == 'Elbonia, Inc.') @@ -1082,9 +1082,9 @@ class _PolymorphicTestBase(object): sess = create_session() palias = aliased(Person) expected = [ - (u'manager', u'dogbert', u'engineer', u'dilbert'), - (u'manager', u'dogbert', u'engineer', u'wally'), - (u'manager', u'dogbert', u'boss', u'pointy haired boss')] + ('manager', 'dogbert', 'engineer', 'dilbert'), + ('manager', 'dogbert', 'engineer', 'wally'), + ('manager', 'dogbert', 'boss', 'pointy haired boss')] eq_(sess.query(Person.type, Person.name, palias.type, palias.name) .filter(Person.company_id == palias.company_id) .filter(Person.name == 'dogbert') @@ -1095,14 +1095,14 @@ class _PolymorphicTestBase(object): def test_mixed_entities_seven(self): sess = create_session() expected = [ - (u'dilbert', u'tps report #1'), - (u'dilbert', u'tps report #2'), - (u'dogbert', u'review #2'), - (u'dogbert', u'review #3'), - (u'pointy haired boss', u'review #1'), - (u'vlad', u'elbonian missive #3'), - (u'wally', u'tps report #3'), - (u'wally', u'tps report #4')] + ('dilbert', 'tps report #1'), + ('dilbert', 'tps report #2'), + ('dogbert', 'review #2'), + ('dogbert', 'review #3'), + ('pointy haired boss', 'review #1'), + ('vlad', 'elbonian missive #3'), + ('wally', 'tps report #3'), + ('wally', 'tps report #4')] eq_(sess.query(Person.name, Paperwork.description) .filter(Person.person_id == Paperwork.person_id) .order_by(Person.name, Paperwork.description).all(), @@ -1116,7 +1116,7 @@ class _PolymorphicTestBase(object): def test_mixed_entities_nine(self): sess = create_session() - expected = [(u'Elbonia, Inc.', 1), (u'MegaCorp, Inc.', 4)] + expected = [('Elbonia, Inc.', 1), ('MegaCorp, Inc.', 4)] eq_(sess.query(Company.name, func.count(Person.person_id)) .filter(Company.company_id == Person.company_id) .group_by(Company.name) @@ -1125,7 +1125,7 @@ class _PolymorphicTestBase(object): def test_mixed_entities_ten(self): sess = create_session() - expected = [(u'Elbonia, Inc.', 1), (u'MegaCorp, Inc.', 4)] + expected = [('Elbonia, Inc.', 1), ('MegaCorp, Inc.', 4)] eq_(sess.query(Company.name, func.count(Person.person_id)) .join(Company.employees) .group_by(Company.name) @@ -1153,14 +1153,14 @@ class _PolymorphicTestBase(object): def test_mixed_entities_eleven(self): sess = create_session() - expected = [(u'java',), (u'c++',), (u'cobol',)] + expected = [('java',), ('c++',), ('cobol',)] eq_(sess.query(Engineer.primary_language) .filter(Person.type == 'engineer').all(), expected) def test_mixed_entities_twelve(self): sess = create_session() - expected = [(u'vlad', u'Elbonia, Inc.')] + expected = [('vlad', 'Elbonia, Inc.')] eq_(sess.query(Person.name, Company.name) .join(Company.employees) .filter(Company.name == 'Elbonia, Inc.').all(), @@ -1168,15 +1168,15 @@ class _PolymorphicTestBase(object): def test_mixed_entities_thirteen(self): sess = create_session() - expected = [(u'pointy haired boss', u'fore')] + expected = [('pointy haired boss', 'fore')] eq_(sess.query(Boss.name, Boss.golf_swing).all(), expected) def test_mixed_entities_fourteen(self): sess = create_session() expected = [ - (u'dilbert', u'java'), - (u'wally', u'c++'), - (u'vlad', u'cobol')] + ('dilbert', 'java'), + ('wally', 'c++'), + ('vlad', 'cobol')] eq_(sess.query(Engineer.name, Engineer.primary_language).all(), expected) @@ -1184,12 +1184,12 @@ class _PolymorphicTestBase(object): sess = create_session() expected = [( - u'Elbonia, Inc.', + 'Elbonia, Inc.', Engineer( - status=u'elbonian engineer', - engineer_name=u'vlad', - name=u'vlad', - primary_language=u'cobol'))] + status='elbonian engineer', + engineer_name='vlad', + name='vlad', + primary_language='cobol'))] eq_(sess.query(Company.name, Person) .join(Company.employees) .filter(Company.name == 'Elbonia, Inc.').all(), @@ -1199,11 +1199,11 @@ class _PolymorphicTestBase(object): sess = create_session() expected = [( Engineer( - status=u'elbonian engineer', - engineer_name=u'vlad', - name=u'vlad', - primary_language=u'cobol'), - u'Elbonia, Inc.')] + status='elbonian engineer', + engineer_name='vlad', + name='vlad', + primary_language='cobol'), + 'Elbonia, Inc.')] eq_(sess.query(Person, Company.name) .join(Company.employees) .filter(Company.name == 'Elbonia, Inc.').all(), diff --git a/test/orm/inheritance/test_productspec.py b/test/orm/inheritance/test_productspec.py index c1e99e338..3de5e3695 100644 --- a/test/orm/inheritance/test_productspec.py +++ b/test/orm/inheritance/test_productspec.py @@ -122,8 +122,8 @@ class InheritTest(fixtures.MappedTest): a1 = session.query(Product).filter_by(name='a1').one() new = repr(a1) - print orig - print new + print(orig) + print(new) assert orig == new == '<Assembly a1> specification=[<SpecLine 1.0 <Product p1>>, <SpecLine 1.0 <Detail d1>>] documents=None' def testtwo(self): @@ -153,8 +153,8 @@ class InheritTest(fixtures.MappedTest): session.flush() session.expunge_all() new = repr(session.query(SpecLine).all()) - print orig - print new + print(orig) + print(new) assert orig == new == '[<SpecLine 1.0 <Product p1>>, <SpecLine 1.0 <Detail d1>>]' def testthree(self): @@ -206,8 +206,8 @@ class InheritTest(fixtures.MappedTest): a1 = session.query(Product).filter_by(name='a1').one() new = repr(a1) - print orig - print new + print(orig) + print(new) assert orig == new == '<Assembly a1> specification=[<SpecLine 1.0 <Detail d1>>] documents=[<Document doc1>, <RasterDocument doc2>]' def testfour(self): @@ -245,8 +245,8 @@ class InheritTest(fixtures.MappedTest): a1 = session.query(Product).filter_by(name='a1').one() new = repr(a1) - print orig - print new + print(orig) + print(new) assert orig == new == '<Assembly a1> specification=None documents=[<RasterDocument doc2>]' del a1.documents[0] @@ -312,7 +312,7 @@ class InheritTest(fixtures.MappedTest): a1 = session.query(Product).filter_by(name='a1').one() new = repr(a1) - print orig - print new + print(orig) + print(new) assert orig == new == '<Assembly a1> specification=[<SpecLine 1.0 <Detail d1>>] documents=[<Document doc1>, <RasterDocument doc2>]' diff --git a/test/orm/inheritance/test_relationship.py b/test/orm/inheritance/test_relationship.py index 36dbb7d27..809884f52 100644 --- a/test/orm/inheritance/test_relationship.py +++ b/test/orm/inheritance/test_relationship.py @@ -287,7 +287,7 @@ class SelfReferentialJ2JSelfTest(fixtures.MappedTest): def _five_obj_fixture(self): sess = Session() e1, e2, e3, e4, e5 = [ - Engineer(name='e%d' % (i + 1)) for i in xrange(5) + Engineer(name='e%d' % (i + 1)) for i in range(5) ] e3.reports_to = e1 e4.reports_to = e2 diff --git a/test/orm/inheritance/test_with_poly.py b/test/orm/inheritance/test_with_poly.py index 93cccee7a..87251a4c2 100644 --- a/test/orm/inheritance/test_with_poly.py +++ b/test/orm/inheritance/test_with_poly.py @@ -11,7 +11,7 @@ from sqlalchemy import testing from sqlalchemy.testing.schema import Table, Column from sqlalchemy.testing import assert_raises, eq_ -from _poly_fixtures import Company, Person, Engineer, Manager, Boss, \ +from ._poly_fixtures import Company, Person, Engineer, Manager, Boss, \ Machine, Paperwork, _PolymorphicFixtureBase, _Polymorphic,\ _PolymorphicPolymorphic, _PolymorphicUnions, _PolymorphicJoins,\ _PolymorphicAliasedJoins @@ -37,8 +37,8 @@ class _WithPolymorphicBase(_PolymorphicFixtureBase): pa.Manager.manager_name=='dogbert')).\ order_by(pa.Engineer.type).all(), [ - (u'dilbert', u'java', None), - (u'dogbert', None, u'dogbert'), + ('dilbert', 'java', None), + ('dogbert', None, 'dogbert'), ] ) @@ -63,10 +63,10 @@ class _WithPolymorphicBase(_PolymorphicFixtureBase): ) ).order_by(pa.name, pa_alias.name)], [ - (u'dilbert', Engineer, u'dilbert', Engineer), - (u'dogbert', Manager, u'pointy haired boss', Boss), - (u'vlad', Engineer, u'vlad', Engineer), - (u'wally', Engineer, u'wally', Engineer) + ('dilbert', Engineer, 'dilbert', Engineer), + ('dogbert', Manager, 'pointy haired boss', Boss), + ('vlad', Engineer, 'vlad', Engineer), + ('wally', Engineer, 'wally', Engineer) ] ) @@ -91,10 +91,10 @@ class _WithPolymorphicBase(_PolymorphicFixtureBase): ) ).order_by(pa.name, pa_alias.name)], [ - (u'dilbert', u'java', u'dilbert', u'java'), - (u'dogbert', None, u'pointy haired boss', None), - (u'vlad', u'cobol', u'vlad', u'cobol'), - (u'wally', u'c++', u'wally', u'c++') + ('dilbert', 'java', 'dilbert', 'java'), + ('dogbert', None, 'pointy haired boss', None), + ('vlad', 'cobol', 'vlad', 'cobol'), + ('wally', 'c++', 'wally', 'c++') ] ) diff --git a/test/orm/test_assorted_eager.py b/test/orm/test_assorted_eager.py index a1c96bdfa..83fccbf7a 100644 --- a/test/orm/test_assorted_eager.py +++ b/test/orm/test_assorted_eager.py @@ -129,7 +129,7 @@ class EagerTest(fixtures.MappedTest): # 3 " # not orm style correct query - print "Obtaining correct results without orm" + print("Obtaining correct results without orm") result = sa.select( [tests.c.id,categories.c.name], sa.and_(tests.c.owner_id == 1, @@ -140,7 +140,7 @@ class EagerTest(fixtures.MappedTest): tests.c.id == options.c.test_id, tests.c.owner_id == options.c.owner_id))] ).execute().fetchall() - eq_(result, [(1, u'Some Category'), (3, u'Some Category')]) + eq_(result, [(1, 'Some Category'), (3, 'Some Category')]) def test_withoutjoinedload(self): Thing, tests, options = (self.classes.Thing, @@ -158,7 +158,7 @@ class EagerTest(fixtures.MappedTest): options.c.someoption==False)))) result = ["%d %s" % ( t.id,t.category.name ) for t in l] - eq_(result, [u'1 Some Category', u'3 Some Category']) + eq_(result, ['1 Some Category', '3 Some Category']) def test_withjoinedload(self): """ @@ -185,7 +185,7 @@ class EagerTest(fixtures.MappedTest): options.c.someoption==False)))) result = ["%d %s" % ( t.id,t.category.name ) for t in l] - eq_(result, [u'1 Some Category', u'3 Some Category']) + eq_(result, ['1 Some Category', '3 Some Category']) def test_dslish(self): """test the same as withjoinedload except using generative""" @@ -203,7 +203,7 @@ class EagerTest(fixtures.MappedTest): ).outerjoin('owner_option') result = ["%d %s" % ( t.id,t.category.name ) for t in l] - eq_(result, [u'1 Some Category', u'3 Some Category']) + eq_(result, ['1 Some Category', '3 Some Category']) @testing.crashes('sybase', 'FIXME: unknown, verify not fails_on') def test_without_outerjoin_literal(self): @@ -219,7 +219,7 @@ class EagerTest(fixtures.MappedTest): join('owner_option')) result = ["%d %s" % ( t.id,t.category.name ) for t in l] - eq_(result, [u'3 Some Category']) + eq_(result, ['3 Some Category']) def test_withoutouterjoin(self): Thing, tests, options = (self.classes.Thing, @@ -234,7 +234,7 @@ class EagerTest(fixtures.MappedTest): ).join('owner_option') result = ["%d %s" % ( t.id,t.category.name ) for t in l] - eq_(result, [u'3 Some Category']) + eq_(result, ['3 Some Category']) class EagerTest2(fixtures.MappedTest): @@ -723,13 +723,13 @@ class EagerTest8(fixtures.MappedTest): Table('prj', metadata, Column('id', Integer, primary_key=True, test_needs_autoincrement=True), Column('created', sa.DateTime ), - Column('title', sa.Unicode(100))) + Column('title', sa.String(100))) Table('task', metadata, Column('id', Integer, primary_key=True, test_needs_autoincrement=True), Column('status_id', Integer, ForeignKey('task_status.id'), nullable=False), - Column('title', sa.Unicode(100)), + Column('title', sa.String(100)), Column('task_type_id', Integer , ForeignKey('task_type.id'), nullable=False), Column('prj_id', Integer , ForeignKey('prj.id'), nullable=False)) @@ -748,8 +748,8 @@ class EagerTest8(fixtures.MappedTest): Table('msg_type', metadata, Column('id', Integer, primary_key=True, test_needs_autoincrement=True), - Column('name', sa.Unicode(20)), - Column('display_name', sa.Unicode(20))) + Column('name', sa.String(20)), + Column('display_name', sa.String(20))) @classmethod def fixtures(cls): @@ -764,7 +764,7 @@ class EagerTest8(fixtures.MappedTest): (1,),), task=(('title', 'task_type_id', 'status_id', 'prj_id'), - (u'task 1', 1, 1, 1))) + ('task 1', 1, 1, 1))) @classmethod def setup_classes(cls): @@ -804,7 +804,7 @@ class EagerTest8(fixtures.MappedTest): session = create_session() eq_(session.query(Joined).limit(10).offset(0).one(), - Joined(id=1, title=u'task 1', props_cnt=0)) + Joined(id=1, title='task 1', props_cnt=0)) class EagerTest9(fixtures.MappedTest): diff --git a/test/orm/test_attributes.py b/test/orm/test_attributes.py index d60c55edd..2b5d14695 100644 --- a/test/orm/test_attributes.py +++ b/test/orm/test_attributes.py @@ -838,7 +838,7 @@ class AttributesTest(fixtures.ORMTest): attributes.register_attribute(Foo, 'collection', uselist=True, typecallable=dict, useobject=True) assert False - except sa_exc.ArgumentError, e: + except sa_exc.ArgumentError as e: assert str(e) \ == 'Type InstrumentedDict must elect an appender '\ 'method to be a collection class' @@ -865,7 +865,7 @@ class AttributesTest(fixtures.ORMTest): attributes.register_attribute(Foo, 'collection', uselist=True, typecallable=MyColl, useobject=True) assert False - except sa_exc.ArgumentError, e: + except sa_exc.ArgumentError as e: assert str(e) \ == 'Type MyColl must elect an appender method to be a '\ 'collection class' @@ -889,7 +889,7 @@ class AttributesTest(fixtures.ORMTest): try: Foo().collection assert True - except sa_exc.ArgumentError, e: + except sa_exc.ArgumentError as e: assert False class GetNoValueTest(fixtures.ORMTest): @@ -1410,7 +1410,7 @@ class HistoryTest(fixtures.TestBase): class Foo(fixtures.BasicEntity): pass class Bar(fixtures.BasicEntity): - def __nonzero__(self): + def __bool__(self): assert False instrumentation.register_class(Foo) diff --git a/test/orm/test_collection.py b/test/orm/test_collection.py index f6cf51037..c9f9f6951 100644 --- a/test/orm/test_collection.py +++ b/test/orm/test_collection.py @@ -1106,7 +1106,7 @@ class CollectionsTest(fixtures.ORMTest): @collection.converter def _convert(self, dictlike): - for key, value in dictlike.iteritems(): + for key, value in dictlike.items(): yield value + 5 class Foo(object): @@ -1145,12 +1145,12 @@ class CollectionsTest(fixtures.ORMTest): def __delitem__(self, key): del self.data[key] def values(self): - return self.data.values() + return list(self.data.values()) def __contains__(self, key): return key in self.data @collection.iterator def itervalues(self): - return self.data.itervalues() + return iter(self.data.values()) __hash__ = object.__hash__ def __eq__(self, other): return self.data == other @@ -1158,7 +1158,7 @@ class CollectionsTest(fixtures.ORMTest): return 'DictLike(%s)' % repr(self.data) self._test_adapter(DictLike, self.dictable_entity, - to_set=lambda c: set(c.itervalues())) + to_set=lambda c: set(c.values())) self._test_dict(DictLike) self._test_dict_bulk(DictLike) self.assert_(getattr(DictLike, '_sa_instrumented') == id(DictLike)) @@ -1185,12 +1185,12 @@ class CollectionsTest(fixtures.ORMTest): def __delitem__(self, key): del self.data[key] def values(self): - return self.data.values() + return list(self.data.values()) def __contains__(self, key): return key in self.data @collection.iterator def itervalues(self): - return self.data.itervalues() + return iter(self.data.values()) __hash__ = object.__hash__ def __eq__(self, other): return self.data == other @@ -1198,7 +1198,7 @@ class CollectionsTest(fixtures.ORMTest): return 'DictIsh(%s)' % repr(self.data) self._test_adapter(DictIsh, self.dictable_entity, - to_set=lambda c: set(c.itervalues())) + to_set=lambda c: set(c.values())) self._test_dict(DictIsh) self._test_dict_bulk(DictIsh) self.assert_(getattr(DictIsh, '_sa_instrumented') == id(DictIsh)) @@ -1859,7 +1859,7 @@ class CustomCollectionsTest(fixtures.MappedTest): f = sess.query(Foo).get(f.col1) assert len(list(f.bars)) == 2 - existing = set([id(b) for b in f.bars.values()]) + existing = set([id(b) for b in list(f.bars.values())]) col = collections.collection_adapter(f.bars) col.append_with_event(Bar('b')) @@ -1869,7 +1869,7 @@ class CustomCollectionsTest(fixtures.MappedTest): f = sess.query(Foo).get(f.col1) assert len(list(f.bars)) == 2 - replaced = set([id(b) for b in f.bars.values()]) + replaced = set([id(b) for b in list(f.bars.values())]) self.assert_(existing != replaced) def test_list(self): diff --git a/test/orm/test_composites.py b/test/orm/test_composites.py index f9af0c702..b6e5d81a1 100644 --- a/test/orm/test_composites.py +++ b/test/orm/test_composites.py @@ -510,11 +510,11 @@ class MappedSelectTest(fixtures.MappedTest): session.commit() eq_( testing.db.execute(descriptions.select()).fetchall(), - [(1, u'Color', u'Number')] + [(1, 'Color', 'Number')] ) eq_( testing.db.execute(values.select()).fetchall(), - [(1, 1, u'Red', u'5'), (2, 1, u'Blue', u'1')] + [(1, 1, 'Red', '5'), (2, 1, 'Blue', '1')] ) class ManyToOneTest(fixtures.MappedTest): diff --git a/test/orm/test_deprecations.py b/test/orm/test_deprecations.py index 95fa9dd29..a8be34971 100644 --- a/test/orm/test_deprecations.py +++ b/test/orm/test_deprecations.py @@ -122,7 +122,7 @@ class QueryAlternativesTest(fixtures.MappedTest): session = sessionmaker(query_cls=MyQuery)() ad1 = session.query(Address).get(1) - assert ad1 in cache.values() + assert ad1 in list(cache.values()) def test_load(self): """x = session.query(Address).load(1) diff --git a/test/orm/test_dynamic.py b/test/orm/test_dynamic.py index c8a675e3b..21dcfd436 100644 --- a/test/orm/test_dynamic.py +++ b/test/orm/test_dynamic.py @@ -119,9 +119,9 @@ class DynamicTest(_DynamicFixture, _fixtures.FixtureTest, AssertsCompiledSQL): eq_( list(u.addresses.order_by(desc(Address.email_address))), [ - Address(email_address=u'ed@wood.com'), - Address(email_address=u'ed@lala.com'), - Address(email_address=u'ed@bettyboop.com') + Address(email_address='ed@wood.com'), + Address(email_address='ed@lala.com'), + Address(email_address='ed@bettyboop.com') ] ) @@ -137,9 +137,9 @@ class DynamicTest(_DynamicFixture, _fixtures.FixtureTest, AssertsCompiledSQL): eq_( list(u.addresses), [ - Address(email_address=u'ed@wood.com'), - Address(email_address=u'ed@lala.com'), - Address(email_address=u'ed@bettyboop.com') + Address(email_address='ed@wood.com'), + Address(email_address='ed@lala.com'), + Address(email_address='ed@bettyboop.com') ] ) @@ -147,9 +147,9 @@ class DynamicTest(_DynamicFixture, _fixtures.FixtureTest, AssertsCompiledSQL): eq_( list(u.addresses.order_by(None).order_by(Address.email_address)), [ - Address(email_address=u'ed@bettyboop.com'), - Address(email_address=u'ed@lala.com'), - Address(email_address=u'ed@wood.com') + Address(email_address='ed@bettyboop.com'), + Address(email_address='ed@lala.com'), + Address(email_address='ed@wood.com') ] ) @@ -157,9 +157,9 @@ class DynamicTest(_DynamicFixture, _fixtures.FixtureTest, AssertsCompiledSQL): eq_( set(u.addresses.order_by(None)), set([ - Address(email_address=u'ed@bettyboop.com'), - Address(email_address=u'ed@lala.com'), - Address(email_address=u'ed@wood.com') + Address(email_address='ed@bettyboop.com'), + Address(email_address='ed@lala.com'), + Address(email_address='ed@wood.com') ]) ) @@ -529,12 +529,12 @@ class UOWTest(_DynamicFixture, _fixtures.FixtureTest, "SELECT addresses.id AS addresses_id, addresses.email_address " "AS addresses_email_address FROM addresses " "WHERE addresses.id = :param_1", - lambda ctx: [{u'param_1': a2_id}] + lambda ctx: [{'param_1': a2_id}] ), CompiledSQL( "UPDATE addresses SET user_id=:user_id WHERE addresses.id = " ":addresses_id", - lambda ctx: [{u'addresses_id': a2_id, 'user_id': None}] + lambda ctx: [{'addresses_id': a2_id, 'user_id': None}] ) ) diff --git a/test/orm/test_eager_relations.py b/test/orm/test_eager_relations.py index b240d29f6..54a106c5a 100644 --- a/test/orm/test_eager_relations.py +++ b/test/orm/test_eager_relations.py @@ -747,11 +747,11 @@ class EagerTest(_fixtures.FixtureTest, testing.AssertsCompiledSQL): sess = create_session() eq_(sess.query(User).first(), - User(name=u'jack',orders=[ - Order(address_id=1,description=u'order 1',isopen=0,user_id=7,id=1), - Order(address_id=1,description=u'order 3',isopen=1,user_id=7,id=3), - Order(address_id=None,description=u'order 5',isopen=0,user_id=7,id=5)], - email_address=u'jack@bean.com',id=7) + User(name='jack',orders=[ + Order(address_id=1,description='order 1',isopen=0,user_id=7,id=1), + Order(address_id=1,description='order 3',isopen=1,user_id=7,id=3), + Order(address_id=None,description='order 5',isopen=0,user_id=7,id=5)], + email_address='jack@bean.com',id=7) ) def test_useget_cancels_eager(self): @@ -2103,7 +2103,7 @@ class SelfReferentialM2MEagerTest(fixtures.MappedTest): def define_tables(cls, metadata): Table('widget', metadata, Column('id', Integer, primary_key=True, test_needs_autoincrement=True), - Column('name', sa.Unicode(40), nullable=False, unique=True), + Column('name', sa.String(40), nullable=False, unique=True), ) Table('widget_rel', metadata, @@ -2127,15 +2127,15 @@ class SelfReferentialM2MEagerTest(fixtures.MappedTest): }) sess = create_session() - w1 = Widget(name=u'w1') - w2 = Widget(name=u'w2') + w1 = Widget(name='w1') + w2 = Widget(name='w2') w1.children.append(w2) sess.add(w1) sess.flush() sess.expunge_all() eq_([Widget(name='w1', children=[Widget(name='w2')])], - sess.query(Widget).filter(Widget.name==u'w1').all()) + sess.query(Widget).filter(Widget.name=='w1').all()) class MixedEntitiesTest(_fixtures.FixtureTest, testing.AssertsCompiledSQL): run_setup_mappers = 'once' @@ -2219,24 +2219,24 @@ class MixedEntitiesTest(_fixtures.FixtureTest, testing.AssertsCompiledSQL): eq_( [ ( - User(addresses=[Address(email_address=u'fred@fred.com')], name=u'fred'), - Order(description=u'order 2', isopen=0, items=[Item(description=u'item 1'), Item(description=u'item 2'), Item(description=u'item 3')]), - User(addresses=[Address(email_address=u'jack@bean.com')], name=u'jack'), - Order(description=u'order 3', isopen=1, items=[Item(description=u'item 3'), Item(description=u'item 4'), Item(description=u'item 5')]) + User(addresses=[Address(email_address='fred@fred.com')], name='fred'), + Order(description='order 2', isopen=0, items=[Item(description='item 1'), Item(description='item 2'), Item(description='item 3')]), + User(addresses=[Address(email_address='jack@bean.com')], name='jack'), + Order(description='order 3', isopen=1, items=[Item(description='item 3'), Item(description='item 4'), Item(description='item 5')]) ), ( - User(addresses=[Address(email_address=u'fred@fred.com')], name=u'fred'), - Order(description=u'order 2', isopen=0, items=[Item(description=u'item 1'), Item(description=u'item 2'), Item(description=u'item 3')]), - User(addresses=[Address(email_address=u'jack@bean.com')], name=u'jack'), - Order(address_id=None, description=u'order 5', isopen=0, items=[Item(description=u'item 5')]) + User(addresses=[Address(email_address='fred@fred.com')], name='fred'), + Order(description='order 2', isopen=0, items=[Item(description='item 1'), Item(description='item 2'), Item(description='item 3')]), + User(addresses=[Address(email_address='jack@bean.com')], name='jack'), + Order(address_id=None, description='order 5', isopen=0, items=[Item(description='item 5')]) ), ( - User(addresses=[Address(email_address=u'fred@fred.com')], name=u'fred'), - Order(description=u'order 4', isopen=1, items=[Item(description=u'item 1'), Item(description=u'item 5')]), - User(addresses=[Address(email_address=u'jack@bean.com')], name=u'jack'), - Order(address_id=None, description=u'order 5', isopen=0, items=[Item(description=u'item 5')]) + User(addresses=[Address(email_address='fred@fred.com')], name='fred'), + Order(description='order 4', isopen=1, items=[Item(description='item 1'), Item(description='item 5')]), + User(addresses=[Address(email_address='jack@bean.com')], name='jack'), + Order(address_id=None, description='order 5', isopen=0, items=[Item(description='item 5')]) ), ], sess.query(User, Order, u1, o1).\ @@ -2647,9 +2647,9 @@ class CyclicalInheritingEagerTestTwo(fixtures.DeclarativeMappedTest, Movie = self.classes.Movie session = Session(testing.db) - rscott = Director(name=u"Ridley Scott") - alien = Movie(title=u"Alien") - brunner = Movie(title=u"Blade Runner") + rscott = Director(name="Ridley Scott") + alien = Movie(title="Alien") + brunner = Movie(title="Blade Runner") rscott.movies.append(brunner) rscott.movies.append(alien) session.add_all([rscott, alien, brunner]) diff --git a/test/orm/test_evaluator.py b/test/orm/test_evaluator.py index 4678100eb..2570f7650 100644 --- a/test/orm/test_evaluator.py +++ b/test/orm/test_evaluator.py @@ -62,6 +62,23 @@ class EvaluateTest(fixtures.MappedTest): (User(name=None), True), ]) + def test_true_false(self): + User = self.classes.User + + eval_eq(User.name == False, testcases=[ + (User(name='foo'), False), + (User(name=True), False), + (User(name=False), True), + ] + ) + + eval_eq(User.name == True, testcases=[ + (User(name='foo'), False), + (User(name=True), True), + (User(name=False), False), + ] + ) + def test_boolean_ops(self): User = self.classes.User diff --git a/test/orm/test_expire.py b/test/orm/test_expire.py index 0b1350e0a..6eb124cd2 100644 --- a/test/orm/test_expire.py +++ b/test/orm/test_expire.py @@ -850,7 +850,7 @@ class ExpireTest(_fixtures.FixtureTest): assert len(u.addresses) == 3 sess.expire(u) assert 'addresses' not in u.__dict__ - print "-------------------------------------------" + print("-------------------------------------------") sess.query(User).filter_by(id=8).all() assert 'addresses' in u.__dict__ assert len(u.addresses) == 3 diff --git a/test/orm/test_froms.py b/test/orm/test_froms.py index c701a7076..1a972d965 100644 --- a/test/orm/test_froms.py +++ b/test/orm/test_froms.py @@ -161,7 +161,7 @@ class RawSelectTest(QueryTest, AssertsCompiledSQL): sess = create_session() - self.assert_compile(sess.query(users).select_from( + self.assert_compile(sess.query(users).select_entity_from( users.select()).with_labels().statement, "SELECT users.id AS users_id, users.name AS users_name FROM users, " "(SELECT users.id AS id, users.name AS name FROM users) AS anon_1", @@ -180,7 +180,7 @@ class RawSelectTest(QueryTest, AssertsCompiledSQL): filter(addresses.c.user_id == users.c.id).correlate(users).\ statement.alias() - self.assert_compile(sess.query(users, s.c.email).select_from( + self.assert_compile(sess.query(users, s.c.email).select_entity_from( users.join(s, s.c.id == users.c.id) ).with_labels().statement, "SELECT users.id AS users_id, users.name AS users_name, " @@ -315,12 +315,12 @@ class FromSelfTest(QueryTest, AssertsCompiledSQL): from_self(User.name, ualias.name). order_by(User.name, ualias.name).all(), [ - (u'chuck', u'ed'), - (u'chuck', u'fred'), - (u'chuck', u'jack'), - (u'ed', u'jack'), - (u'fred', u'ed'), - (u'fred', u'jack') + ('chuck', 'ed'), + ('chuck', 'fred'), + ('chuck', 'jack'), + ('ed', 'jack'), + ('fred', 'ed'), + ('fred', 'jack') ] ) @@ -330,7 +330,7 @@ class FromSelfTest(QueryTest, AssertsCompiledSQL): from_self(User.name, ualias.name). filter(ualias.name=='ed')\ .order_by(User.name, ualias.name).all(), - [(u'chuck', u'ed'), (u'fred', u'ed')] + [('chuck', 'ed'), ('fred', 'ed')] ) eq_( @@ -340,11 +340,11 @@ class FromSelfTest(QueryTest, AssertsCompiledSQL): join(ualias.addresses). order_by(ualias.name, Address.email_address).all(), [ - (u'ed', u'fred@fred.com'), - (u'jack', u'ed@bettyboop.com'), - (u'jack', u'ed@lala.com'), - (u'jack', u'ed@wood.com'), - (u'jack', u'fred@fred.com')] + ('ed', 'fred@fred.com'), + ('jack', 'ed@bettyboop.com'), + ('jack', 'ed@lala.com'), + ('jack', 'ed@wood.com'), + ('jack', 'fred@fred.com')] ) @@ -425,12 +425,12 @@ class ColumnAccessTest(QueryTest, AssertsCompiledSQL): "WHERE anon_1.anon_2_users_name = :name_1" ) - def test_select_from(self): + def test_select_entity_from(self): User = self.classes.User sess = create_session() q = sess.query(User) - q = sess.query(User).select_from(q.statement) + q = sess.query(User).select_entity_from(q.statement) self.assert_compile( q.filter(User.name=='ed'), "SELECT anon_1.id AS anon_1_id, anon_1.name AS anon_1_name " @@ -438,6 +438,31 @@ class ColumnAccessTest(QueryTest, AssertsCompiledSQL): "users) AS anon_1 WHERE anon_1.name = :name_1" ) + def test_select_entity_from_no_entities(self): + User = self.classes.User + sess = create_session() + + q = sess.query(User) + assert_raises_message( + sa.exc.ArgumentError, + r"A selectable \(FromClause\) instance is " + "expected when the base alias is being set", + sess.query(User).select_entity_from, User + ) + + def test_select_from_no_aliasing(self): + User = self.classes.User + sess = create_session() + + q = sess.query(User) + q = sess.query(User).select_from(q.statement) + self.assert_compile( + q.filter(User.name=='ed'), + "SELECT users.id AS users_id, users.name AS users_name " + "FROM users, (SELECT users.id AS id, users.name AS name FROM " + "users) AS anon_1 WHERE users.name = :name_1" + ) + def test_anonymous_expression(self): from sqlalchemy.sql import column @@ -629,8 +654,8 @@ class AddEntityEquivalenceTest(fixtures.MappedTest, AssertsCompiledSQL): eq_( q.all(), [( - A(bid=2, id=1, name=u'a1', type=u'a'), - C(age=3, id=2, name=u'c1', type=u'c') + A(bid=2, id=1, name='a1', type='a'), + C(age=3, id=2, name='c1', type='c') )] ) @@ -642,8 +667,8 @@ class AddEntityEquivalenceTest(fixtures.MappedTest, AssertsCompiledSQL): eq_( q.all(), [( - C(age=3, id=2, name=u'c1', type=u'c'), - A(bid=2, id=1, name=u'a1', type=u'a') + C(age=3, id=2, name='c1', type='c'), + A(bid=2, id=1, name='a1', type='a') )] ) @@ -681,15 +706,15 @@ class InstancesTest(QueryTest, AssertsCompiledSQL): assert self.static.user_address_result == l self.assert_sql_count(testing.db, go, 1) - # better way. use select_from() + # better way. use select_entity_from() def go(): - l = sess.query(User).select_from(query).\ + l = sess.query(User).select_entity_from(query).\ options(contains_eager('addresses')).all() assert self.static.user_address_result == l self.assert_sql_count(testing.db, go, 1) # same thing, but alias addresses, so that the adapter - # generated by select_from() is wrapped within + # generated by select_entity_from() is wrapped within # the adapter created by contains_eager() adalias = addresses.alias() query = users.select(users.c.id==7).\ @@ -699,7 +724,7 @@ class InstancesTest(QueryTest, AssertsCompiledSQL): select(use_labels=True, order_by=['ulist.id', adalias.c.id]) def go(): - l = sess.query(User).select_from(query).\ + l = sess.query(User).select_entity_from(query).\ options(contains_eager('addresses', alias=adalias)).all() assert self.static.user_address_result == l self.assert_sql_count(testing.db, go, 1) @@ -733,7 +758,7 @@ class InstancesTest(QueryTest, AssertsCompiledSQL): adalias = addresses.alias() q = sess.query(User).\ - select_from(users.outerjoin(adalias)).\ + select_entity_from(users.outerjoin(adalias)).\ options(contains_eager(User.addresses, alias=adalias)).\ order_by(User.id, adalias.c.id) def go(): @@ -929,12 +954,12 @@ class InstancesTest(QueryTest, AssertsCompiledSQL): # have a Dingaling here due to using the inner # join for the eager load [ - User(name=u'ed', addresses=[ - Address(email_address=u'ed@wood.com', + User(name='ed', addresses=[ + Address(email_address='ed@wood.com', dingaling=Dingaling(data='ding 1/2')), ]), - User(name=u'fred', addresses=[ - Address(email_address=u'fred@fred.com', + User(name='fred', addresses=[ + Address(email_address='fred@fred.com', dingaling=Dingaling(data='ding 2/5')) ]) ] @@ -965,12 +990,12 @@ class InstancesTest(QueryTest, AssertsCompiledSQL): # have a Dingaling here due to using the inner # join for the eager load [ - User(name=u'ed', addresses=[ - Address(email_address=u'ed@wood.com', + User(name='ed', addresses=[ + Address(email_address='ed@wood.com', dingaling=Dingaling(data='ding 1/2')), ]), - User(name=u'fred', addresses=[ - Address(email_address=u'fred@fred.com', + User(name='fred', addresses=[ + Address(email_address='fred@fred.com', dingaling=Dingaling(data='ding 2/5')) ]) ] @@ -998,11 +1023,11 @@ class InstancesTest(QueryTest, AssertsCompiledSQL): contains_eager(User.orders)).order_by(User.id, Order.id).offset(1).limit(2).all() eq_(l, [User(id=7, - addresses=[Address(email_address=u'jack@bean.com', - user_id=7, id=1)], name=u'jack', + addresses=[Address(email_address='jack@bean.com', + user_id=7, id=1)], name='jack', orders=[Order(address_id=1, user_id=7, - description=u'order 3', isopen=1, id=3), - Order(address_id=None, user_id=7, description=u'order 5' + description='order 3', isopen=1, id=3), + Order(address_id=None, user_id=7, description='order 5' , isopen=0, id=5)])]) self.assert_sql_count(testing.db, go, 1) @@ -1020,11 +1045,11 @@ class InstancesTest(QueryTest, AssertsCompiledSQL): order_by(User.id, oalias.id).\ offset(1).limit(2).all() eq_(l, [User(id=7, - addresses=[Address(email_address=u'jack@bean.com', - user_id=7, id=1)], name=u'jack', + addresses=[Address(email_address='jack@bean.com', + user_id=7, id=1)], name='jack', orders=[Order(address_id=1, user_id=7, - description=u'order 3', isopen=1, id=3), - Order(address_id=None, user_id=7, description=u'order 5' + description='order 3', isopen=1, id=3), + Order(address_id=None, user_id=7, description='order 5' , isopen=0, id=5)])]) self.assert_sql_count(testing.db, go, 1) @@ -1044,16 +1069,16 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): sel = users.select(User.id.in_([7, 8])).alias() q = sess.query(User) - q2 = q.select_from(sel).values(User.name) - eq_(list(q2), [(u'jack',), (u'ed',)]) + q2 = q.select_entity_from(sel).values(User.name) + eq_(list(q2), [('jack',), ('ed',)]) q = sess.query(User) q2 = q.order_by(User.id).\ values(User.name, User.name + " " + cast(User.id, String(50))) eq_( list(q2), - [(u'jack', u'jack 7'), (u'ed', u'ed 8'), - (u'fred', u'fred 9'), (u'chuck', u'chuck 10')] + [('jack', 'jack 7'), ('ed', 'ed 8'), + ('fred', 'fred 9'), ('chuck', 'chuck 10')] ) q2 = q.join('addresses').\ @@ -1061,39 +1086,39 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): order_by(User.id, Address.id).\ values(User.name, Address.email_address) eq_(list(q2), - [(u'ed', u'ed@wood.com'), (u'ed', u'ed@bettyboop.com'), - (u'ed', u'ed@lala.com'), (u'fred', u'fred@fred.com')]) + [('ed', 'ed@wood.com'), ('ed', 'ed@bettyboop.com'), + ('ed', 'ed@lala.com'), ('fred', 'fred@fred.com')]) q2 = q.join('addresses').\ filter(User.name.like('%e%')).\ order_by(desc(Address.email_address)).\ slice(1, 3).values(User.name, Address.email_address) - eq_(list(q2), [(u'ed', u'ed@wood.com'), (u'ed', u'ed@lala.com')]) + eq_(list(q2), [('ed', 'ed@wood.com'), ('ed', 'ed@lala.com')]) adalias = aliased(Address) q2 = q.join(adalias, 'addresses').\ filter(User.name.like('%e%')).order_by(adalias.email_address).\ values(User.name, adalias.email_address) - eq_(list(q2), [(u'ed', u'ed@bettyboop.com'), (u'ed', u'ed@lala.com'), - (u'ed', u'ed@wood.com'), (u'fred', u'fred@fred.com')]) + eq_(list(q2), [('ed', 'ed@bettyboop.com'), ('ed', 'ed@lala.com'), + ('ed', 'ed@wood.com'), ('fred', 'fred@fred.com')]) q2 = q.values(func.count(User.name)) - assert q2.next() == (4,) + assert next(q2) == (4,) - q2 = q.select_from(sel).filter(User.id==8).values(User.name, sel.c.name, User.name) - eq_(list(q2), [(u'ed', u'ed', u'ed')]) + q2 = q.select_entity_from(sel).filter(User.id==8).values(User.name, sel.c.name, User.name) + eq_(list(q2), [('ed', 'ed', 'ed')]) # using User.xxx is alised against "sel", so this query returns nothing - q2 = q.select_from(sel).\ + q2 = q.select_entity_from(sel).\ filter(User.id==8).\ filter(User.id>sel.c.id).values(User.name, sel.c.name, User.name) eq_(list(q2), []) # whereas this uses users.c.xxx, is not aliased and creates a new join - q2 = q.select_from(sel).\ + q2 = q.select_entity_from(sel).\ filter(users.c.id==8).\ filter(users.c.id>sel.c.id).values(users.c.name, sel.c.name, User.name) - eq_(list(q2), [(u'ed', u'jack', u'jack')]) + eq_(list(q2), [('ed', 'jack', 'jack')]) def test_alias_naming(self): User = self.classes.User @@ -1119,14 +1144,14 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): sel = users.select(User.id.in_([7, 8])).alias() q = sess.query(User) u2 = aliased(User) - q2 = q.select_from(sel).\ + q2 = q.select_entity_from(sel).\ filter(u2.id>1).\ order_by(User.id, sel.c.id, u2.id).\ values(User.name, sel.c.name, u2.name) - eq_(list(q2), [(u'jack', u'jack', u'jack'), (u'jack', u'jack', u'ed'), - (u'jack', u'jack', u'fred'), (u'jack', u'jack', u'chuck'), - (u'ed', u'ed', u'jack'), (u'ed', u'ed', u'ed'), - (u'ed', u'ed', u'fred'), (u'ed', u'ed', u'chuck')]) + eq_(list(q2), [('jack', 'jack', 'jack'), ('jack', 'jack', 'ed'), + ('jack', 'jack', 'fred'), ('jack', 'jack', 'chuck'), + ('ed', 'ed', 'jack'), ('ed', 'ed', 'ed'), + ('ed', 'ed', 'fred'), ('ed', 'ed', 'chuck')]) @testing.fails_on('mssql', 'FIXME: unknown') @testing.fails_on('oracle', @@ -1177,8 +1202,8 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): # we don't want Address to be outside of the subquery here eq_( list(sess.query(User, subq)[0:3]), - [(User(id=7,name=u'jack'), 1), (User(id=8,name=u'ed'), 3), - (User(id=9,name=u'fred'), 1)] + [(User(id=7,name='jack'), 1), (User(id=8,name='ed'), 3), + (User(id=9,name='fred'), 1)] ) # same thing without the correlate, as it should @@ -1190,8 +1215,8 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): # we don't want Address to be outside of the subquery here eq_( list(sess.query(User, subq)[0:3]), - [(User(id=7,name=u'jack'), 1), (User(id=8,name=u'ed'), 3), - (User(id=9,name=u'fred'), 1)] + [(User(id=7,name='jack'), 1), (User(id=8,name='ed'), 3), + (User(id=9,name='fred'), 1)] ) @@ -1202,23 +1227,23 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): sess = create_session() - eq_(sess.query(User.name).all(), [(u'jack',), (u'ed',), (u'fred',), (u'chuck',)]) + eq_(sess.query(User.name).all(), [('jack',), ('ed',), ('fred',), ('chuck',)]) sel = users.select(User.id.in_([7, 8])).alias() q = sess.query(User.name) - q2 = q.select_from(sel).all() - eq_(list(q2), [(u'jack',), (u'ed',)]) + q2 = q.select_entity_from(sel).all() + eq_(list(q2), [('jack',), ('ed',)]) eq_(sess.query(User.name, Address.email_address).filter(User.id==Address.user_id).all(), [ - (u'jack', u'jack@bean.com'), (u'ed', u'ed@wood.com'), - (u'ed', u'ed@bettyboop.com'), (u'ed', u'ed@lala.com'), - (u'fred', u'fred@fred.com') + ('jack', 'jack@bean.com'), ('ed', 'ed@wood.com'), + ('ed', 'ed@bettyboop.com'), ('ed', 'ed@lala.com'), + ('fred', 'fred@fred.com') ]) eq_(sess.query(User.name, func.count(Address.email_address)).\ outerjoin(User.addresses).group_by(User.id, User.name).\ order_by(User.id).all(), - [(u'jack', 1), (u'ed', 3), (u'fred', 1), (u'chuck', 0)] + [('jack', 1), ('ed', 3), ('fred', 1), ('chuck', 0)] ) eq_(sess.query(User, func.count(Address.email_address)).\ @@ -1246,8 +1271,8 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): eq_(sess.query(func.count(adalias.email_address), User).\ outerjoin(adalias, User.addresses).group_by(User).\ order_by(User.id).all(), - [(1, User(name=u'jack',id=7)), (3, User(name=u'ed',id=8)), - (1, User(name=u'fred',id=9)), (0, User(name=u'chuck',id=10))] + [(1, User(name='jack',id=7)), (3, User(name='ed',id=8)), + (1, User(name='fred',id=9)), (0, User(name='chuck',id=10))] ) # select from aliasing + explicit aliasing @@ -1257,12 +1282,12 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): from_self(User, adalias.email_address).\ order_by(User.id, adalias.id).all(), [ - (User(name=u'jack',id=7), u'jack@bean.com'), - (User(name=u'ed',id=8), u'ed@wood.com'), - (User(name=u'ed',id=8), u'ed@bettyboop.com'), - (User(name=u'ed',id=8), u'ed@lala.com'), - (User(name=u'fred',id=9), u'fred@fred.com'), - (User(name=u'chuck',id=10), None) + (User(name='jack',id=7), 'jack@bean.com'), + (User(name='ed',id=8), 'ed@wood.com'), + (User(name='ed',id=8), 'ed@bettyboop.com'), + (User(name='ed',id=8), 'ed@lala.com'), + (User(name='fred',id=9), 'fred@fred.com'), + (User(name='chuck',id=10), None) ] ) @@ -1272,12 +1297,12 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): filter(Address.email_address.like('%ed%')).\ from_self().all(), [ - User(name=u'ed',id=8), - User(name=u'fred',id=9), + User(name='ed',id=8), + User(name='fred',id=9), ] ) - # test eager aliasing, with/without select_from aliasing + # test eager aliasing, with/without select_entity_from aliasing for q in [ sess.query(User, adalias.email_address).\ outerjoin(adalias, User.addresses).\ @@ -1293,27 +1318,27 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): q.all(), [(User(addresses=[ - Address(user_id=7,email_address=u'jack@bean.com',id=1)], - name=u'jack',id=7), u'jack@bean.com'), + Address(user_id=7,email_address='jack@bean.com',id=1)], + name='jack',id=7), 'jack@bean.com'), (User(addresses=[ - Address(user_id=8,email_address=u'ed@wood.com',id=2), - Address(user_id=8,email_address=u'ed@bettyboop.com',id=3), - Address(user_id=8,email_address=u'ed@lala.com',id=4)], - name=u'ed',id=8), u'ed@wood.com'), + Address(user_id=8,email_address='ed@wood.com',id=2), + Address(user_id=8,email_address='ed@bettyboop.com',id=3), + Address(user_id=8,email_address='ed@lala.com',id=4)], + name='ed',id=8), 'ed@wood.com'), (User(addresses=[ - Address(user_id=8,email_address=u'ed@wood.com',id=2), - Address(user_id=8,email_address=u'ed@bettyboop.com',id=3), - Address(user_id=8,email_address=u'ed@lala.com',id=4)],name=u'ed',id=8), - u'ed@bettyboop.com'), + Address(user_id=8,email_address='ed@wood.com',id=2), + Address(user_id=8,email_address='ed@bettyboop.com',id=3), + Address(user_id=8,email_address='ed@lala.com',id=4)],name='ed',id=8), + 'ed@bettyboop.com'), (User(addresses=[ - Address(user_id=8,email_address=u'ed@wood.com',id=2), - Address(user_id=8,email_address=u'ed@bettyboop.com',id=3), - Address(user_id=8,email_address=u'ed@lala.com',id=4)],name=u'ed',id=8), - u'ed@lala.com'), - (User(addresses=[Address(user_id=9,email_address=u'fred@fred.com',id=5)],name=u'fred',id=9), - u'fred@fred.com'), + Address(user_id=8,email_address='ed@wood.com',id=2), + Address(user_id=8,email_address='ed@bettyboop.com',id=3), + Address(user_id=8,email_address='ed@lala.com',id=4)],name='ed',id=8), + 'ed@lala.com'), + (User(addresses=[Address(user_id=9,email_address='fred@fred.com',id=5)],name='fred',id=9), + 'fred@fred.com'), - (User(addresses=[],name=u'chuck',id=10), None)] + (User(addresses=[],name='chuck',id=10), None)] ) def test_column_from_limited_joinedload(self): @@ -1367,19 +1392,19 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): eq_( q.all(), [ - (Order(address_id=1,description=u'order 3',isopen=1,user_id=7,id=3), - Order(address_id=1,description=u'order 1',isopen=0,user_id=7,id=1)), - (Order(address_id=None,description=u'order 5',isopen=0,user_id=7,id=5), - Order(address_id=1,description=u'order 1',isopen=0,user_id=7,id=1)), - (Order(address_id=None,description=u'order 5',isopen=0,user_id=7,id=5), - Order(address_id=1,description=u'order 3',isopen=1,user_id=7,id=3)) + (Order(address_id=1,description='order 3',isopen=1,user_id=7,id=3), + Order(address_id=1,description='order 1',isopen=0,user_id=7,id=1)), + (Order(address_id=None,description='order 5',isopen=0,user_id=7,id=5), + Order(address_id=1,description='order 1',isopen=0,user_id=7,id=1)), + (Order(address_id=None,description='order 5',isopen=0,user_id=7,id=5), + Order(address_id=1,description='order 3',isopen=1,user_id=7,id=3)) ] ) # ensure column expressions are taken from inside the subquery, not restated at the top q = sess.query(Order.id, Order.description, literal_column("'q'").label('foo')).\ - filter(Order.description == u'order 3').from_self() + filter(Order.description == 'order 3').from_self() self.assert_compile(q, "SELECT anon_1.orders_id AS " "anon_1_orders_id, anon_1.orders_descriptio" @@ -1392,7 +1417,7 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): "anon_1") eq_( q.all(), - [(3, u'order 3', 'q')] + [(3, 'order 3', 'q')] ) @@ -1466,14 +1491,14 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): q = sess.query(User) adalias = addresses.alias('adalias') - q = q.add_entity(Address, alias=adalias).select_from(users.outerjoin(adalias)) + q = q.add_entity(Address, alias=adalias).select_entity_from(users.outerjoin(adalias)) l = q.order_by(User.id, adalias.c.id).all() assert l == expected sess.expunge_all() q = sess.query(User).add_entity(Address, alias=adalias) - l = q.select_from(users.outerjoin(adalias)).filter(adalias.c.email_address=='ed@bettyboop.com').all() + l = q.select_entity_from(users.outerjoin(adalias)).filter(adalias.c.email_address=='ed@bettyboop.com').all() assert l == [(user8, address3)] def test_with_entities(self): @@ -1518,7 +1543,7 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): eq_( sess.query(User.id).add_column(users).all(), - [(7, 7, u'jack'), (8, 8, u'ed'), (9, 9, u'fred'), (10, 10, u'chuck')] + [(7, 7, 'jack'), (8, 8, 'ed'), (9, 9, 'fred'), (10, 10, 'chuck')] ) def test_multi_columns_2(self): @@ -1591,9 +1616,9 @@ class MixedEntitiesTest(QueryTest, AssertsCompiledSQL): sess.expunge_all() - # test with select_from() + # test with select_entity_from() q = create_session().query(User).add_column(func.count(addresses.c.id))\ - .add_column(("Name:" + users.c.name)).select_from(users.outerjoin(addresses))\ + .add_column(("Name:" + users.c.name)).select_entity_from(users.outerjoin(addresses))\ .group_by(users).order_by(users.c.id) assert q.all() == expected @@ -1718,19 +1743,19 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): sel = users.select(users.c.id.in_([7, 8])).alias() sess = create_session() - eq_(sess.query(User).select_from(sel).all(), [User(id=7), User(id=8)]) + eq_(sess.query(User).select_entity_from(sel).all(), [User(id=7), User(id=8)]) - eq_(sess.query(User).select_from(sel).filter(User.id==8).all(), [User(id=8)]) + eq_(sess.query(User).select_entity_from(sel).filter(User.id==8).all(), [User(id=8)]) - eq_(sess.query(User).select_from(sel).order_by(desc(User.name)).all(), [ + eq_(sess.query(User).select_entity_from(sel).order_by(desc(User.name)).all(), [ User(name='jack',id=7), User(name='ed',id=8) ]) - eq_(sess.query(User).select_from(sel).order_by(asc(User.name)).all(), [ + eq_(sess.query(User).select_entity_from(sel).order_by(asc(User.name)).all(), [ User(name='ed',id=8), User(name='jack',id=7) ]) - eq_(sess.query(User).select_from(sel).options(joinedload('addresses')).first(), + eq_(sess.query(User).select_entity_from(sel).options(joinedload('addresses')).first(), User(name='jack', addresses=[Address(id=1)]) ) @@ -1745,7 +1770,7 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): sel = users.select(users.c.id.in_([7, 8])) sess = create_session() - eq_(sess.query(User).select_from(sel).all(), + eq_(sess.query(User).select_entity_from(sel).all(), [ User(name='jack',id=7), User(name='ed',id=8) ] @@ -1772,14 +1797,14 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): ) self.assert_compile( - sess.query(ualias).select_from(sel).filter(ualias.id>sel.c.id), + sess.query(ualias).select_entity_from(sel).filter(ualias.id>sel.c.id), "SELECT users_1.id AS users_1_id, users_1.name AS users_1_name FROM " "users AS users_1, (SELECT users.id AS id, users.name AS name FROM " "users WHERE users.id IN (:id_1, :id_2)) AS anon_1 WHERE users_1.id > anon_1.id", ) self.assert_compile( - sess.query(ualias).select_from(sel).join(ualias, ualias.id>sel.c.id), + sess.query(ualias).select_entity_from(sel).join(ualias, ualias.id>sel.c.id), "SELECT users_1.id AS users_1_id, users_1.name AS users_1_name " "FROM (SELECT users.id AS id, users.name AS name " "FROM users WHERE users.id IN (:id_1, :id_2)) AS anon_1 " @@ -1787,7 +1812,7 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): ) self.assert_compile( - sess.query(ualias).select_from(sel).join(ualias, ualias.id>User.id), + sess.query(ualias).select_entity_from(sel).join(ualias, ualias.id>User.id), "SELECT users_1.id AS users_1_id, users_1.name AS users_1_name " "FROM (SELECT users.id AS id, users.name AS name FROM " "users WHERE users.id IN (:id_1, :id_2)) AS anon_1 " @@ -1805,7 +1830,7 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): # this one uses an explicit join(left, right, onclause) so works self.assert_compile( - sess.query(ualias).select_from(join(sel, ualias, ualias.id>sel.c.id)), + sess.query(ualias).select_entity_from(join(sel, ualias, ualias.id>sel.c.id)), "SELECT users_1.id AS users_1_id, users_1.name AS users_1_name FROM " "(SELECT users.id AS id, users.name AS name FROM users WHERE users.id " "IN (:id_1, :id_2)) AS anon_1 JOIN users AS users_1 ON users_1.id > anon_1.id", @@ -1848,7 +1873,7 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): # here for comparison self.assert_compile( sess.query(User.name).\ - select_from(users.select().where(users.c.id > 5)), + select_entity_from(users.select().where(users.c.id > 5)), "SELECT anon_1.name AS anon_1_name FROM (SELECT users.id AS id, " "users.name AS name FROM users WHERE users.id > :id_1) AS anon_1" ) @@ -1861,7 +1886,7 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): sel = users.select(users.c.id.in_([7, 8])) sess = create_session() - eq_(sess.query(User).select_from(sel).all(), + eq_(sess.query(User).select_entity_from(sel).all(), [ User(name='jack',id=7), User(name='ed',id=8) ] @@ -1881,7 +1906,7 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): sel = users.select(users.c.id.in_([7, 8])) sess = create_session() - eq_(sess.query(User).select_from(sel).join('addresses'). + eq_(sess.query(User).select_entity_from(sel).join('addresses'). add_entity(Address).order_by(User.id).order_by(Address.id).all(), [ (User(name='jack',id=7), Address(user_id=7,email_address='jack@bean.com',id=1)), @@ -1892,7 +1917,7 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): ) adalias = aliased(Address) - eq_(sess.query(User).select_from(sel).join(adalias, 'addresses'). + eq_(sess.query(User).select_entity_from(sel).join(adalias, 'addresses'). add_entity(adalias).order_by(User.id).order_by(adalias.id).all(), [ (User(name='jack',id=7), Address(user_id=7,email_address='jack@bean.com',id=1)), @@ -1931,82 +1956,82 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): sess = create_session() sel = users.select(users.c.id.in_([7, 8])) - eq_(sess.query(User).select_from(sel).\ + eq_(sess.query(User).select_entity_from(sel).\ join('orders', 'items', 'keywords').\ filter(Keyword.name.in_(['red', 'big', 'round'])).\ all(), [ - User(name=u'jack',id=7) + User(name='jack',id=7) ]) - eq_(sess.query(User).select_from(sel).\ + eq_(sess.query(User).select_entity_from(sel).\ join('orders', 'items', 'keywords', aliased=True).\ filter(Keyword.name.in_(['red', 'big', 'round'])).\ all(), [ - User(name=u'jack',id=7) + User(name='jack',id=7) ]) def go(): eq_( - sess.query(User).select_from(sel). + sess.query(User).select_entity_from(sel). options(joinedload_all('orders.items.keywords')). join('orders', 'items', 'keywords', aliased=True). filter(Keyword.name.in_(['red', 'big', 'round'])).\ all(), [ - User(name=u'jack',orders=[ - Order(description=u'order 1',items=[ - Item(description=u'item 1', + User(name='jack',orders=[ + Order(description='order 1',items=[ + Item(description='item 1', keywords=[ - Keyword(name=u'red'), - Keyword(name=u'big'), - Keyword(name=u'round') + Keyword(name='red'), + Keyword(name='big'), + Keyword(name='round') ]), - Item(description=u'item 2', + Item(description='item 2', keywords=[ - Keyword(name=u'red',id=2), - Keyword(name=u'small',id=5), - Keyword(name=u'square') + Keyword(name='red',id=2), + Keyword(name='small',id=5), + Keyword(name='square') ]), - Item(description=u'item 3', + Item(description='item 3', keywords=[ - Keyword(name=u'green',id=3), - Keyword(name=u'big',id=4), - Keyword(name=u'round',id=6)]) + Keyword(name='green',id=3), + Keyword(name='big',id=4), + Keyword(name='round',id=6)]) ]), - Order(description=u'order 3',items=[ - Item(description=u'item 3', + Order(description='order 3',items=[ + Item(description='item 3', keywords=[ - Keyword(name=u'green',id=3), - Keyword(name=u'big',id=4), - Keyword(name=u'round',id=6) + Keyword(name='green',id=3), + Keyword(name='big',id=4), + Keyword(name='round',id=6) ]), - Item(description=u'item 4',keywords=[],id=4), - Item(description=u'item 5',keywords=[],id=5) + Item(description='item 4',keywords=[],id=4), + Item(description='item 5',keywords=[],id=5) ]), - Order(description=u'order 5', + Order(description='order 5', items=[ - Item(description=u'item 5',keywords=[])]) + Item(description='item 5',keywords=[])]) ]) ]) self.assert_sql_count(testing.db, go, 1) sess.expunge_all() sel2 = orders.select(orders.c.id.in_([1,2,3])) - eq_(sess.query(Order).select_from(sel2).\ + eq_(sess.query(Order).select_entity_from(sel2).\ join('items', 'keywords').\ filter(Keyword.name == 'red').\ order_by(Order.id).all(), [ - Order(description=u'order 1',id=1), - Order(description=u'order 2',id=2), + Order(description='order 1',id=1), + Order(description='order 2',id=2), ]) - eq_(sess.query(Order).select_from(sel2).\ + eq_(sess.query(Order).select_entity_from(sel2).\ join('items', 'keywords', aliased=True).\ filter(Keyword.name == 'red').\ order_by(Order.id).all(), [ - Order(description=u'order 1',id=1), - Order(description=u'order 2',id=2), + Order(description='order 1',id=1), + Order(description='order 2',id=2), ]) @@ -2025,7 +2050,9 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): sess = create_session() def go(): - eq_(sess.query(User).options(joinedload('addresses')).select_from(sel).order_by(User.id).all(), + eq_(sess.query(User).options( + joinedload('addresses') + ).select_entity_from(sel).order_by(User.id).all(), [ User(id=7, addresses=[Address(id=1)]), User(id=8, addresses=[Address(id=2), Address(id=3), Address(id=4)]) @@ -2035,14 +2062,19 @@ class SelectFromTest(QueryTest, AssertsCompiledSQL): sess.expunge_all() def go(): - eq_(sess.query(User).options(joinedload('addresses')).select_from(sel).filter(User.id==8).order_by(User.id).all(), + eq_(sess.query(User).options( + joinedload('addresses') + ).select_entity_from(sel).filter(User.id==8).order_by(User.id).all(), [User(id=8, addresses=[Address(id=2), Address(id=3), Address(id=4)])] ) self.assert_sql_count(testing.db, go, 1) sess.expunge_all() def go(): - eq_(sess.query(User).options(joinedload('addresses')).select_from(sel).order_by(User.id)[1], User(id=8, addresses=[Address(id=2), Address(id=3), Address(id=4)])) + eq_(sess.query(User).options( + joinedload('addresses') + ).select_entity_from(sel).order_by(User.id)[1], + User(id=8, addresses=[Address(id=2), Address(id=3), Address(id=4)])) self.assert_sql_count(testing.db, go, 1) class CustomJoinTest(QueryTest): @@ -2180,7 +2212,7 @@ class ExternalColumnsTest(QueryTest): ua = aliased(User) eq_(sess.query(Address, ua.concat, ua.count). - select_from(join(Address, ua, 'user')). + select_entity_from(join(Address, ua, 'user')). options(joinedload(Address.user)).order_by(Address.id).all(), [ (Address(id=1, user=User(id=7, concat=14, count=1)), 14, 1), @@ -2195,7 +2227,7 @@ class ExternalColumnsTest(QueryTest): [(1, 7, 14, 1), (2, 8, 16, 3), (3, 8, 16, 3), (4, 8, 16, 3), (5, 9, 18, 1)] ) - eq_(list(sess.query(Address, ua).select_from(join(Address,ua, 'user')).values(Address.id, ua.id, ua.concat, ua.count)), + eq_(list(sess.query(Address, ua).select_entity_from(join(Address,ua, 'user')).values(Address.id, ua.id, ua.concat, ua.count)), [(1, 7, 14, 1), (2, 8, 16, 3), (3, 8, 16, 3), (4, 8, 16, 3), (5, 9, 18, 1)] ) diff --git a/test/orm/test_generative.py b/test/orm/test_generative.py index 3f5da69c3..52858cc26 100644 --- a/test/orm/test_generative.py +++ b/test/orm/test_generative.py @@ -78,13 +78,8 @@ class GenerativeQueryTest(fixtures.MappedTest): assert sess.query(func.min(foo.c.bar)).filter(foo.c.bar<30).one() == (0,) assert sess.query(func.max(foo.c.bar)).filter(foo.c.bar<30).one() == (29,) - # Py3K - #assert query.filter(foo.c.bar<30).values(sa.func.max(foo.c.bar)).__next__()[0] == 29 - #assert query.filter(foo.c.bar<30).values(sa.func.max(foo.c.bar)).__next__()[0] == 29 - # Py2K - assert query.filter(foo.c.bar<30).values(sa.func.max(foo.c.bar)).next()[0] == 29 - assert query.filter(foo.c.bar<30).values(sa.func.max(foo.c.bar)).next()[0] == 29 - # end Py2K + assert next(query.filter(foo.c.bar<30).values(sa.func.max(foo.c.bar)))[0] == 29 + assert next(query.filter(foo.c.bar<30).values(sa.func.max(foo.c.bar)))[0] == 29 @testing.fails_if(lambda:testing.against('mysql+mysqldb') and testing.db.dialect.dbapi.version_info[:4] == (1, 2, 1, 'gamma'), @@ -111,18 +106,10 @@ class GenerativeQueryTest(fixtures.MappedTest): query = create_session().query(Foo) - # Py3K - #avg_f = query.filter(foo.c.bar<30).values(sa.func.avg(foo.c.bar)).__next__()[0] - # Py2K - avg_f = query.filter(foo.c.bar<30).values(sa.func.avg(foo.c.bar)).next()[0] - # end Py2K + avg_f = next(query.filter(foo.c.bar<30).values(sa.func.avg(foo.c.bar)))[0] assert float(round(avg_f, 1)) == 14.5 - # Py3K - #avg_o = query.filter(foo.c.bar<30).values(sa.func.avg(foo.c.bar)).__next__()[0] - # Py2K - avg_o = query.filter(foo.c.bar<30).values(sa.func.avg(foo.c.bar)).next()[0] - # end Py2K + avg_o = next(query.filter(foo.c.bar<30).values(sa.func.avg(foo.c.bar)))[0] assert float(round(avg_o, 1)) == 14.5 def test_filter(self): diff --git a/test/orm/test_inspect.py b/test/orm/test_inspect.py index 2a401f91d..78d84f505 100644 --- a/test/orm/test_inspect.py +++ b/test/orm/test_inspect.py @@ -142,7 +142,7 @@ class TestORMInspection(_fixtures.FixtureTest): [insp.get_property('id'), insp.get_property('name')] ) eq_( - insp.column_attrs.keys(), + list(insp.column_attrs.keys()), ['id', 'name'] ) is_( @@ -274,7 +274,7 @@ class TestORMInspection(_fixtures.FixtureTest): insp = inspect(SomeSubClass) eq_( dict((k, v.extension_type) - for k, v in insp.all_orm_descriptors.items() + for k, v in list(insp.all_orm_descriptors.items()) ), { 'id': NOT_EXTENSION, diff --git a/test/orm/test_instrumentation.py b/test/orm/test_instrumentation.py index 1a1a70e8b..c3d24ebe7 100644 --- a/test/orm/test_instrumentation.py +++ b/test/orm/test_instrumentation.py @@ -540,30 +540,6 @@ class NativeInstrumentationTest(fixtures.ORMTest): class Py3KFunctionInstTest(fixtures.ORMTest): __requires__ = ("python3", ) - # Py3K - #def _kw_only_fixture(self): - # class A(object): - # def __init__(self, a, *, b, c): - # self.a = a - # self.b = b - # self.c = c - # return self._instrument(A) - # - #def _kw_plus_posn_fixture(self): - # class A(object): - # def __init__(self, a, *args, b, c): - # self.a = a - # self.b = b - # self.c = c - # return self._instrument(A) - # - #def _kw_opt_fixture(self): - # class A(object): - # def __init__(self, a, *, b, c="c"): - # self.a = a - # self.b = b - # self.c = c - # return self._instrument(A) def _instrument(self, cls): manager = instrumentation.register_class(cls) @@ -614,6 +590,36 @@ class Py3KFunctionInstTest(fixtures.ORMTest): cls, "a", "b", c="c" ) +if util.py3k: + _locals = {} + exec(""" +def _kw_only_fixture(self): + class A(object): + def __init__(self, a, *, b, c): + self.a = a + self.b = b + self.c = c + return self._instrument(A) + +def _kw_plus_posn_fixture(self): + class A(object): + def __init__(self, a, *args, b, c): + self.a = a + self.b = b + self.c = c + return self._instrument(A) + +def _kw_opt_fixture(self): + class A(object): + def __init__(self, a, *, b, c="c"): + self.a = a + self.b = b + self.c = c + return self._instrument(A) +""", _locals) + for k in _locals: + setattr(Py3KFunctionInstTest, k, _locals[k]) + class MiscTest(fixtures.ORMTest): """Seems basic, but not directly covered elsewhere!""" diff --git a/test/orm/test_joins.py b/test/orm/test_joins.py index 2bf0d8d92..2dac59150 100644 --- a/test/orm/test_joins.py +++ b/test/orm/test_joins.py @@ -774,14 +774,14 @@ class JoinTest(QueryTest, AssertsCompiledSQL): eq_( sess.query(User).join(Address.user).\ filter(Address.email_address=='ed@wood.com').all(), - [User(id=8,name=u'ed')] + [User(id=8,name='ed')] ) # its actually not so controversial if you view it in terms # of multiple entities. eq_( sess.query(User, Address).join(Address.user).filter(Address.email_address=='ed@wood.com').all(), - [(User(id=8,name=u'ed'), Address(email_address='ed@wood.com'))] + [(User(id=8,name='ed'), Address(email_address='ed@wood.com'))] ) # this was the controversial part. now, raise an error if the feature is abused. @@ -1066,7 +1066,7 @@ class JoinTest(QueryTest, AssertsCompiledSQL): # same with an explicit select_from() eq_( - sess.query(User).select_from(select([users]). + sess.query(User).select_entity_from(select([users]). order_by(User.id).offset(2).alias()). join(Order, User.id==Order.user_id). all(), @@ -1162,9 +1162,9 @@ class JoinTest(QueryTest, AssertsCompiledSQL): sess.query(OrderAlias).join('items').filter_by(description='item 3').\ order_by(OrderAlias.id).all(), [ - Order(address_id=1,description=u'order 1',isopen=0,user_id=7,id=1), - Order(address_id=4,description=u'order 2',isopen=0,user_id=9,id=2), - Order(address_id=1,description=u'order 3',isopen=1,user_id=7,id=3) + Order(address_id=1,description='order 1',isopen=0,user_id=7,id=1), + Order(address_id=4,description='order 2',isopen=0,user_id=9,id=2), + Order(address_id=1,description='order 3',isopen=1,user_id=7,id=3) ] ) @@ -1175,9 +1175,9 @@ class JoinTest(QueryTest, AssertsCompiledSQL): filter_by(description='item 3').\ order_by(User.id, OrderAlias.id).all(), [ - (User(name=u'jack',id=7), Order(address_id=1,description=u'order 1',isopen=0,user_id=7,id=1), u'item 3'), - (User(name=u'jack',id=7), Order(address_id=1,description=u'order 3',isopen=1,user_id=7,id=3), u'item 3'), - (User(name=u'fred',id=9), Order(address_id=4,description=u'order 2',isopen=0,user_id=9,id=2), u'item 3') + (User(name='jack',id=7), Order(address_id=1,description='order 1',isopen=0,user_id=7,id=1), 'item 3'), + (User(name='jack',id=7), Order(address_id=1,description='order 3',isopen=1,user_id=7,id=3), 'item 3'), + (User(name='fred',id=9), Order(address_id=4,description='order 2',isopen=0,user_id=9,id=2), 'item 3') ] ) @@ -1334,12 +1334,12 @@ class JoinTest(QueryTest, AssertsCompiledSQL): eq_( sess.query(User, ualias).filter(User.id > ualias.id).order_by(desc(ualias.id), User.name).all(), [ - (User(id=10,name=u'chuck'), User(id=9,name=u'fred')), - (User(id=10,name=u'chuck'), User(id=8,name=u'ed')), - (User(id=9,name=u'fred'), User(id=8,name=u'ed')), - (User(id=10,name=u'chuck'), User(id=7,name=u'jack')), - (User(id=8,name=u'ed'), User(id=7,name=u'jack')), - (User(id=9,name=u'fred'), User(id=7,name=u'jack')) + (User(id=10,name='chuck'), User(id=9,name='fred')), + (User(id=10,name='chuck'), User(id=8,name='ed')), + (User(id=9,name='fred'), User(id=8,name='ed')), + (User(id=10,name='chuck'), User(id=7,name='jack')), + (User(id=8,name='ed'), User(id=7,name='jack')), + (User(id=9,name='fred'), User(id=7,name='jack')) ] ) @@ -1351,7 +1351,7 @@ class JoinTest(QueryTest, AssertsCompiledSQL): eq_( sess.query(User.name).join(addresses, User.id==addresses.c.user_id).order_by(User.id).all(), - [(u'jack',), (u'ed',), (u'ed',), (u'ed',), (u'fred',)] + [('jack',), ('ed',), ('ed',), ('ed',), ('fred',)] ) def test_no_joinpoint_expr(self): @@ -2066,13 +2066,13 @@ class SelfReferentialTest(fixtures.MappedTest, AssertsCompiledSQL): # using 'n1.parent' implicitly joins to unaliased Node eq_( sess.query(n1).join(n1.parent).filter(Node.data=='n1').all(), - [Node(parent_id=1,data=u'n11',id=2), Node(parent_id=1,data=u'n12',id=3), Node(parent_id=1,data=u'n13',id=4)] + [Node(parent_id=1,data='n11',id=2), Node(parent_id=1,data='n12',id=3), Node(parent_id=1,data='n13',id=4)] ) # explicit (new syntax) eq_( sess.query(n1).join(Node, n1.parent).filter(Node.data=='n1').all(), - [Node(parent_id=1,data=u'n11',id=2), Node(parent_id=1,data=u'n12',id=3), Node(parent_id=1,data=u'n13',id=4)] + [Node(parent_id=1,data='n11',id=2), Node(parent_id=1,data='n12',id=3), Node(parent_id=1,data='n13',id=4)] ) diff --git a/test/orm/test_loading.py b/test/orm/test_loading.py index 20dc082c5..97c08ea29 100644 --- a/test/orm/test_loading.py +++ b/test/orm/test_loading.py @@ -73,7 +73,7 @@ class MergeResultTest(_fixtures.FixtureTest): [(x.id, y) for x, y in it], [(1, 1), (2, 2), (7, 7), (8, 8)] ) - eq_(it[0].keys(), ['User', 'id']) + eq_(list(it[0].keys()), ['User', 'id']) def test_entity_col_mix_keyed_tuple(self): s, (u1, u2, u3, u4) = self._fixture() @@ -91,7 +91,7 @@ class MergeResultTest(_fixtures.FixtureTest): [(x.id, y) for x, y in it], [(1, 1), (2, 2), (7, 7), (8, 8)] ) - eq_(it[0].keys(), ['User', 'id']) + eq_(list(it[0].keys()), ['User', 'id']) def test_none_entity(self): s, (u1, u2, u3, u4) = self._fixture() diff --git a/test/orm/test_mapper.py b/test/orm/test_mapper.py index 6b97fb135..19ff78004 100644 --- a/test/orm/test_mapper.py +++ b/test/orm/test_mapper.py @@ -272,13 +272,14 @@ class MapperTest(_fixtures.FixtureTest, AssertsCompiledSQL): ) m = mapper(Foo, foo_t) class DontCompareMeToString(int): - # Py3K - # pass - # Py2K - def __lt__(self, other): - assert not isinstance(other, basestring) - return int(self) < other - # end Py2K +# start Py3K + pass +# end Py3K +# start Py2K +# def __lt__(self, other): +# assert not isinstance(other, basestring) +# return int(self) < other +# end Py2K foos = [Foo(id='f%d' % i) for i in range(5)] states = [attributes.instance_state(f) for f in foos] @@ -847,7 +848,7 @@ class MapperTest(_fixtures.FixtureTest, AssertsCompiledSQL): def test_we_dont_call_bool(self): class NoBoolAllowed(object): - def __nonzero__(self): + def __bool__(self): raise Exception("nope") mapper(NoBoolAllowed, self.tables.users) u1 = NoBoolAllowed() @@ -1057,12 +1058,12 @@ class MapperTest(_fixtures.FixtureTest, AssertsCompiledSQL): eq_( create_session().query(User).all(), - [User(id=7, name=u'jack'), User(id=9, name=u'fred'), User(id=8, name=u'ed'), User(id=10, name=u'chuck')] + [User(id=7, name='jack'), User(id=9, name='fred'), User(id=8, name='ed'), User(id=10, name='chuck')] ) eq_( create_session().query(User).order_by(User.name).all(), - [User(id=10, name=u'chuck'), User(id=8, name=u'ed'), User(id=9, name=u'fred'), User(id=7, name=u'jack')] + [User(id=10, name='chuck'), User(id=8, name='ed'), User(id=9, name='fred'), User(id=7, name='jack')] ) # 'Raises a "expression evaluation not supported" error at prepare time @@ -2123,7 +2124,7 @@ class ValidatorTest(_fixtures.FixtureTest): mapper(Address, addresses) eq_( - dict((k, v[0].__name__) for k, v in u_m.validators.items()), + dict((k, v[0].__name__) for k, v in list(u_m.validators.items())), {'name':'validate_name', 'addresses':'validate_address'} ) @@ -2992,29 +2993,29 @@ class RequirementsTest(fixtures.MappedTest): Column('ht1b_id', Integer, ForeignKey('ht1.id'), primary_key=True), Column('value', String(10))) - # Py2K - def test_baseclass(self): - ht1 = self.tables.ht1 - - class OldStyle: - pass - - assert_raises(sa.exc.ArgumentError, mapper, OldStyle, ht1) - - assert_raises(sa.exc.ArgumentError, mapper, 123) - - class NoWeakrefSupport(str): - pass - - # TODO: is weakref support detectable without an instance? - #self.assertRaises(sa.exc.ArgumentError, mapper, NoWeakrefSupport, t2) - # end Py2K +# start Py2K +# def test_baseclass(self): +# ht1 = self.tables.ht1 +# +# class OldStyle: +# pass +# +# assert_raises(sa.exc.ArgumentError, mapper, OldStyle, ht1) +# +# assert_raises(sa.exc.ArgumentError, mapper, 123) +# +# class NoWeakrefSupport(str): +# pass +# +# # TODO: is weakref support detectable without an instance? +# #self.assertRaises(sa.exc.ArgumentError, mapper, NoWeakrefSupport, t2) +# end Py2K class _ValueBase(object): def __init__(self, value='abc', id=None): self.id = id self.value = value - def __nonzero__(self): + def __bool__(self): return False def __hash__(self): return hash(self.value) @@ -3173,7 +3174,7 @@ class RequirementsTest(fixtures.MappedTest): return self.value class H2(object): - def __nonzero__(self): + def __bool__(self): return bool(self.get_value()) def get_value(self): @@ -3224,7 +3225,7 @@ class IsUserlandTest(fixtures.MappedTest): self._test("someprop") def test_unicode(self): - self._test(u"someprop") + self._test("someprop") def test_int(self): self._test(5) diff --git a/test/orm/test_merge.py b/test/orm/test_merge.py index e1474f39b..494f5e349 100644 --- a/test/orm/test_merge.py +++ b/test/orm/test_merge.py @@ -346,7 +346,7 @@ class MergeTest(_fixtures.FixtureTest): sess = create_session() sess.merge(u1) sess.flush() - assert u1.addresses.keys() == ['foo@bar.com'] + assert list(u1.addresses.keys()) == ['foo@bar.com'] def test_attribute_cascade(self): """Merge of a persistent entity with two child @@ -803,7 +803,7 @@ class MergeTest(_fixtures.FixtureTest): try: sess2.merge(u, load=False) assert False - except sa.exc.InvalidRequestError, e: + except sa.exc.InvalidRequestError as e: assert "merge() with load=False option does not support "\ "objects marked as 'dirty'. flush() all changes on "\ "mapped instances before merging with load=False." \ @@ -924,7 +924,7 @@ class MergeTest(_fixtures.FixtureTest): sess2.expunge_all() eq_(sess2.query(User).get(u2.id).addresses[0].email_address, 'somenewaddress') - except sa.exc.InvalidRequestError, e: + except sa.exc.InvalidRequestError as e: assert "load=False option does not support" in str(e) def test_synonym_comparable(self): diff --git a/test/orm/test_naturalpks.py b/test/orm/test_naturalpks.py index d30cdc598..173408b82 100644 --- a/test/orm/test_naturalpks.py +++ b/test/orm/test_naturalpks.py @@ -2,7 +2,7 @@ Primary key changing capabilities and passive/non-passive cascading updates. """ -from __future__ import with_statement + from sqlalchemy.testing import eq_, ne_, \ assert_raises, assert_raises_message import sqlalchemy as sa @@ -499,7 +499,7 @@ class ReversePKsTest(fixtures.MappedTest): 'user', metadata, Column('code', Integer, primary_key=True), Column('status', Integer, primary_key=True), - Column('username', Unicode(50), nullable=False), + Column('username', String(50), nullable=False), ) @classmethod @@ -519,11 +519,11 @@ class ReversePKsTest(fixtures.MappedTest): session = sa.orm.sessionmaker()() - a_published = User(1, PUBLISHED, u'a') + a_published = User(1, PUBLISHED, 'a') session.add(a_published) session.commit() - a_editable = User(1, EDITABLE, u'a') + a_editable = User(1, EDITABLE, 'a') session.add(a_editable) session.commit() diff --git a/test/orm/test_pickled.py b/test/orm/test_pickled.py index 5ac34f914..b54af93f2 100644 --- a/test/orm/test_pickled.py +++ b/test/orm/test_pickled.py @@ -443,21 +443,21 @@ class TupleLabelTest(_fixtures.FixtureTest): if pickled is not False: row = pickle.loads(pickle.dumps(row, pickled)) - eq_(row.keys(), ['User', 'Address']) + eq_(list(row.keys()), ['User', 'Address']) eq_(row.User, row[0]) eq_(row.Address, row[1]) for row in sess.query(User.name, User.id.label('foobar')): if pickled is not False: row = pickle.loads(pickle.dumps(row, pickled)) - eq_(row.keys(), ['name', 'foobar']) + eq_(list(row.keys()), ['name', 'foobar']) eq_(row.name, row[0]) eq_(row.foobar, row[1]) for row in sess.query(User).values(User.name, User.id.label('foobar')): if pickled is not False: row = pickle.loads(pickle.dumps(row, pickled)) - eq_(row.keys(), ['name', 'foobar']) + eq_(list(row.keys()), ['name', 'foobar']) eq_(row.name, row[0]) eq_(row.foobar, row[1]) @@ -465,21 +465,21 @@ class TupleLabelTest(_fixtures.FixtureTest): for row in sess.query(User, oalias).join(User.orders).all(): if pickled is not False: row = pickle.loads(pickle.dumps(row, pickled)) - eq_(row.keys(), ['User']) + eq_(list(row.keys()), ['User']) eq_(row.User, row[0]) oalias = aliased(Order, name='orders') for row in sess.query(User, oalias).join(oalias, User.orders).all(): if pickled is not False: row = pickle.loads(pickle.dumps(row, pickled)) - eq_(row.keys(), ['User', 'orders']) + eq_(list(row.keys()), ['User', 'orders']) eq_(row.User, row[0]) eq_(row.orders, row[1]) # test here that first col is not labeled, only # one name in keys, matches correctly for row in sess.query(User.name + 'hoho', User.name): - eq_(row.keys(), ['name']) + eq_(list(row.keys()), ['name']) eq_(row[0], row.name + 'hoho') if pickled is not False: diff --git a/test/orm/test_query.py b/test/orm/test_query.py index 3882ec4b5..2e0355e63 100644 --- a/test/orm/test_query.py +++ b/test/orm/test_query.py @@ -407,11 +407,12 @@ class GetTest(QueryTest): Column('data', Unicode(40))) try: metadata.create_all() - # Py3K - #ustring = b'petit voix m\xe2\x80\x99a'.decode('utf-8') - # Py2K - ustring = 'petit voix m\xe2\x80\x99a'.decode('utf-8') - # end Py2K +# start Py3K + ustring = b'petit voix m\xe2\x80\x99a'.decode('utf-8') +# end Py3K +# start Py2K +# ustring = 'petit voix m\xe2\x80\x99a'.decode('utf-8') +# end Py2K table.insert().execute(id=ustring, data=ustring) class LocalFoo(self.classes.Base): @@ -637,9 +638,9 @@ class OperatorTest(QueryTest, AssertsCompiledSQL): (operator.sub, '-'), # Py3k #(operator.truediv, '/'), - # Py2K - (operator.div, '/'), - # end Py2K +# start Py2K +# (operator.div, '/'), +# end Py2K ): for (lhs, rhs, res) in ( (5, User.id, ':id_1 %s users.id'), @@ -1047,7 +1048,7 @@ class ExpressionTest(QueryTest, AssertsCompiledSQL): adalias = aliased(Address, q1.subquery()) eq_( s.query(User, adalias).join(adalias, User.id==adalias.user_id).all(), - [(User(id=7,name=u'jack'), Address(email_address=u'jack@bean.com',user_id=7,id=1))] + [(User(id=7,name='jack'), Address(email_address='jack@bean.com',user_id=7,id=1))] ) # more slice tests are available in test/orm/generative.py @@ -1196,7 +1197,7 @@ class FilterTest(QueryTest, AssertsCompiledSQL): "users.name AS users_name FROM users WHERE users.id = :param_1 " "UNION SELECT users.id AS users_id, users.name AS users_name " "FROM users WHERE users.id = :param_2) AS anon_1", - checkparams = {u'param_1': 7, u'param_2': 8} + checkparams = {'param_1': 7, 'param_2': 8} ) def test_any(self): @@ -1319,7 +1320,7 @@ class FilterTest(QueryTest, AssertsCompiledSQL): "FROM users JOIN addresses ON users.id = addresses.user_id " "WHERE users.name = :name_1 AND " "addresses.email_address = :email_address_1", - checkparams={u'email_address_1': 'ed@ed.com', u'name_1': 'ed'} + checkparams={'email_address_1': 'ed@ed.com', 'name_1': 'ed'} ) def test_filter_by_no_property(self): @@ -1468,14 +1469,14 @@ class SetOpsTest(QueryTest, AssertsCompiledSQL): for q in (q3.order_by(User.id, "anon_1_param_1"), q6.order_by(User.id, "foo")): eq_(q.all(), [ - (User(id=7, name=u'jack'), u'x'), - (User(id=7, name=u'jack'), u'y'), - (User(id=8, name=u'ed'), u'x'), - (User(id=8, name=u'ed'), u'y'), - (User(id=9, name=u'fred'), u'x'), - (User(id=9, name=u'fred'), u'y'), - (User(id=10, name=u'chuck'), u'x'), - (User(id=10, name=u'chuck'), u'y') + (User(id=7, name='jack'), 'x'), + (User(id=7, name='jack'), 'y'), + (User(id=8, name='ed'), 'x'), + (User(id=8, name='ed'), 'y'), + (User(id=9, name='fred'), 'x'), + (User(id=9, name='fred'), 'y'), + (User(id=10, name='chuck'), 'x'), + (User(id=10, name='chuck'), 'y') ] ) @@ -1603,7 +1604,7 @@ class AggregateTest(QueryTest): sess = create_session() orders = sess.query(Order).filter(Order.id.in_([2, 3, 4])) - eq_(orders.values(func.sum(Order.user_id * Order.address_id)).next(), (79,)) + eq_(next(orders.values(func.sum(Order.user_id * Order.address_id))), (79,)) eq_(orders.value(func.sum(Order.user_id * Order.address_id)), 79) def test_apply(self): @@ -1616,9 +1617,9 @@ class AggregateTest(QueryTest): User, Address = self.classes.User, self.classes.Address sess = create_session() - assert [User(name=u'ed',id=8)] == sess.query(User).order_by(User.id).group_by(User).join('addresses').having(func.count(Address.id)> 2).all() + assert [User(name='ed',id=8)] == sess.query(User).order_by(User.id).group_by(User).join('addresses').having(func.count(Address.id)> 2).all() - assert [User(name=u'jack',id=7), User(name=u'fred',id=9)] == sess.query(User).order_by(User.id).group_by(User).join('addresses').having(func.count(Address.id)< 2).all() + assert [User(name='jack',id=7), User(name='fred',id=9)] == sess.query(User).order_by(User.id).group_by(User).join('addresses').having(func.count(Address.id)< 2).all() class ExistsTest(QueryTest, AssertsCompiledSQL): @@ -1798,14 +1799,14 @@ class YieldTest(QueryTest): ret = [] eq_(len(sess.identity_map), 0) - ret.append(q.next()) - ret.append(q.next()) + ret.append(next(q)) + ret.append(next(q)) eq_(len(sess.identity_map), 2) - ret.append(q.next()) - ret.append(q.next()) + ret.append(next(q)) + ret.append(next(q)) eq_(len(sess.identity_map), 4) try: - q.next() + next(q) assert False except StopIteration: pass @@ -1915,7 +1916,7 @@ class TextTest(QueryTest): User.id, text("users.name")) eq_(s.query(User.id, "name").order_by(User.id).all(), - [(7, u'jack'), (8, u'ed'), (9, u'fred'), (10, u'chuck')]) + [(7, 'jack'), (8, 'ed'), (9, 'fred'), (10, 'chuck')]) def test_via_select(self): User = self.classes.User @@ -1973,7 +1974,7 @@ class ParentTest(QueryTest, AssertsCompiledSQL): try: q = sess.query(Item).with_parent(u1) assert False - except sa_exc.InvalidRequestError, e: + except sa_exc.InvalidRequestError as e: assert str(e) \ == "Could not locate a property which relates "\ "instances of class 'Item' to instances of class 'User'" @@ -2058,7 +2059,7 @@ class ParentTest(QueryTest, AssertsCompiledSQL): "addresses.id AS addresses_id, addresses.user_id AS " "addresses_user_id, addresses.email_address AS addresses_email_address " "FROM addresses WHERE :param_2 = addresses.user_id) AS anon_1", - checkparams={u'param_1': 7, u'param_2': 8}, + checkparams={'param_1': 7, 'param_2': 8}, ) def test_unique_binds_or(self): @@ -2075,7 +2076,7 @@ class ParentTest(QueryTest, AssertsCompiledSQL): "addresses_user_id, addresses.email_address AS " "addresses_email_address FROM addresses WHERE " ":param_1 = addresses.user_id OR :param_2 = addresses.user_id", - checkparams={u'param_1': 7, u'param_2': 8}, + checkparams={'param_1': 7, 'param_2': 8}, ) class SynonymTest(QueryTest): @@ -2117,9 +2118,9 @@ class SynonymTest(QueryTest): options(joinedload(User.orders_syn)).all() eq_(result, [ User(id=7, name='jack', orders=[ - Order(description=u'order 1'), - Order(description=u'order 3'), - Order(description=u'order 5') + Order(description='order 1'), + Order(description='order 3'), + Order(description='order 5') ]) ]) self.assert_sql_count(testing.db, go, 1) @@ -2133,9 +2134,9 @@ class SynonymTest(QueryTest): options(joinedload(User.orders_syn_2)).all() eq_(result, [ User(id=7, name='jack', orders=[ - Order(description=u'order 1'), - Order(description=u'order 3'), - Order(description=u'order 5') + Order(description='order 1'), + Order(description='order 3'), + Order(description='order 5') ]) ]) self.assert_sql_count(testing.db, go, 1) @@ -2149,9 +2150,9 @@ class SynonymTest(QueryTest): options(joinedload('orders_syn_2')).all() eq_(result, [ User(id=7, name='jack', orders=[ - Order(description=u'order 1'), - Order(description=u'order 3'), - Order(description=u'order 5') + Order(description='order 1'), + Order(description='order 3'), + Order(description='order 5') ]) ]) self.assert_sql_count(testing.db, go, 1) @@ -2355,7 +2356,7 @@ class OptionsTest(QueryTest): if isinstance(item, type): item = class_mapper(item) else: - if isinstance(item, basestring): + if isinstance(item, str): item = inspect(r[-1]).mapper.attrs[item] r.append(item) return tuple(r) diff --git a/test/orm/test_relationships.py b/test/orm/test_relationships.py index e5789eb7c..124202dcc 100644 --- a/test/orm/test_relationships.py +++ b/test/orm/test_relationships.py @@ -426,13 +426,13 @@ class CompositeSelfRefFKTest(fixtures.MappedTest): c1 = Company('c1') c2 = Company('c2') - e1 = Employee(u'emp1', c1, 1) - e2 = Employee(u'emp2', c1, 2, e1) - e3 = Employee(u'emp3', c1, 3, e1) - e4 = Employee(u'emp4', c1, 4, e3) - e5 = Employee(u'emp5', c2, 1) - e6 = Employee(u'emp6', c2, 2, e5) - e7 = Employee(u'emp7', c2, 3, e5) + e1 = Employee('emp1', c1, 1) + e2 = Employee('emp2', c1, 2, e1) + e3 = Employee('emp3', c1, 3, e1) + e4 = Employee('emp4', c1, 4, e3) + e5 = Employee('emp5', c2, 1) + e6 = Employee('emp6', c2, 2, e5) + e7 = Employee('emp7', c2, 3, e5) sess.add_all((c1, c2)) sess.commit() @@ -642,7 +642,7 @@ class FKsAsPksTest(fixtures.MappedTest): try: sess.flush() assert False - except AssertionError, e: + except AssertionError as e: startswith_(str(e), "Dependency rule tried to blank-out " "primary key column 'tableB.id' on instance ") @@ -667,7 +667,7 @@ class FKsAsPksTest(fixtures.MappedTest): try: sess.flush() assert False - except AssertionError, e: + except AssertionError as e: startswith_(str(e), "Dependency rule tried to blank-out " "primary key column 'tableB.id' on instance ") @@ -1106,9 +1106,9 @@ class AmbiguousJoinInterpretedAsSelfRef(fixtures.MappedTest): eq_( sess.query(Subscriber).order_by(Subscriber.type).all(), [ - Subscriber(id=1, type=u'A'), - Subscriber(id=2, type=u'B'), - Subscriber(id=2, type=u'C') + Subscriber(id=1, type='A'), + Subscriber(id=2, type='B'), + Subscriber(id=2, type='C') ] ) @@ -1365,7 +1365,7 @@ class TypeMatchTest(fixtures.MappedTest): try: sess.add(a1) assert False - except AssertionError, err: + except AssertionError as err: eq_(str(err), "Attribute 'bs' on class '%s' doesn't handle " "objects of type '%s'" % (A, C)) diff --git a/test/orm/test_session.py b/test/orm/test_session.py index 7c2e8a3b8..45164483b 100644 --- a/test/orm/test_session.py +++ b/test/orm/test_session.py @@ -232,7 +232,7 @@ class ExecutionTest(_fixtures.FixtureTest): # use :bindparam style eq_(sess.execute("select * from users where id=:id", {'id': 7}).fetchall(), - [(7, u'jack')]) + [(7, 'jack')]) # use :bindparam style @@ -733,7 +733,7 @@ class SessionStateTest(_fixtures.FixtureTest): # withstand a change? should this be # more directly attempting to manipulate the identity_map ? u1, u2, u3 = sess.query(User).all() - for i, (key, value) in enumerate(sess.identity_map.iteritems()): + for i, (key, value) in enumerate(iter(sess.identity_map.items())): if i == 2: del u3 gc_collect() @@ -747,7 +747,7 @@ class SessionStateTest(_fixtures.FixtureTest): @event.listens_for(m, "after_update") def e(mapper, conn, target): sess = object_session(target) - for entry in sess.identity_map.values(): + for entry in list(sess.identity_map.values()): entry.name = "5" a1, a2 = User(name="1"), User(name="2") @@ -845,7 +845,7 @@ class SessionStateWFixtureTest(_fixtures.FixtureTest): u = session.query(User).filter_by(id=7).one() # get everything to load in both directions - print [a.user for a in u.addresses] + print([a.user for a in u.addresses]) # then see if expunge fails session.expunge(u) @@ -1187,7 +1187,7 @@ class StrongIdentityMapTest(_fixtures.FixtureTest): s.flush() user = s.query(User).one() user = None - print s.identity_map + print(s.identity_map) gc_collect() assert len(s.identity_map) == 1 @@ -1207,7 +1207,7 @@ class StrongIdentityMapTest(_fixtures.FixtureTest): s = create_session(weak_identity_map=False) mapper(User, users) - for o in [User(name='u%s' % x) for x in xrange(10)]: + for o in [User(name='u%s' % x) for x in range(10)]: s.add(o) # o is still live after this loop... diff --git a/test/orm/test_subquery_relations.py b/test/orm/test_subquery_relations.py index 3ee94cae9..d493e0b7e 100644 --- a/test/orm/test_subquery_relations.py +++ b/test/orm/test_subquery_relations.py @@ -1552,9 +1552,9 @@ class CyclicalInheritingEagerTestTwo(fixtures.DeclarativeMappedTest, Movie = self.classes.Movie session = Session(testing.db) - rscott = Director(name=u"Ridley Scott") - alien = Movie(title=u"Alien") - brunner = Movie(title=u"Blade Runner") + rscott = Director(name="Ridley Scott") + alien = Movie(title="Alien") + brunner = Movie(title="Blade Runner") rscott.movies.append(brunner) rscott.movies.append(alien) session.add_all([rscott, alien, brunner]) diff --git a/test/orm/test_unitofwork.py b/test/orm/test_unitofwork.py index 6be1672e1..4b9799d47 100644 --- a/test/orm/test_unitofwork.py +++ b/test/orm/test_unitofwork.py @@ -6,6 +6,7 @@ import datetime from sqlalchemy.orm import mapper as orm_mapper import sqlalchemy as sa +from sqlalchemy.util import u, ue, b from sqlalchemy import Integer, String, ForeignKey, literal_column, event from sqlalchemy.testing import engines from sqlalchemy import testing @@ -87,7 +88,7 @@ class UnicodeTest(fixtures.MappedTest): mapper(Test, uni_t1) - txt = u"\u0160\u0110\u0106\u010c\u017d" + txt = ue("\u0160\u0110\u0106\u010c\u017d") t1 = Test(id=1, txt=txt) self.assert_(t1.txt == txt) @@ -107,7 +108,7 @@ class UnicodeTest(fixtures.MappedTest): 't2s': relationship(Test2)}) mapper(Test2, uni_t2) - txt = u"\u0160\u0110\u0106\u010c\u017d" + txt = ue("\u0160\u0110\u0106\u010c\u017d") t1 = Test(txt=txt) t1.t2s.append(Test2()) t1.t2s.append(Test2()) @@ -132,16 +133,16 @@ class UnicodeSchemaTest(fixtures.MappedTest): @classmethod def define_tables(cls, metadata): t1 = Table('unitable1', metadata, - Column(u'méil', Integer, primary_key=True, key='a', test_needs_autoincrement=True), - Column(u'\u6e2c\u8a66', Integer, key='b'), + Column(u('méil'), Integer, primary_key=True, key='a', test_needs_autoincrement=True), + Column(ue('\u6e2c\u8a66'), Integer, key='b'), Column('type', String(20)), test_needs_fk=True, test_needs_autoincrement=True) - t2 = Table(u'Unitéble2', metadata, - Column(u'méil', Integer, primary_key=True, key="cc", test_needs_autoincrement=True), - Column(u'\u6e2c\u8a66', Integer, - ForeignKey(u'unitable1.a'), key="d"), - Column(u'\u6e2c\u8a66_2', Integer, key="e"), + t2 = Table(u('Unitéble2'), metadata, + Column(u('méil'), Integer, primary_key=True, key="cc", test_needs_autoincrement=True), + Column(ue('\u6e2c\u8a66'), Integer, + ForeignKey('unitable1.a'), key="d"), + Column(ue('\u6e2c\u8a66_2'), Integer, key="e"), test_needs_fk=True, test_needs_autoincrement=True) @@ -237,12 +238,7 @@ class BinaryHistTest(fixtures.MappedTest, testing.AssertsExecutionResults): def test_binary_equality(self): Foo, t1 = self.classes.Foo, self.tables.t1 - - # Py3K - #data = b"this is some data" - # Py2K - data = "this is some data" - # end Py2K + data = b("this is some data") mapper(Foo, t1) @@ -1054,13 +1050,13 @@ class OneToManyTest(_fixtures.FixtureTest): session.flush() user_rows = users.select(users.c.id.in_([u.id])).execute().fetchall() - eq_(user_rows[0].values(), [u.id, 'one2manytester']) + eq_(list(user_rows[0].values()), [u.id, 'one2manytester']) address_rows = addresses.select( addresses.c.id.in_([a.id, a2.id]), order_by=[addresses.c.email_address]).execute().fetchall() - eq_(address_rows[0].values(), [a2.id, u.id, 'lala@test.org']) - eq_(address_rows[1].values(), [a.id, u.id, 'one2many@test.org']) + eq_(list(address_rows[0].values()), [a2.id, u.id, 'lala@test.org']) + eq_(list(address_rows[1].values()), [a.id, u.id, 'one2many@test.org']) userid = u.id addressid = a2.id @@ -1071,7 +1067,7 @@ class OneToManyTest(_fixtures.FixtureTest): address_rows = addresses.select( addresses.c.id == addressid).execute().fetchall() - eq_(address_rows[0].values(), + eq_(list(address_rows[0].values()), [addressid, userid, 'somethingnew@foo.com']) self.assert_(u.id == userid and a2.id == addressid) @@ -1501,18 +1497,18 @@ class SaveTest(_fixtures.FixtureTest): assert u.name == 'multitester' user_rows = users.select(users.c.id.in_([u.foo_id])).execute().fetchall() - eq_(user_rows[0].values(), [u.foo_id, 'multitester']) + eq_(list(user_rows[0].values()), [u.foo_id, 'multitester']) address_rows = addresses.select(addresses.c.id.in_([u.id])).execute().fetchall() - eq_(address_rows[0].values(), [u.id, u.foo_id, 'multi@test.org']) + eq_(list(address_rows[0].values()), [u.id, u.foo_id, 'multi@test.org']) u.email = 'lala@hey.com' u.name = 'imnew' session.flush() user_rows = users.select(users.c.id.in_([u.foo_id])).execute().fetchall() - eq_(user_rows[0].values(), [u.foo_id, 'imnew']) + eq_(list(user_rows[0].values()), [u.foo_id, 'imnew']) address_rows = addresses.select(addresses.c.id.in_([u.id])).execute().fetchall() - eq_(address_rows[0].values(), [u.id, u.foo_id, 'lala@hey.com']) + eq_(list(address_rows[0].values()), [u.id, u.foo_id, 'lala@hey.com']) session.expunge_all() u = session.query(User).get(id) @@ -1650,7 +1646,7 @@ class ManyToOneTest(_fixtures.FixtureTest): l = sa.select([users, addresses], sa.and_(users.c.id==addresses.c.user_id, addresses.c.id==a.id)).execute() - eq_(l.first().values(), + eq_(list(l.first().values()), [a.user.id, 'asdf8d', a.id, a.user_id, 'theater@foo.com']) def test_many_to_one_1(self): @@ -2127,7 +2123,6 @@ class SaveTest3(fixtures.MappedTest): assert assoc.count().scalar() == 2 i.keywords = [] - print i.keywords session.flush() assert assoc.count().scalar() == 0 diff --git a/test/orm/test_unitofworkv2.py b/test/orm/test_unitofworkv2.py index 7c7337c62..34648c256 100644 --- a/test/orm/test_unitofworkv2.py +++ b/test/orm/test_unitofworkv2.py @@ -26,7 +26,7 @@ class AssertsUOW(object): def _assert_uow_size(self, session, expected ): uow = self._get_test_uow(session) postsort_actions = uow._generate_actions() - print postsort_actions + print(postsort_actions) eq_(len(postsort_actions), expected, postsort_actions) class UOWTest(_fixtures.FixtureTest, @@ -125,12 +125,12 @@ class RudimentaryFlushTest(UOWTest): CompiledSQL( "UPDATE addresses SET user_id=:user_id WHERE " "addresses.id = :addresses_id", - lambda ctx: [{u'addresses_id': a1.id, 'user_id': None}] + lambda ctx: [{'addresses_id': a1.id, 'user_id': None}] ), CompiledSQL( "UPDATE addresses SET user_id=:user_id WHERE " "addresses.id = :addresses_id", - lambda ctx: [{u'addresses_id': a2.id, 'user_id': None}] + lambda ctx: [{'addresses_id': a2.id, 'user_id': None}] ), CompiledSQL( "DELETE FROM users WHERE users.id = :id", @@ -235,12 +235,12 @@ class RudimentaryFlushTest(UOWTest): CompiledSQL( "UPDATE addresses SET user_id=:user_id WHERE " "addresses.id = :addresses_id", - lambda ctx: [{u'addresses_id': a1.id, 'user_id': None}] + lambda ctx: [{'addresses_id': a1.id, 'user_id': None}] ), CompiledSQL( "UPDATE addresses SET user_id=:user_id WHERE " "addresses.id = :addresses_id", - lambda ctx: [{u'addresses_id': a2.id, 'user_id': None}] + lambda ctx: [{'addresses_id': a2.id, 'user_id': None}] ), CompiledSQL( "DELETE FROM users WHERE users.id = :id", @@ -1149,7 +1149,7 @@ class SingleCycleM2MTest(fixtures.MappedTest, "nodes, node_to_nodes WHERE :param_1 = " "node_to_nodes.right_node_id AND nodes.id = " "node_to_nodes.left_node_id" , - lambda ctx:{u'param_1': n1.id}, + lambda ctx:{'param_1': n1.id}, ), CompiledSQL( "DELETE FROM node_to_nodes WHERE " diff --git a/test/orm/test_update_delete.py b/test/orm/test_update_delete.py index 65d69538e..6915ac8a2 100644 --- a/test/orm/test_update_delete.py +++ b/test/orm/test_update_delete.py @@ -188,22 +188,22 @@ class UpdateDeleteTest(fixtures.MappedTest): update({'age': User.age - 10}, synchronize_session='evaluate') eq_([john.age, jack.age, jill.age, jane.age], [25,37,29,27]) - eq_(sess.query(User.age).order_by(User.id).all(), zip([25,37,29,27])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([25,37,29,27]))) sess.query(User).filter(User.age > 29).\ update({User.age: User.age - 10}, synchronize_session='evaluate') eq_([john.age, jack.age, jill.age, jane.age], [25,27,29,27]) - eq_(sess.query(User.age).order_by(User.id).all(), zip([25,27,29,27])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([25,27,29,27]))) sess.query(User).filter(User.age > 27).\ update({users.c.age: User.age - 10}, synchronize_session='evaluate') eq_([john.age, jack.age, jill.age, jane.age], [25,27,19,27]) - eq_(sess.query(User.age).order_by(User.id).all(), zip([25,27,19,27])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([25,27,19,27]))) sess.query(User).filter(User.age == 25).\ update({User.age: User.age - 10}, synchronize_session='fetch') eq_([john.age, jack.age, jill.age, jane.age], [15,27,19,27]) - eq_(sess.query(User.age).order_by(User.id).all(), zip([15,27,19,27])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([15,27,19,27]))) def test_update_against_metadata(self): User, users = self.classes.User, self.tables.users @@ -211,7 +211,7 @@ class UpdateDeleteTest(fixtures.MappedTest): sess = Session() sess.query(users).update({users.c.age: 29}, synchronize_session=False) - eq_(sess.query(User.age).order_by(User.id).all(), zip([29,29,29,29])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([29,29,29,29]))) def test_update_with_bindparams(self): User = self.classes.User @@ -224,7 +224,7 @@ class UpdateDeleteTest(fixtures.MappedTest): update({'age': User.age - 10}, synchronize_session='fetch') eq_([john.age, jack.age, jill.age, jane.age], [25,37,29,27]) - eq_(sess.query(User.age).order_by(User.id).all(), zip([25,37,29,27])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([25,37,29,27]))) def test_update_without_load(self): User = self.classes.User @@ -233,7 +233,7 @@ class UpdateDeleteTest(fixtures.MappedTest): sess.query(User).filter(User.id == 3).\ update({'age': 44}, synchronize_session='fetch') - eq_(sess.query(User.age).order_by(User.id).all(), zip([25,47,44,37])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([25,47,44,37]))) def test_update_changes_resets_dirty(self): User = self.classes.User @@ -300,7 +300,7 @@ class UpdateDeleteTest(fixtures.MappedTest): update({'age': User.age - 10}, synchronize_session='fetch') eq_([john.age, jack.age, jill.age, jane.age], [25,37,29,27]) - eq_(sess.query(User.age).order_by(User.id).all(), zip([25,37,29,27])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([25,37,29,27]))) @testing.fails_if(lambda: not testing.db.dialect.supports_sane_rowcount) def test_update_returns_rowcount(self): @@ -334,7 +334,7 @@ class UpdateDeleteTest(fixtures.MappedTest): sess.query(User).update({'age': 42}, synchronize_session='evaluate') eq_([john.age, jack.age, jill.age, jane.age], [42,42,42,42]) - eq_(sess.query(User.age).order_by(User.id).all(), zip([42,42,42,42])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([42,42,42,42]))) def test_delete_all(self): User = self.classes.User @@ -516,7 +516,7 @@ class UpdateDeleteIgnoresLoadersTest(fixtures.MappedTest): eq_([foo.title, bar.title, baz.title], ['foofoo','barbar', 'baz']) eq_(sess.query(Document.title).order_by(Document.id).all(), - zip(['foofoo','barbar', 'baz'])) + list(zip(['foofoo','barbar', 'baz']))) def test_update_with_explicit_joinedload(self): User = self.classes.User @@ -528,7 +528,7 @@ class UpdateDeleteIgnoresLoadersTest(fixtures.MappedTest): update({'age': User.age - 10}, synchronize_session='fetch') eq_([john.age, jack.age, jill.age, jane.age], [25,37,29,27]) - eq_(sess.query(User.age).order_by(User.id).all(), zip([25,37,29,27])) + eq_(sess.query(User.age).order_by(User.id).all(), list(zip([25,37,29,27]))) def test_delete_with_eager_relationships(self): Document = self.classes.Document @@ -538,7 +538,7 @@ class UpdateDeleteIgnoresLoadersTest(fixtures.MappedTest): sess.query(Document).filter(Document.user_id == 1).\ delete(synchronize_session=False) - eq_(sess.query(Document.title).all(), zip(['baz'])) + eq_(sess.query(Document.title).all(), list(zip(['baz']))) class UpdateDeleteFromTest(fixtures.MappedTest): @classmethod diff --git a/test/orm/test_utils.py b/test/orm/test_utils.py index b2853a8b8..49ff61ea0 100644 --- a/test/orm/test_utils.py +++ b/test/orm/test_utils.py @@ -52,10 +52,10 @@ class AliasedClassTest(fixtures.TestBase): alias = aliased(Point) assert Point.zero - # Py2K - # TODO: what is this testing ?? - assert not getattr(alias, 'zero') - # end Py2K +# start Py2K +# # TODO: what is this testing ?? +# assert not getattr(alias, 'zero') +# end Py2K def test_classmethods(self): class Point(object): @@ -123,17 +123,19 @@ class AliasedClassTest(fixtures.TestBase): self.func = func def __get__(self, instance, owner): if instance is None: - # Py3K - #args = (self.func, owner) - # Py2K - args = (self.func, owner, owner.__class__) - # end Py2K +# start Py3K + args = (self.func, owner) +# end Py3K +# start Py2K +# args = (self.func, owner, owner.__class__) +# end Py2K else: - # Py3K - #args = (self.func, instance) - # Py2K - args = (self.func, instance, owner) - # end Py2K +# start Py3K + args = (self.func, instance) +# end Py3K +# start Py2K +# args = (self.func, instance, owner) +# end Py2K return types.MethodType(*args) class PropertyDescriptor(object): @@ -368,13 +370,6 @@ class PathRegistryTest(_fixtures.FixtureTest): assert p1.contains_mapper(umapper) assert not p1.contains_mapper(amapper) - def _registry(self): - class Reg(dict): - @property - def _attributes(self): - return self - return Reg() - def test_path(self): umapper = inspect(self.classes.User) amapper = inspect(self.classes.Address) @@ -394,7 +389,7 @@ class PathRegistryTest(_fixtures.FixtureTest): ) def test_registry_set(self): - reg = self._registry() + reg = {} umapper = inspect(self.classes.User) amapper = inspect(self.classes.Address) @@ -415,7 +410,7 @@ class PathRegistryTest(_fixtures.FixtureTest): ) def test_registry_get(self): - reg = self._registry() + reg = {} umapper = inspect(self.classes.User) amapper = inspect(self.classes.Address) @@ -437,7 +432,7 @@ class PathRegistryTest(_fixtures.FixtureTest): eq_(p3.get(reg, "p1key"), None) def test_registry_contains(self): - reg = self._registry() + reg = {} umapper = inspect(self.classes.User) amapper = inspect(self.classes.Address) @@ -457,7 +452,7 @@ class PathRegistryTest(_fixtures.FixtureTest): assert not p2.contains(reg, "fake") def test_registry_setdefault(self): - reg = self._registry() + reg = {} umapper = inspect(self.classes.User) amapper = inspect(self.classes.Address) diff --git a/test/perf/README b/test/perf/README deleted file mode 100644 index 1d03cb5ce..000000000 --- a/test/perf/README +++ /dev/null @@ -1,17 +0,0 @@ -This directory contains informal scripts used to stress test various -library subsections over the years, including testing of memory usage, -function call count, threading behavior. - -The scripts here are *not* part of the automated test suite, and instead -were used at the time of development for particular features or -performance enhancements in an ad-hoc fashion. Ideally -the various functionality tested within would be brought under the -umbrella of controlled, automated tests. Many of the scripts here -are out of date and are possibly covered by formal performance tests -elsewhere. - -Current automated stress and performance tests are in test/aaa_profiling/, -which test either for expected function call count, or flat growth in memory -usage over time. These tests are part of the automated test suite -and are maintained for 100% success rate along Python versions from 2.4 through -current 3 versions.
\ No newline at end of file diff --git a/test/perf/insertspeed.py b/test/perf/insertspeed.py deleted file mode 100644 index 03d2c4144..000000000 --- a/test/perf/insertspeed.py +++ /dev/null @@ -1,109 +0,0 @@ -import sys, time -from sqlalchemy import * -from sqlalchemy.orm import * -from sqlalchemy.testing import profiling - -db = create_engine('sqlite://') -metadata = MetaData(db) -Person_table = Table('Person', metadata, - Column('name', String(40)), - Column('sex', Integer), - Column('age', Integer)) - - -def sa_unprofiled_insertmany(n): - i = Person_table.insert() - i.execute([{'name':'John Doe','sex':1,'age':35} for j in xrange(n)]) - -def sqlite_unprofiled_insertmany(n): - conn = db.connect().connection - c = conn.cursor() - persons = [('john doe', 1, 35) for i in xrange(n)] - c.executemany("insert into Person(name, sex, age) values (?,?,?)", persons) - -@profiling.profiled('sa_profiled_insert_many', always=True) -def sa_profiled_insert_many(n): - i = Person_table.insert() - i.execute([{'name':'John Doe','sex':1,'age':35} for j in xrange(n)]) - s = Person_table.select() - r = s.execute() - res = [[value for value in row] for row in r.fetchall()] - -def sqlite_unprofiled_insert(n): - conn = db.connect().connection - c = conn.cursor() - for j in xrange(n): - c.execute("insert into Person(name, sex, age) values (?,?,?)", - ('john doe', 1, 35)) - -def sa_unprofiled_insert(n): - # Another option is to build Person_table.insert() outside of the - # loop. But it doesn't make much of a difference, so might as well - # use the worst-case/naive version here. - for j in xrange(n): - Person_table.insert().execute({'name':'John Doe','sex':1,'age':35}) - -@profiling.profiled('sa_profiled_insert', always=True) -def sa_profiled_insert(n): - i = Person_table.insert() - for j in xrange(n): - i.execute({'name':'John Doe','sex':1,'age':35}) - s = Person_table.select() - r = s.execute() - res = [[value for value in row] for row in r.fetchall()] - -def run_timed(fn, label, *args, **kw): - metadata.drop_all() - metadata.create_all() - - sys.stdout.write("%s (%s): " % (label, ', '.join([str(a) for a in args]))) - sys.stdout.flush() - - t = time.clock() - fn(*args, **kw) - t2 = time.clock() - - sys.stdout.write("%0.2f seconds\n" % (t2 - t)) - -def run_profiled(fn, label, *args, **kw): - metadata.drop_all() - metadata.create_all() - - print "%s (%s)" % (label, ', '.join([str(a) for a in args])) - fn(*args, **kw) - -def all(): - try: - print "Bulk INSERTS via executemany():\n" - - run_timed(sqlite_unprofiled_insertmany, - 'pysqlite bulk insert', - 50000) - - run_timed(sa_unprofiled_insertmany, - 'SQLAlchemy bulk insert', - 50000) - - run_profiled(sa_profiled_insert_many, - 'SQLAlchemy bulk insert/select, profiled', - 50000) - - print "\nIndividual INSERTS via execute():\n" - - run_timed(sqlite_unprofiled_insert, - "pysqlite individual insert", - 50000) - - run_timed(sa_unprofiled_insert, - "SQLAlchemy individual insert", - 50000) - - run_profiled(sa_profiled_insert, - 'SQLAlchemy individual insert/select, profiled', - 50000) - - finally: - metadata.drop_all() - -if __name__ == '__main__': - all() diff --git a/test/perf/large_flush.py b/test/perf/large_flush.py deleted file mode 100644 index b1ecce852..000000000 --- a/test/perf/large_flush.py +++ /dev/null @@ -1,84 +0,0 @@ -import sqlalchemy as sa -from sqlalchemy import create_engine, MetaData, orm -from sqlalchemy import Column, ForeignKey -from sqlalchemy import Integer, String -from sqlalchemy.orm import mapper -from sqlalchemy.testing import profiling - -class Object(object): - pass - -class Q(Object): - pass - -class A(Object): - pass - -class C(Object): - pass - -class WC(C): - pass - -engine = create_engine('sqlite:///:memory:', echo=True) - -sm = orm.sessionmaker(bind=engine) - -SA_Session = orm.scoped_session(sm) - -SA_Metadata = MetaData() - -object_table = sa.Table('Object', - SA_Metadata, - Column('ObjectID', Integer,primary_key=True), - Column('Type', String(1), nullable=False)) - -q_table = sa.Table('Q', - SA_Metadata, - Column('QID', Integer, ForeignKey('Object.ObjectID'),primary_key=True)) - -c_table = sa.Table('C', - SA_Metadata, - Column('CID', Integer, ForeignKey('Object.ObjectID'),primary_key=True)) - -wc_table = sa.Table('WC', - SA_Metadata, - Column('WCID', Integer, ForeignKey('C.CID'), primary_key=True)) - -a_table = sa.Table('A', - SA_Metadata, - Column('AID', Integer, ForeignKey('Object.ObjectID'),primary_key=True), - Column('QID', Integer, ForeignKey('Q.QID')), - Column('CID', Integer, ForeignKey('C.CID'))) - -mapper(Object, object_table, polymorphic_on=object_table.c.Type, polymorphic_identity='O') - -mapper(Q, q_table, inherits=Object, polymorphic_identity='Q') -mapper(C, c_table, inherits=Object, polymorphic_identity='C') -mapper(WC, wc_table, inherits=C, polymorphic_identity='W') - -mapper(A, a_table, inherits=Object, polymorphic_identity='A', - properties = { - 'Q' : orm.relation(Q,primaryjoin=a_table.c.QID==q_table.c.QID, - backref='As' - ), - 'C' : orm.relation(C,primaryjoin=a_table.c.CID==c_table.c.CID, - backref='A', - uselist=False) - } - ) - -SA_Metadata.create_all(engine) - -@profiling.profiled('large_flush', always=True, sort=['file']) -def generate_error(): - q = Q() - for j in range(100): #at 306 the error does not pop out (depending on recursion depth) - a = A() - a.Q = q - a.C = WC() - - SA_Session.add(q) - SA_Session.commit() #here the error pops out - -generate_error()
\ No newline at end of file diff --git a/test/perf/objselectspeed.py b/test/perf/objselectspeed.py deleted file mode 100644 index c0ed88444..000000000 --- a/test/perf/objselectspeed.py +++ /dev/null @@ -1,146 +0,0 @@ -import time, resource -from sqlalchemy import * -from sqlalchemy.orm import * -from sqlalchemy.testing.util import gc_collect -from sqlalchemy.testing import profiling - -db = create_engine('sqlite://') -metadata = MetaData(db) -Person_table = Table('Person', metadata, - Column('id', Integer, primary_key=True), - Column('type', String(10)), - Column('name', String(40)), - Column('sex', Integer), - Column('age', Integer)) - - -Employee_table = Table('Employee', metadata, - Column('id', Integer, ForeignKey('Person.id'), primary_key=True), - Column('foo', String(40)), - Column('bar', Integer), - Column('bat', Integer)) - -class RawPerson(object): pass -class Person(object): pass -mapper(Person, Person_table) - -class JoinedPerson(object):pass -class Employee(JoinedPerson):pass -mapper(JoinedPerson, Person_table, \ - polymorphic_on=Person_table.c.type, polymorphic_identity='person') -mapper(Employee, Employee_table, \ - inherits=JoinedPerson, polymorphic_identity='employee') -compile_mappers() - -def setup(): - metadata.create_all() - i = Person_table.insert() - data = [{'name':'John Doe','sex':1,'age':35, 'type':'employee'}] * 100 - for j in xrange(500): - i.execute(data) - - # note we arent fetching from employee_table, - # so we can leave it empty even though its "incorrect" - #i = Employee_table.insert() - #data = [{'foo':'foo', 'bar':'bar':'bat':'bat'}] * 100 - #for j in xrange(500): - # i.execute(data) - - print "Inserted 50,000 rows" - -def sqlite_select(entity_cls): - conn = db.connect().connection - cr = conn.cursor() - cr.execute("SELECT id, name, sex, age FROM Person") - people = [] - for row in cr.fetchall(): - person = entity_cls() - person.id = row[0] - person.name = row[1] - person.sex = row[2] - person.age = row[3] - people.append(person) - cr.close() - conn.close() - -def sql_select(entity_cls): - people = [] - for row in Person_table.select().execute().fetchall(): - person = entity_cls() - person.id = row['id'] - person.name = row['name'] - person.sex = row['sex'] - person.age = row['age'] - people.append(person) - -#@profiling.profiled(report=True, always=True) -def orm_select(): - session = create_session() - people = session.query(Person).all() - -#@profiling.profiled(report=True, always=True) -def joined_orm_select(): - session = create_session() - people = session.query(JoinedPerson).all() - -def all(): - setup() - try: - t, t2 = 0, 0 - def usage(label): - now = resource.getrusage(resource.RUSAGE_SELF) - print "%s: %0.3fs real, %0.3fs user, %0.3fs sys" % ( - label, t2 - t, - now.ru_utime - usage.last.ru_utime, - now.ru_stime - usage.last.ru_stime) - usage.snap(now) - usage.snap = lambda stats=None: setattr( - usage, 'last', stats or resource.getrusage(resource.RUSAGE_SELF)) - - gc_collect() - usage.snap() - t = time.clock() - sqlite_select(RawPerson) - t2 = time.clock() - usage('sqlite select/native') - - gc_collect() - usage.snap() - t = time.clock() - sqlite_select(Person) - t2 = time.clock() - usage('sqlite select/instrumented') - - gc_collect() - usage.snap() - t = time.clock() - sql_select(RawPerson) - t2 = time.clock() - usage('sqlalchemy.sql select/native') - - gc_collect() - usage.snap() - t = time.clock() - sql_select(Person) - t2 = time.clock() - usage('sqlalchemy.sql select/instrumented') - - gc_collect() - usage.snap() - t = time.clock() - orm_select() - t2 = time.clock() - usage('sqlalchemy.orm fetch') - - gc_collect() - usage.snap() - t = time.clock() - joined_orm_select() - t2 = time.clock() - usage('sqlalchemy.orm "joined" fetch') - finally: - metadata.drop_all() - - -if __name__ == '__main__': - all() diff --git a/test/perf/objupdatespeed.py b/test/perf/objupdatespeed.py deleted file mode 100644 index 98d10180e..000000000 --- a/test/perf/objupdatespeed.py +++ /dev/null @@ -1,94 +0,0 @@ -import time, resource -from sqlalchemy import * -from sqlalchemy.orm import * -from sqlalchemy.testing import * -from sqlalchemy.testing.util import gc_collect - - -NUM = 100 - -metadata = MetaData(testing.db) -Person_table = Table('Person', metadata, - Column('id', Integer, primary_key=True), - Column('name', String(40)), - Column('sex', Integer), - Column('age', Integer)) - -Email_table = Table('Email', metadata, - Column('id', Integer, primary_key=True), - Column('person_id', Integer, ForeignKey('Person.id')), - Column('address', String(300))) - -class Person(object): - pass -class Email(object): - def __repr__(self): - return '<email %s %s>' % (getattr(self, 'id', None), - getattr(self, 'address', None)) - -mapper(Person, Person_table, properties={ - 'emails': relationship(Email, backref='owner', lazy='joined') - }) -mapper(Email, Email_table) -compile_mappers() - -def setup(): - metadata.create_all() - i = Person_table.insert() - data = [{'name':'John Doe','sex':1,'age':35}] * NUM - i.execute(data) - - i = Email_table.insert() - for j in xrange(1, NUM + 1): - i.execute(address='foo@bar', person_id=j) - if j % 2: - i.execute(address='baz@quux', person_id=j) - - print "Inserted %d rows." % (NUM + NUM + (NUM // 2)) - -def orm_select(session): - return session.query(Person).all() - -@profiling.profiled('update_and_flush') -def update_and_flush(session, people): - for p in people: - p.name = 'Exene Cervenka' - p.sex = 2 - p.emails[0].address = 'hoho@lala' - session.flush() - -def all(): - setup() - try: - t, t2 = 0, 0 - def usage(label): - now = resource.getrusage(resource.RUSAGE_SELF) - print "%s: %0.3fs real, %0.3fs user, %0.3fs sys" % ( - label, t2 - t, - now.ru_utime - usage.last.ru_utime, - now.ru_stime - usage.last.ru_stime) - usage.snap(now) - usage.snap = lambda stats=None: setattr( - usage, 'last', stats or resource.getrusage(resource.RUSAGE_SELF)) - - session = create_session() - - gc_collect() - usage.snap() - t = time.clock() - people = orm_select(session) - t2 = time.clock() - usage('load objects') - - gc_collect() - usage.snap() - t = time.clock() - update_and_flush(session, people) - t2 = time.clock() - usage('update and flush') - finally: - metadata.drop_all() - - -if __name__ == '__main__': - all() diff --git a/test/perf/orm2010.py b/test/perf/orm2010.py index 47257ba87..937e6ddff 100644 --- a/test/perf/orm2010.py +++ b/test/perf/orm2010.py @@ -95,7 +95,7 @@ def runit(): name="Boss %d" % i, golf_average=Decimal(random.randint(40, 150)) ) - for i in xrange(1000) + for i in range(1000) ] sess.add_all(bosses) @@ -107,7 +107,7 @@ def runit(): name="Grunt %d" % i, savings=Decimal(random.randint(5000000, 15000000) / 100) ) - for i in xrange(10000) + for i in range(10000) ] # Assign each Grunt a Boss. Look them up in the DB @@ -149,15 +149,15 @@ stats = pstats.Stats(filename) counts_by_methname = dict((key[2], stats.stats[key][0]) for key in stats.stats) -print "SQLA Version: %s" % __version__ -print "Total calls %d" % stats.total_calls -print "Total cpu seconds: %.2f" % stats.total_tt -print 'Total execute calls: %d' \ +print("SQLA Version: %s" % __version__) +print("Total calls %d" % stats.total_calls) +print("Total cpu seconds: %.2f" % stats.total_tt) +print('Total execute calls: %d' \ % counts_by_methname["<method 'execute' of 'sqlite3.Cursor' " - "objects>"] -print 'Total executemany calls: %d' \ + "objects>"]) +print('Total executemany calls: %d' \ % counts_by_methname.get("<method 'executemany' of 'sqlite3.Cursor' " - "objects>", 0) + "objects>", 0)) #stats.sort_stats('time', 'calls') #stats.print_stats() diff --git a/test/perf/ormsession.py b/test/perf/ormsession.py deleted file mode 100644 index 5e38d6e80..000000000 --- a/test/perf/ormsession.py +++ /dev/null @@ -1,225 +0,0 @@ -import time -from datetime import datetime - -from sqlalchemy import * -from sqlalchemy.orm import * -from sqlalchemy.testing import * -from sqlalchemy.testing.profiling import profiled - -class Item(object): - def __repr__(self): - return 'Item<#%s "%s">' % (self.id, self.name) -class SubItem(object): - def __repr__(self): - return 'SubItem<#%s "%s">' % (self.id, self.name) -class Customer(object): - def __repr__(self): - return 'Customer<#%s "%s">' % (self.id, self.name) -class Purchase(object): - def __repr__(self): - return 'Purchase<#%s "%s">' % (self.id, self.purchase_date) - -items, subitems, customers, purchases, purchaseitems = \ - None, None, None, None, None - -metadata = MetaData() - -@profiled('table') -def define_tables(): - global items, subitems, customers, purchases, purchaseitems - items = Table('items', metadata, - Column('id', Integer, primary_key=True), - Column('name', String(100)), - test_needs_acid=True) - subitems = Table('subitems', metadata, - Column('id', Integer, primary_key=True), - Column('item_id', Integer, ForeignKey('items.id'), - nullable=False), - Column('name', String(100), server_default='no name'), - test_needs_acid=True) - customers = Table('customers', metadata, - Column('id', Integer, primary_key=True), - Column('name', String(100)), - *[Column("col_%s" % chr(i), String(64), default=str(i)) - for i in range(97,117)], - **dict(test_needs_acid=True)) - purchases = Table('purchases', metadata, - Column('id', Integer, primary_key=True), - Column('customer_id', Integer, - ForeignKey('customers.id'), nullable=False), - Column('purchase_date', DateTime, - default=datetime.now), - test_needs_acid=True) - purchaseitems = Table('purchaseitems', metadata, - Column('purchase_id', Integer, - ForeignKey('purchases.id'), - nullable=False, primary_key=True), - Column('item_id', Integer, ForeignKey('items.id'), - nullable=False, primary_key=True), - test_needs_acid=True) - -@profiled('mapper') -def setup_mappers(): - mapper(Item, items, properties={ - 'subitems': relationship(SubItem, backref='item', lazy='select') - }) - mapper(SubItem, subitems) - mapper(Customer, customers, properties={ - 'purchases': relationship(Purchase, lazy='select', backref='customer') - }) - mapper(Purchase, purchases, properties={ - 'items': relationship(Item, lazy='select', secondary=purchaseitems) - }) - -@profiled('inserts') -def insert_data(): - q_items = 1000 - q_sub_per_item = 10 - q_customers = 1000 - - con = testing.db.connect() - - transaction = con.begin() - data, subdata = [], [] - for item_id in xrange(1, q_items + 1): - data.append({'name': "item number %s" % item_id}) - for subitem_id in xrange(1, (item_id % q_sub_per_item) + 1): - subdata.append({'item_id': item_id, - 'name': "subitem number %s" % subitem_id}) - if item_id % 100 == 0: - items.insert().execute(*data) - subitems.insert().execute(*subdata) - del data[:] - del subdata[:] - if data: - items.insert().execute(*data) - if subdata: - subitems.insert().execute(*subdata) - transaction.commit() - - transaction = con.begin() - data = [] - for customer_id in xrange(1, q_customers): - data.append({'name': "customer number %s" % customer_id}) - if customer_id % 100 == 0: - customers.insert().execute(*data) - del data[:] - if data: - customers.insert().execute(*data) - transaction.commit() - - transaction = con.begin() - data, subdata = [], [] - order_t = int(time.time()) - (5000 * 5 * 60) - current = xrange(1, q_customers) - step, purchase_id = 1, 0 - while current: - next = [] - for customer_id in current: - order_t += 300 - data.append({'customer_id': customer_id, - 'purchase_date': datetime.fromtimestamp(order_t)}) - purchase_id += 1 - for item_id in range(customer_id % 200, customer_id + 1, 200): - if item_id != 0: - subdata.append({'purchase_id': purchase_id, - 'item_id': item_id}) - if customer_id % 10 > step: - next.append(customer_id) - - if len(data) >= 100: - purchases.insert().execute(*data) - if subdata: - purchaseitems.insert().execute(*subdata) - del data[:] - del subdata[:] - step, current = step + 1, next - - if data: - purchases.insert().execute(*data) - if subdata: - purchaseitems.insert().execute(*subdata) - transaction.commit() - -@profiled('queries') -def run_queries(): - session = create_session() - # no explicit transaction here. - - # build a report of summarizing the last 50 purchases and - # the top 20 items from all purchases - - q = session.query(Purchase). \ - order_by(desc(Purchase.purchase_date)). \ - limit(50).\ - options(joinedload('items'), joinedload('items.subitems'), - joinedload('customer')) - - report = [] - # "write" the report. pretend it's going to a web template or something, - # the point is to actually pull data through attributes and collections. - for purchase in q: - report.append(purchase.customer.name) - report.append(purchase.customer.col_a) - report.append(purchase.purchase_date) - for item in purchase.items: - report.append(item.name) - report.extend([s.name for s in item.subitems]) - - # mix a little low-level with orm - # pull a report of the top 20 items of all time - _item_id = purchaseitems.c.item_id - top_20_q = select([func.distinct(_item_id).label('id')], - group_by=[purchaseitems.c.purchase_id, _item_id], - order_by=[desc(func.count(_item_id)), _item_id], - limit=20) - ids = [r.id for r in top_20_q.execute().fetchall()] - q2 = session.query(Item).filter(Item.id.in_(ids)) - - for num, item in enumerate(q2): - report.append("number %s: %s" % (num + 1, item.name)) - -@profiled('creating') -def create_purchase(): - # commit a purchase - customer_id = 100 - item_ids = (10,22,34,46,58) - - session = create_session() - session.begin() - - customer = session.query(Customer).get(customer_id) - items = session.query(Item).filter(Item.id.in_(item_ids)) - - purchase = Purchase() - purchase.customer = customer - purchase.items.extend(items) - - session.flush() - session.commit() - session.expire(customer) - -def setup_db(): - metadata.drop_all() - metadata.create_all() -def cleanup_db(): - metadata.drop_all() - -@profiled('default') -def default(): - run_queries() - create_purchase() - -@profiled('all') -def main(): - metadata.bind = testing.db - try: - define_tables() - setup_mappers() - setup_db() - insert_data() - default() - finally: - cleanup_db() - -main() diff --git a/test/perf/sessions.py b/test/perf/sessions.py deleted file mode 100644 index 80553fe48..000000000 --- a/test/perf/sessions.py +++ /dev/null @@ -1,95 +0,0 @@ -from sqlalchemy import * -from sqlalchemy.orm import * - -from sqlalchemy.testing.compat import gc_collect -from sqlalchemy.testing import AssertsExecutionResults, profiling, testing -from test.orm import _fixtures - -# in this test we are specifically looking for time spent in the attributes.InstanceState.__cleanup() method. - -ITERATIONS = 100 - -class SessionTest(fixtures.TestBase, AssertsExecutionResults): - @classmethod - def setup_class(cls): - global t1, t2, metadata,T1, T2 - metadata = MetaData(testing.db) - t1 = Table('t1', metadata, - Column('c1', Integer, primary_key=True), - Column('c2', String(30))) - - t2 = Table('t2', metadata, - Column('c1', Integer, primary_key=True), - Column('c2', String(30)), - Column('t1id', Integer, ForeignKey('t1.c1')) - ) - - metadata.create_all() - - l = [] - for x in range(1,51): - l.append({'c2':'this is t1 #%d' % x}) - t1.insert().execute(*l) - for x in range(1, 51): - l = [] - for y in range(1, 100): - l.append({'c2':'this is t2 #%d' % y, 't1id':x}) - t2.insert().execute(*l) - - class T1(fixtures.ComparableEntity): - pass - class T2(fixtures.ComparableEntity): - pass - - mapper(T1, t1, properties={ - 't2s':relationship(T2, backref='t1') - }) - mapper(T2, t2) - - @classmethod - def teardown_class(cls): - metadata.drop_all() - clear_mappers() - - @profiling.profiled('clean', report=True) - def test_session_clean(self): - for x in range(0, ITERATIONS): - sess = create_session() - t1s = sess.query(T1).filter(T1.c1.between(15, 48)).all() - for index in [2, 7, 12, 15, 18, 20]: - t1s[index].t2s - - sess.close() - del sess - gc_collect() - - @profiling.profiled('dirty', report=True) - def test_session_dirty(self): - for x in range(0, ITERATIONS): - sess = create_session() - t1s = sess.query(T1).filter(T1.c1.between(15, 48)).all() - - for index in [2, 7, 12, 15, 18, 20]: - t1s[index].c2 = 'this is some modified text' - for t2 in t1s[index].t2s: - t2.c2 = 'this is some modified text' - - del t1s - gc_collect() - - sess.close() - del sess - gc_collect() - - @profiling.profiled('noclose', report=True) - def test_session_noclose(self): - for x in range(0, ITERATIONS): - sess = create_session() - t1s = sess.query(T1).filter(T1.c1.between(15, 48)).all() - for index in [2, 7, 12, 15, 18, 20]: - t1s[index].t2s - - del sess - gc_collect() - - diff --git a/test/perf/stress_all.py b/test/perf/stress_all.py deleted file mode 100644 index 890ef24a3..000000000 --- a/test/perf/stress_all.py +++ /dev/null @@ -1,226 +0,0 @@ -# -*- encoding: utf8 -*- -from datetime import * -import decimal -#from fastdec import mpd as Decimal -from cPickle import dumps, loads - -#from sqlalchemy.dialects.postgresql.base import ARRAY - -from stresstest import * - -# --- -test_types = False -test_methods = True -test_pickle = False -test_orm = False -# --- -verbose = True - -def values_results(raw_results): - return [tuple(r.values()) for r in raw_results] - -def getitem_str_results(raw_results): - return [ - (r['id'], - r['field0'], r['field1'], r['field2'], r['field3'], r['field4'], - r['field5'], r['field6'], r['field7'], r['field8'], r['field9']) - for r in raw_results] - -def getitem_fallback_results(raw_results): - return [ - (r['ID'], - r['FIELD0'], r['FIELD1'], r['FIELD2'], r['FIELD3'], r['FIELD4'], - r['FIELD5'], r['FIELD6'], r['FIELD7'], r['FIELD8'], r['FIELD9']) - for r in raw_results] - -def getitem_int_results(raw_results): - return [ - (r[0], - r[1], r[2], r[3], r[4], r[5], - r[6], r[7], r[8], r[9], r[10]) - for r in raw_results] - -def getitem_long_results(raw_results): - return [ - (r[0L], - r[1L], r[2L], r[3L], r[4L], r[5L], - r[6L], r[7L], r[8L], r[9L], r[10L]) - for r in raw_results] - -def getitem_obj_results(raw_results): - c = test_table.c - fid, f0, f1, f2, f3, f4, f5, f6, f7, f8, f9 = ( - c.id, c.field0, c.field1, c.field2, c.field3, c.field4, - c.field5, c.field6, c.field7, c.field8, c.field9) - return [ - (r[fid], - r[f0], r[f1], r[f2], r[f3], r[f4], - r[f5], r[f6], r[f7], r[f8], r[f9]) - for r in raw_results] - -def slice_results(raw_results): - return [row[0:6] + row[6:11] for row in raw_results] - -# ---------- # -# Test types # -# ---------- # - -# Array -#def genarrayvalue(rnum, fnum): -# return [fnum, fnum + 1, fnum + 2] -#arraytest = (ARRAY(Integer), genarrayvalue, -# dict(num_fields=100, num_records=1000, -# engineurl='postgresql:///test')) - -# Boolean -def genbooleanvalue(rnum, fnum): - if rnum % 4: - return bool(fnum % 2) - else: - return None -booleantest = (Boolean, genbooleanvalue, dict(num_records=100000)) - -# Datetime -def gendatetimevalue(rnum, fnum): - return (rnum % 4) and datetime(2005, 3, 3) or None -datetimetest = (DateTime, gendatetimevalue, dict(num_records=10000)) - -# Decimal -def gendecimalvalue(rnum, fnum): - if rnum % 4: - return Decimal(str(0.25 * fnum)) - else: - return None -decimaltest = (Numeric(10, 2), gendecimalvalue, dict(num_records=10000)) - -# Interval - -# no microseconds because Postgres does not seem to support it -from_epoch = timedelta(14643, 70235) -def genintervalvalue(rnum, fnum): - return from_epoch -intervaltest = (Interval, genintervalvalue, - dict(num_fields=2, num_records=100000)) - -# PickleType -def genpicklevalue(rnum, fnum): - return (rnum % 4) and {'str': "value%d" % fnum, 'int': rnum} or None -pickletypetest = (PickleType, genpicklevalue, - dict(num_fields=1, num_records=100000)) - -# TypeDecorator -class MyIntType(TypeDecorator): - impl = Integer - - def process_bind_param(self, value, dialect): - return value * 10 - - def process_result_value(self, value, dialect): - return value / 10 - - def copy(self): - return MyIntType() - -def genmyintvalue(rnum, fnum): - return rnum + fnum -typedecoratortest = (MyIntType, genmyintvalue, - dict(num_records=100000)) - -# Unicode -def genunicodevalue(rnum, fnum): - return (rnum % 4) and (u"value%d" % fnum) or None -unicodetest = (Unicode(20, ), genunicodevalue, - dict(num_records=100000)) -# dict(engineurl='mysql:///test', freshdata=False)) - -# do the tests -if test_types: - tests = [booleantest, datetimetest, decimaltest, intervaltest, - pickletypetest, typedecoratortest, unicodetest] - for engineurl in ('postgresql://scott:tiger@localhost/test', - 'sqlite://', 'mysql://scott:tiger@localhost/test'): - print "\n%s\n" % engineurl - for datatype, genvalue, kwargs in tests: - print "%s:" % getattr(datatype, '__name__', - datatype.__class__.__name__), - profile_and_time_dbfunc(iter_results, datatype, genvalue, - profile=False, engineurl=engineurl, - verbose=verbose, **kwargs) - -# ---------------------- # -# test row proxy methods # -# ---------------------- # - -if test_methods: - methods = [iter_results, values_results, getattr_results, - getitem_str_results, getitem_fallback_results, - getitem_int_results, getitem_long_results, getitem_obj_results, - slice_results] - for engineurl in ('postgresql://scott:tiger@localhost/test', - 'sqlite://', 'mysql://scott:tiger@localhost/test'): - print "\n%s\n" % engineurl - test_table = prepare(Unicode(20,), - genunicodevalue, - num_fields=10, num_records=100000, - verbose=verbose, engineurl=engineurl) - for method in methods: - print "%s:" % method.__name__, - time_dbfunc(test_table, method, genunicodevalue, - num_fields=10, num_records=100000, profile=False, - verbose=verbose) - -# -------------------------------- -# test pickling Rowproxy instances -# -------------------------------- - -def pickletofile_results(raw_results): - from cPickle import dump, load - for protocol in (0, 1, 2): - print "dumping protocol %d..." % protocol - f = file('noext.pickle%d' % protocol, 'wb') - dump(raw_results, f, protocol) - f.close() - return raw_results - -def pickle_results(raw_results): - return loads(dumps(raw_results, 2)) - -def pickle_meta(raw_results): - pickled = dumps(raw_results[0]._parent, 2) - metadata = loads(pickled) - return raw_results - -def pickle_rows(raw_results): - return [loads(dumps(row, 2)) for row in raw_results] - -if test_pickle: - test_table = prepare(Unicode, genunicodevalue, - num_fields=10, num_records=10000) - funcs = [pickle_rows, pickle_results] - for func in funcs: - print "%s:" % func.__name__, - time_dbfunc(test_table, func, genunicodevalue, - num_records=10000, profile=False, verbose=verbose) - -# -------------------------------- -# test ORM -# -------------------------------- - -if test_orm: - from sqlalchemy.orm import * - - class Test(object): - pass - - Session = sessionmaker() - session = Session() - - def get_results(): - return session.query(Test).all() - print "ORM:", - for engineurl in ('postgresql:///test', 'sqlite://', 'mysql:///test'): - print "\n%s\n" % engineurl - profile_and_time_dbfunc(getattr_results, Unicode(20), genunicodevalue, - class_=Test, getresults_func=get_results, - engineurl=engineurl, #freshdata=False, - num_records=10000, verbose=verbose) diff --git a/test/perf/stresstest.py b/test/perf/stresstest.py deleted file mode 100644 index cf9404f53..000000000 --- a/test/perf/stresstest.py +++ /dev/null @@ -1,174 +0,0 @@ -import gc -import sys -import timeit -import cProfile - -from sqlalchemy import MetaData, Table, Column -from sqlalchemy.types import * -from sqlalchemy.orm import mapper, clear_mappers - -metadata = MetaData() - -def gen_table(num_fields, field_type, metadata): - return Table('test', metadata, - Column('id', Integer, primary_key=True), - *[Column("field%d" % fnum, field_type) - for fnum in range(num_fields)]) - -def insert(test_table, num_fields, num_records, genvalue, verbose=True): - if verbose: - print "building insert values...", - sys.stdout.flush() - values = [dict(("field%d" % fnum, genvalue(rnum, fnum)) - for fnum in range(num_fields)) - for rnum in range(num_records)] - if verbose: - print "inserting...", - sys.stdout.flush() - def db_insert(): - test_table.insert().execute(values) - sys.modules['__main__'].db_insert = db_insert - timing = timeit.timeit("db_insert()", - "from __main__ import db_insert", - number=1) - if verbose: - print "%s" % round(timing, 3) - -def check_result(results, num_fields, genvalue, verbose=True): - if verbose: - print "checking...", - sys.stdout.flush() - for rnum, row in enumerate(results): - expected = tuple([rnum + 1] + - [genvalue(rnum, fnum) for fnum in range(num_fields)]) - assert row == expected, "got: %s\nexpected: %s" % (row, expected) - return True - -def avgdev(values, comparison): - return sum(value - comparison for value in values) / len(values) - -def nicer_res(values, printvalues=False): - if printvalues: - print values - min_time = min(values) - return round(min_time, 3), round(avgdev(values, min_time), 2) - -def profile_func(func_name, verbose=True): - if verbose: - print "profiling...", - sys.stdout.flush() - cProfile.run('%s()' % func_name, 'prof') - -def time_func(func_name, num_tests=1, verbose=True): - if verbose: - print "timing...", - sys.stdout.flush() - timings = timeit.repeat('%s()' % func_name, - "from __main__ import %s" % func_name, - number=num_tests, repeat=5) - avg, dev = nicer_res(timings) - if verbose: - print "%s (%s)" % (avg, dev) - else: - print avg - -def profile_and_time(func_name, num_tests=1): - profile_func(func_name) - time_func(func_name, num_tests) - -def iter_results(raw_results): - return [tuple(row) for row in raw_results] - -def getattr_results(raw_results): - return [ - (r.id, - r.field0, r.field1, r.field2, r.field3, r.field4, - r.field5, r.field6, r.field7, r.field8, r.field9) - for r in raw_results] - -def fetchall(test_table): - def results(): - return test_table.select().order_by(test_table.c.id).execute() \ - .fetchall() - return results - -def hashable_set(l): - hashables = [] - for o in l: - try: - hash(o) - hashables.append(o) - except: - pass - return set(hashables) - -def prepare(field_type, genvalue, engineurl='sqlite://', - num_fields=10, num_records=1000, freshdata=True, verbose=True): - global metadata - metadata.clear() - metadata.bind = engineurl - test_table = gen_table(num_fields, field_type, metadata) - if freshdata: - metadata.drop_all() - metadata.create_all() - insert(test_table, num_fields, num_records, genvalue, verbose) - return test_table - -def time_dbfunc(test_table, test_func, genvalue, - class_=None, - getresults_func=None, - num_fields=10, num_records=1000, num_tests=1, - check_results=check_result, profile=True, - check_leaks=True, print_leaks=False, verbose=True): - if verbose: - print "testing '%s'..." % test_func.__name__, - sys.stdout.flush() - if class_ is not None: - clear_mappers() - mapper(class_, test_table) - if getresults_func is None: - getresults_func = fetchall(test_table) - def test(): - return test_func(getresults_func()) - sys.modules['__main__'].test = test - if check_leaks: - gc.collect() - objects_before = gc.get_objects() - num_objects_before = len(objects_before) - hashable_objects_before = hashable_set(objects_before) -# gc.set_debug(gc.DEBUG_LEAK) - if check_results: - check_results(test(), num_fields, genvalue, verbose) - if check_leaks: - gc.collect() - objects_after = gc.get_objects() - num_objects_after = len(objects_after) - num_leaks = num_objects_after - num_objects_before - hashable_objects_after = hashable_set(objects_after) - diff = hashable_objects_after - hashable_objects_before - ldiff = len(diff) - if print_leaks and ldiff < num_records: - print "\n*** hashable objects leaked (%d) ***" % ldiff - print '\n'.join(map(str, diff)) - print "***\n" - - if num_leaks > num_records: - print "(leaked: %d !)" % num_leaks, - if profile: - profile_func('test', verbose) - time_func('test', num_tests, verbose) - -def profile_and_time_dbfunc(test_func, field_type, genvalue, - class_=None, - getresults_func=None, - engineurl='sqlite://', freshdata=True, - num_fields=10, num_records=1000, num_tests=1, - check_results=check_result, profile=True, - check_leaks=True, print_leaks=False, verbose=True): - test_table = prepare(field_type, genvalue, engineurl, - num_fields, num_records, freshdata, verbose) - time_dbfunc(test_table, test_func, genvalue, class_, - getresults_func, - num_fields, num_records, num_tests, - check_results, profile, - check_leaks, print_leaks, verbose) diff --git a/test/perf/threaded_compile.py b/test/perf/threaded_compile.py deleted file mode 100644 index 0facf0ec2..000000000 --- a/test/perf/threaded_compile.py +++ /dev/null @@ -1,75 +0,0 @@ -"""test that mapper compilation is threadsafe, including -when additional mappers are created while the existing -collection is being compiled.""" - -from sqlalchemy import * -from sqlalchemy.orm import * -import thread, time -from sqlalchemy.orm import mapperlib - - -meta = MetaData('sqlite:///foo.db') - -t1 = Table('t1', meta, - Column('c1', Integer, primary_key=True), - Column('c2', String(30)) - ) - -t2 = Table('t2', meta, - Column('c1', Integer, primary_key=True), - Column('c2', String(30)), - Column('t1c1', None, ForeignKey('t1.c1')) -) -t3 = Table('t3', meta, - Column('c1', Integer, primary_key=True), - Column('c2', String(30)), -) -meta.create_all() - -class T1(object): - pass - -class T2(object): - pass - -class FakeLock(object): - def acquire(self):pass - def release(self):pass - -# uncomment this to disable the mutex in mapper compilation; -# should produce thread collisions -#mapperlib._COMPILE_MUTEX = FakeLock() - -def run1(): - for i in range(50): - print "T1", thread.get_ident() - class_mapper(T1) - time.sleep(.05) - -def run2(): - for i in range(50): - print "T2", thread.get_ident() - class_mapper(T2) - time.sleep(.057) - -def run3(): - for i in range(50): - def foo(): - print "FOO", thread.get_ident() - class Foo(object):pass - mapper(Foo, t3) - class_mapper(Foo).compile() - foo() - time.sleep(.05) - -mapper(T1, t1, properties={'t2':relationship(T2, backref="t1")}) -mapper(T2, t2) -print "START" -for j in range(0, 5): - thread.start_new_thread(run1, ()) - thread.start_new_thread(run2, ()) - thread.start_new_thread(run3, ()) - thread.start_new_thread(run3, ()) - thread.start_new_thread(run3, ()) -print "WAIT" -time.sleep(5) diff --git a/test/profiles.txt b/test/profiles.txt index c9ba04c24..090cf5ea1 100644 --- a/test/profiles.txt +++ b/test/profiles.txt @@ -1,67 +1,88 @@ # /Users/classic/dev/sqlalchemy/test/profiles.txt # This file is written out on a per-environment basis. -# For each test in aaa_profiling, the corresponding function and +# For each test in aaa_profiling, the corresponding function and # environment is located within this file. If it doesn't exist, # the test is skipped. -# If a callcount does exist, it is compared to what we received. +# If a callcount does exist, it is compared to what we received. # assertions are raised if the counts do not match. -# -# To add a new callcount test, apply the function_call_count -# decorator and re-run the tests using the --write-profiles +# +# To add a new callcount test, apply the function_call_count +# decorator and re-run the tests using the --write-profiles # option - this file will be rewritten including the new count. -# +# # TEST: test.aaa_profiling.test_compiler.CompileTest.test_insert -test.aaa_profiling.test_compiler.CompileTest.test_insert 2.5_sqlite_pysqlite_nocextensions 62 -test.aaa_profiling.test_compiler.CompileTest.test_insert 2.6_sqlite_pysqlite_nocextensions 62 -test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_mysql_mysqldb_cextensions 62 -test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_mysql_mysqldb_nocextensions 62 -test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_postgresql_psycopg2_cextensions 62 -test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_postgresql_psycopg2_nocextensions 62 -test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_sqlite_pysqlite_cextensions 62 -test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_sqlite_pysqlite_nocextensions 62 +test.aaa_profiling.test_compiler.CompileTest.test_insert 2.6_sqlite_pysqlite_nocextensions 67 +test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_mysql_mysqldb_cextensions 67 +test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_mysql_mysqldb_nocextensions 67 +test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_postgresql_psycopg2_cextensions 67 +test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_postgresql_psycopg2_nocextensions 67 +test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_sqlite_pysqlite_cextensions 67 +test.aaa_profiling.test_compiler.CompileTest.test_insert 2.7_sqlite_pysqlite_nocextensions 67 +test.aaa_profiling.test_compiler.CompileTest.test_insert 3.2_postgresql_psycopg2_nocextensions 69 +test.aaa_profiling.test_compiler.CompileTest.test_insert 3.2_sqlite_pysqlite_nocextensions 69 +test.aaa_profiling.test_compiler.CompileTest.test_insert 3.3_postgresql_psycopg2_nocextensions 69 +test.aaa_profiling.test_compiler.CompileTest.test_insert 3.3_sqlite_pysqlite_nocextensions 69 # TEST: test.aaa_profiling.test_compiler.CompileTest.test_select -test.aaa_profiling.test_compiler.CompileTest.test_select 2.5_sqlite_pysqlite_nocextensions 134 -test.aaa_profiling.test_compiler.CompileTest.test_select 2.6_sqlite_pysqlite_nocextensions 135 -test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_mysql_mysqldb_cextensions 135 -test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_mysql_mysqldb_nocextensions 135 -test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_postgresql_psycopg2_cextensions 135 -test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_postgresql_psycopg2_nocextensions 135 -test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_sqlite_pysqlite_cextensions 135 -test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_sqlite_pysqlite_nocextensions 135 +test.aaa_profiling.test_compiler.CompileTest.test_select 2.6_sqlite_pysqlite_nocextensions 141 +test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_mysql_mysqldb_cextensions 141 +test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_mysql_mysqldb_nocextensions 141 +test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_postgresql_psycopg2_cextensions 141 +test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_postgresql_psycopg2_nocextensions 141 +test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_sqlite_pysqlite_cextensions 141 +test.aaa_profiling.test_compiler.CompileTest.test_select 2.7_sqlite_pysqlite_nocextensions 141 +test.aaa_profiling.test_compiler.CompileTest.test_select 3.2_postgresql_psycopg2_nocextensions 151 +test.aaa_profiling.test_compiler.CompileTest.test_select 3.2_sqlite_pysqlite_nocextensions 151 +test.aaa_profiling.test_compiler.CompileTest.test_select 3.3_postgresql_psycopg2_nocextensions 151 +test.aaa_profiling.test_compiler.CompileTest.test_select 3.3_sqlite_pysqlite_nocextensions 151 # TEST: test.aaa_profiling.test_compiler.CompileTest.test_select_labels -test.aaa_profiling.test_compiler.CompileTest.test_select_labels 2.7_sqlite_pysqlite_nocextensions 177 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 2.6_sqlite_pysqlite_nocextensions 175 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 2.7_mysql_mysqldb_cextensions 175 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 2.7_mysql_mysqldb_nocextensions 175 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 2.7_postgresql_psycopg2_cextensions 175 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 2.7_postgresql_psycopg2_nocextensions 175 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 2.7_sqlite_pysqlite_cextensions 175 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 2.7_sqlite_pysqlite_nocextensions 175 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 3.2_postgresql_psycopg2_nocextensions 185 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 3.2_sqlite_pysqlite_nocextensions 185 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 3.3_postgresql_psycopg2_nocextensions 185 +test.aaa_profiling.test_compiler.CompileTest.test_select_labels 3.3_sqlite_pysqlite_nocextensions 185 # TEST: test.aaa_profiling.test_compiler.CompileTest.test_update -test.aaa_profiling.test_compiler.CompileTest.test_update 2.5_sqlite_pysqlite_nocextensions 65 -test.aaa_profiling.test_compiler.CompileTest.test_update 2.6_sqlite_pysqlite_nocextensions 65 -test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_mysql_mysqldb_cextensions 65 -test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_mysql_mysqldb_nocextensions 65 -test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_postgresql_psycopg2_cextensions 65 -test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_postgresql_psycopg2_nocextensions 65 -test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_sqlite_pysqlite_cextensions 65 -test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_sqlite_pysqlite_nocextensions 65 +test.aaa_profiling.test_compiler.CompileTest.test_update 2.6_sqlite_pysqlite_nocextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_mysql_mysqldb_cextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_mysql_mysqldb_nocextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_postgresql_psycopg2_cextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_postgresql_psycopg2_nocextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_sqlite_pysqlite_cextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 2.7_sqlite_pysqlite_nocextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 3.2_postgresql_psycopg2_nocextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 3.2_sqlite_pysqlite_nocextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 3.3_postgresql_psycopg2_nocextensions 70 +test.aaa_profiling.test_compiler.CompileTest.test_update 3.3_sqlite_pysqlite_nocextensions 70 # TEST: test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause -test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.5_sqlite_pysqlite_nocextensions 129 -test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.6_sqlite_pysqlite_nocextensions 130 -test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_mysql_mysqldb_cextensions 130 -test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_mysql_mysqldb_nocextensions 130 -test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_postgresql_psycopg2_cextensions 130 -test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_postgresql_psycopg2_nocextensions 130 -test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_sqlite_pysqlite_cextensions 130 -test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_sqlite_pysqlite_nocextensions 130 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.6_sqlite_pysqlite_nocextensions 137 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_mysql_mysqldb_cextensions 137 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_mysql_mysqldb_nocextensions 137 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_postgresql_psycopg2_cextensions 137 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_postgresql_psycopg2_nocextensions 137 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_sqlite_pysqlite_cextensions 137 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 2.7_sqlite_pysqlite_nocextensions 137 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 3.2_postgresql_psycopg2_nocextensions 136 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 3.2_sqlite_pysqlite_nocextensions 136 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 3.3_postgresql_psycopg2_nocextensions 136 +test.aaa_profiling.test_compiler.CompileTest.test_update_whereclause 3.3_sqlite_pysqlite_nocextensions 136 # TEST: test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity -test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 2.5_sqlite_pysqlite_nocextensions 17987 test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 2.6_sqlite_pysqlite_nocextensions 17987 test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 2.7_mysql_mysqldb_cextensions 17987 test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 2.7_mysql_mysqldb_nocextensions 17987 @@ -69,39 +90,52 @@ test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_ test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 2.7_postgresql_psycopg2_nocextensions 17987 test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 2.7_sqlite_pysqlite_cextensions 17987 test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 2.7_sqlite_pysqlite_nocextensions 17987 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 3.2_postgresql_psycopg2_nocextensions 18987 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 3.2_sqlite_pysqlite_nocextensions 18987 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 3.3_postgresql_psycopg2_nocextensions 18987 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_identity 3.3_sqlite_pysqlite_nocextensions 18987 # TEST: test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity -test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.5_sqlite_pysqlite_nocextensions 116289 -test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.6_sqlite_pysqlite_nocextensions 116790 -test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_mysql_mysqldb_cextensions 122540 -test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_mysql_mysqldb_nocextensions 125290 -test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_postgresql_psycopg2_cextensions 115040 -test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_postgresql_psycopg2_nocextensions 117790 -test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_sqlite_pysqlite_cextensions 114040 -test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_sqlite_pysqlite_nocextensions 116790 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.6_sqlite_pysqlite_nocextensions 118319 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_mysql_mysqldb_cextensions 124069 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_mysql_mysqldb_nocextensions 126819 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_postgresql_psycopg2_cextensions 116569 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_postgresql_psycopg2_nocextensions 119319 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_sqlite_pysqlite_cextensions 115569 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 2.7_sqlite_pysqlite_nocextensions 118319 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 3.2_postgresql_psycopg2_nocextensions 121790 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 3.2_sqlite_pysqlite_nocextensions 121822 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 3.3_postgresql_psycopg2_nocextensions 121822 +test.aaa_profiling.test_orm.LoadManyToOneFromIdentityTest.test_many_to_one_load_no_identity 3.3_sqlite_pysqlite_nocextensions 121822 # TEST: test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks -test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.5_sqlite_pysqlite_nocextensions 19852 -test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.6_sqlite_pysqlite_nocextensions 19217 -test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_mysql_mysqldb_cextensions 19491 -test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_mysql_mysqldb_nocextensions 19781 -test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_postgresql_psycopg2_cextensions 18878 -test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_postgresql_psycopg2_nocextensions 19168 -test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_sqlite_pysqlite_cextensions 18957 -test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_sqlite_pysqlite_nocextensions 19217 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.6_sqlite_pysqlite_nocextensions 19534 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_mysql_mysqldb_cextensions 19838 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_mysql_mysqldb_nocextensions 20098 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_postgresql_psycopg2_cextensions 19237 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_postgresql_psycopg2_nocextensions 19467 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_sqlite_pysqlite_cextensions 19274 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 2.7_sqlite_pysqlite_nocextensions 19534 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 3.2_postgresql_psycopg2_nocextensions 20424 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 3.2_sqlite_pysqlite_nocextensions 20447 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 3.3_postgresql_psycopg2_nocextensions 20344 +test.aaa_profiling.test_orm.MergeBackrefsTest.test_merge_pending_with_all_pks 3.3_sqlite_pysqlite_nocextensions 20433 # TEST: test.aaa_profiling.test_orm.MergeTest.test_merge_load -test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.5_sqlite_pysqlite_nocextensions 1178 -test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.6_sqlite_pysqlite_nocextensions 1174 -test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_mysql_mysqldb_cextensions 1341 -test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_mysql_mysqldb_nocextensions 1366 -test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_postgresql_psycopg2_cextensions 1275 -test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_postgresql_psycopg2_nocextensions 1307 -test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_sqlite_pysqlite_cextensions 1149 -test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_sqlite_pysqlite_nocextensions 1174 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.6_sqlite_pysqlite_nocextensions 1221 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_mysql_mysqldb_cextensions 1388 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_mysql_mysqldb_nocextensions 1413 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_postgresql_psycopg2_cextensions 1296 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_postgresql_psycopg2_nocextensions 1321 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_sqlite_pysqlite_cextensions 1196 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 2.7_sqlite_pysqlite_nocextensions 1221 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 3.2_postgresql_psycopg2_nocextensions 1332 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 3.2_sqlite_pysqlite_nocextensions 1243 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 3.3_postgresql_psycopg2_nocextensions 1357 +test.aaa_profiling.test_orm.MergeTest.test_merge_load 3.3_sqlite_pysqlite_nocextensions 1243 # TEST: test.aaa_profiling.test_orm.MergeTest.test_merge_no_load @@ -112,6 +146,10 @@ test.aaa_profiling.test_orm.MergeTest.test_merge_no_load 2.7_postgresql_psycopg2 test.aaa_profiling.test_orm.MergeTest.test_merge_no_load 2.7_postgresql_psycopg2_nocextensions 122,18 test.aaa_profiling.test_orm.MergeTest.test_merge_no_load 2.7_sqlite_pysqlite_cextensions 122,18 test.aaa_profiling.test_orm.MergeTest.test_merge_no_load 2.7_sqlite_pysqlite_nocextensions 122,18 +test.aaa_profiling.test_orm.MergeTest.test_merge_no_load 3.2_postgresql_psycopg2_nocextensions 127,19 +test.aaa_profiling.test_orm.MergeTest.test_merge_no_load 3.2_sqlite_pysqlite_nocextensions 127,19 +test.aaa_profiling.test_orm.MergeTest.test_merge_no_load 3.3_postgresql_psycopg2_nocextensions 127,19 +test.aaa_profiling.test_orm.MergeTest.test_merge_no_load 3.3_sqlite_pysqlite_nocextensions 127,19 # TEST: test.aaa_profiling.test_pool.QueuePoolTest.test_first_connect @@ -122,10 +160,13 @@ test.aaa_profiling.test_pool.QueuePoolTest.test_first_connect 2.7_postgresql_psy test.aaa_profiling.test_pool.QueuePoolTest.test_first_connect 2.7_postgresql_psycopg2_nocextensions 82 test.aaa_profiling.test_pool.QueuePoolTest.test_first_connect 2.7_sqlite_pysqlite_cextensions 82 test.aaa_profiling.test_pool.QueuePoolTest.test_first_connect 2.7_sqlite_pysqlite_nocextensions 82 +test.aaa_profiling.test_pool.QueuePoolTest.test_first_connect 3.2_postgresql_psycopg2_nocextensions 70 +test.aaa_profiling.test_pool.QueuePoolTest.test_first_connect 3.2_sqlite_pysqlite_nocextensions 70 +test.aaa_profiling.test_pool.QueuePoolTest.test_first_connect 3.3_postgresql_psycopg2_nocextensions 69 +test.aaa_profiling.test_pool.QueuePoolTest.test_first_connect 3.3_sqlite_pysqlite_nocextensions 69 # TEST: test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect -test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 2.5_sqlite_pysqlite_nocextensions 32 test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 2.6_sqlite_pysqlite_nocextensions 29 test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 2.7_mysql_mysqldb_cextensions 29 test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 2.7_mysql_mysqldb_nocextensions 29 @@ -133,10 +174,13 @@ test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 2.7_postgresql_ps test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 2.7_postgresql_psycopg2_nocextensions 29 test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 2.7_sqlite_pysqlite_cextensions 29 test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 2.7_sqlite_pysqlite_nocextensions 29 +test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 3.2_postgresql_psycopg2_nocextensions 23 +test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 3.2_sqlite_pysqlite_nocextensions 23 +test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 3.3_postgresql_psycopg2_nocextensions 22 +test.aaa_profiling.test_pool.QueuePoolTest.test_second_connect 3.3_sqlite_pysqlite_nocextensions 22 # TEST: test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect -test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 2.5_sqlite_pysqlite_nocextensions 6 test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 2.6_sqlite_pysqlite_nocextensions 6 test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 2.7_mysql_mysqldb_cextensions 6 test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 2.7_mysql_mysqldb_nocextensions 6 @@ -144,32 +188,41 @@ test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 2.7_po test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 2.7_postgresql_psycopg2_nocextensions 6 test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 2.7_sqlite_pysqlite_cextensions 6 test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 2.7_sqlite_pysqlite_nocextensions 6 +test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 3.2_postgresql_psycopg2_nocextensions 7 +test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 3.2_sqlite_pysqlite_nocextensions 7 +test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 3.3_postgresql_psycopg2_nocextensions 7 +test.aaa_profiling.test_pool.QueuePoolTest.test_second_samethread_connect 3.3_sqlite_pysqlite_nocextensions 7 # TEST: test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.5_sqlite_pysqlite_nocextensions 41 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.6_sqlite_pysqlite_nocextensions 42 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_mysql_mysqldb_cextensions 40 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_mysql_mysqldb_nocextensions 42 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_postgresql_psycopg2_cextensions 40 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_postgresql_psycopg2_nocextensions 42 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_sqlite_pysqlite_cextensions 40 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_sqlite_pysqlite_nocextensions 42 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.6_sqlite_pysqlite_nocextensions 43 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_mysql_mysqldb_cextensions 41 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_mysql_mysqldb_nocextensions 43 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_postgresql_psycopg2_cextensions 41 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_postgresql_psycopg2_nocextensions 43 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_sqlite_pysqlite_cextensions 41 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 2.7_sqlite_pysqlite_nocextensions 43 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 3.2_postgresql_psycopg2_nocextensions 41 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 3.2_sqlite_pysqlite_nocextensions 41 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 3.3_postgresql_psycopg2_nocextensions 41 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_connection_execute 3.3_sqlite_pysqlite_nocextensions 41 # TEST: test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.5_sqlite_pysqlite_nocextensions 64 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.6_sqlite_pysqlite_nocextensions 65 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_mysql_mysqldb_cextensions 63 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_mysql_mysqldb_nocextensions 65 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_postgresql_psycopg2_cextensions 63 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_postgresql_psycopg2_nocextensions 65 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_sqlite_pysqlite_cextensions 63 -test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_sqlite_pysqlite_nocextensions 65 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.6_sqlite_pysqlite_nocextensions 68 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_mysql_mysqldb_cextensions 66 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_mysql_mysqldb_nocextensions 68 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_postgresql_psycopg2_cextensions 66 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_postgresql_psycopg2_nocextensions 68 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_sqlite_pysqlite_cextensions 66 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 2.7_sqlite_pysqlite_nocextensions 68 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 3.2_postgresql_psycopg2_nocextensions 66 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 3.2_sqlite_pysqlite_nocextensions 66 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 3.3_postgresql_psycopg2_nocextensions 66 +test.aaa_profiling.test_resultset.ExecutionTest.test_minimal_engine_execute 3.3_sqlite_pysqlite_nocextensions 66 # TEST: test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile -test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 2.5_sqlite_pysqlite_nocextensions 14 test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 2.6_sqlite_pysqlite_nocextensions 14 test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 2.7_mysql_mysqldb_cextensions 14 test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 2.7_mysql_mysqldb_nocextensions 14 @@ -177,90 +230,125 @@ test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 2.7 test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 2.7_postgresql_psycopg2_nocextensions 14 test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 2.7_sqlite_pysqlite_cextensions 14 test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 2.7_sqlite_pysqlite_nocextensions 14 +test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 3.2_postgresql_psycopg2_nocextensions 15 +test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 3.2_sqlite_pysqlite_nocextensions 15 +test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 3.3_postgresql_psycopg2_nocextensions 15 +test.aaa_profiling.test_resultset.ResultSetTest.test_contains_doesnt_compile 3.3_sqlite_pysqlite_nocextensions 15 # TEST: test.aaa_profiling.test_resultset.ResultSetTest.test_string -test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.5_sqlite_pysqlite_nocextensions 14413 -test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.6_sqlite_pysqlite_nocextensions 14414 -test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_mysql_mysqldb_cextensions 476 -test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_mysql_mysqldb_nocextensions 14472 -test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_postgresql_psycopg2_cextensions 20438 -test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_postgresql_psycopg2_nocextensions 34458 -test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_sqlite_pysqlite_cextensions 416 -test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_sqlite_pysqlite_nocextensions 14414 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.6_sqlite_pysqlite_nocextensions 15447 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_mysql_mysqldb_cextensions 485 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_mysql_mysqldb_nocextensions 15505 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_postgresql_psycopg2_cextensions 20471 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_postgresql_psycopg2_nocextensions 35491 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_sqlite_pysqlite_cextensions 427 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 2.7_sqlite_pysqlite_nocextensions 15447 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 3.2_postgresql_psycopg2_nocextensions 14459 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 3.2_sqlite_pysqlite_nocextensions 14430 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 3.3_postgresql_psycopg2_nocextensions 14457 +test.aaa_profiling.test_resultset.ResultSetTest.test_string 3.3_sqlite_pysqlite_nocextensions 14430 # TEST: test.aaa_profiling.test_resultset.ResultSetTest.test_unicode -test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.5_sqlite_pysqlite_nocextensions 14413 -test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.6_sqlite_pysqlite_nocextensions 14414 -test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_mysql_mysqldb_cextensions 476 -test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_mysql_mysqldb_nocextensions 44472 -test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_postgresql_psycopg2_cextensions 20438 -test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_postgresql_psycopg2_nocextensions 34458 -test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_sqlite_pysqlite_cextensions 416 -test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_sqlite_pysqlite_nocextensions 14414 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.6_sqlite_pysqlite_nocextensions 15447 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_mysql_mysqldb_cextensions 485 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_mysql_mysqldb_nocextensions 45505 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_postgresql_psycopg2_cextensions 20471 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_postgresql_psycopg2_nocextensions 35491 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_sqlite_pysqlite_cextensions 427 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 2.7_sqlite_pysqlite_nocextensions 15447 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 3.2_postgresql_psycopg2_nocextensions 14459 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 3.2_sqlite_pysqlite_nocextensions 14430 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 3.3_postgresql_psycopg2_nocextensions 14457 +test.aaa_profiling.test_resultset.ResultSetTest.test_unicode 3.3_sqlite_pysqlite_nocextensions 14430 # TEST: test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_1a_populate -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_1a_populate 2.7_postgresql_psycopg2_cextensions 5044 -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_1a_populate 2.7_postgresql_psycopg2_nocextensions 5088 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_1a_populate 2.7_postgresql_psycopg2_nocextensions 5175 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_1a_populate 3.2_postgresql_psycopg2_nocextensions 4828 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_1a_populate 3.3_postgresql_psycopg2_nocextensions 4792 # TEST: test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_2_insert -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_2_insert 2.7_postgresql_psycopg2_cextensions 247 -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_2_insert 2.7_postgresql_psycopg2_nocextensions 247 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_2_insert 2.7_postgresql_psycopg2_cextensions 256 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_2_insert 2.7_postgresql_psycopg2_nocextensions 256 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_2_insert 3.2_postgresql_psycopg2_nocextensions 239 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_2_insert 3.3_postgresql_psycopg2_nocextensions 237 # TEST: test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_3_properties -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_3_properties 2.7_postgresql_psycopg2_cextensions 3366 -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_3_properties 2.7_postgresql_psycopg2_nocextensions 3590 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_3_properties 2.7_postgresql_psycopg2_cextensions 3425 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_3_properties 2.7_postgresql_psycopg2_nocextensions 3749 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_3_properties 3.2_postgresql_psycopg2_nocextensions 3401 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_3_properties 3.3_postgresql_psycopg2_nocextensions 3385 # TEST: test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_4_expressions -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_4_expressions 2.7_postgresql_psycopg2_cextensions 10915 -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_4_expressions 2.7_postgresql_psycopg2_nocextensions 11982 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_4_expressions 2.7_postgresql_psycopg2_cextensions 11045 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_4_expressions 2.7_postgresql_psycopg2_nocextensions 12747 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_4_expressions 3.2_postgresql_psycopg2_nocextensions 11849 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_4_expressions 3.3_postgresql_psycopg2_nocextensions 11803 # TEST: test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_5_aggregates -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_5_aggregates 2.7_postgresql_psycopg2_cextensions 1005 -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_5_aggregates 2.7_postgresql_psycopg2_nocextensions 1109 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_5_aggregates 2.7_postgresql_psycopg2_cextensions 1050 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_5_aggregates 2.7_postgresql_psycopg2_nocextensions 1167 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_5_aggregates 3.2_postgresql_psycopg2_nocextensions 1114 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_5_aggregates 3.3_postgresql_psycopg2_nocextensions 1106 # TEST: test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_6_editing -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_6_editing 2.7_postgresql_psycopg2_cextensions 1736 -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_6_editing 2.7_postgresql_psycopg2_nocextensions 1779 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_6_editing 2.7_postgresql_psycopg2_cextensions 1811 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_6_editing 2.7_postgresql_psycopg2_nocextensions 1858 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_6_editing 3.2_postgresql_psycopg2_nocextensions 1731 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_6_editing 3.3_postgresql_psycopg2_nocextensions 1721 # TEST: test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_7_multiview -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_7_multiview 2.7_postgresql_psycopg2_cextensions 2219 -test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_7_multiview 2.7_postgresql_psycopg2_nocextensions 2449 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_7_multiview 2.7_postgresql_psycopg2_cextensions 2300 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_7_multiview 2.7_postgresql_psycopg2_nocextensions 2559 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_7_multiview 3.2_postgresql_psycopg2_nocextensions 2483 +test.aaa_profiling.test_zoomark.ZooMarkTest.test_profile_7_multiview 3.3_postgresql_psycopg2_nocextensions 2473 # TEST: test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_1a_populate -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_1a_populate 2.7_postgresql_psycopg2_cextensions 5977 -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_1a_populate 2.7_postgresql_psycopg2_nocextensions 6096 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_1a_populate 2.7_postgresql_psycopg2_cextensions 6157 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_1a_populate 2.7_postgresql_psycopg2_nocextensions 6276 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_1a_populate 3.2_postgresql_psycopg2_nocextensions 6252 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_1a_populate 3.3_postgresql_psycopg2_nocextensions 6251 # TEST: test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_2_insert -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_2_insert 2.7_postgresql_psycopg2_cextensions 392 -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_2_insert 2.7_postgresql_psycopg2_nocextensions 399 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_2_insert 2.7_postgresql_psycopg2_cextensions 391 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_2_insert 2.7_postgresql_psycopg2_nocextensions 398 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_2_insert 3.2_postgresql_psycopg2_nocextensions 395 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_2_insert 3.3_postgresql_psycopg2_nocextensions 394 # TEST: test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_3_properties -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_3_properties 2.7_postgresql_psycopg2_cextensions 6124 -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_3_properties 2.7_postgresql_psycopg2_nocextensions 6356 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_3_properties 2.7_postgresql_psycopg2_cextensions 6422 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_3_properties 2.7_postgresql_psycopg2_nocextensions 6654 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_3_properties 3.2_postgresql_psycopg2_nocextensions 6560 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_3_properties 3.3_postgresql_psycopg2_nocextensions 6560 # TEST: test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_4_expressions -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_4_expressions 2.7_postgresql_psycopg2_cextensions 18140 -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_4_expressions 2.7_postgresql_psycopg2_nocextensions 19571 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_4_expressions 2.7_postgresql_psycopg2_cextensions 19145 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_4_expressions 2.7_postgresql_psycopg2_nocextensions 20576 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_4_expressions 3.2_postgresql_psycopg2_nocextensions 20279 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_4_expressions 3.3_postgresql_psycopg2_nocextensions 20279 # TEST: test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_5_aggregates -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_5_aggregates 2.7_postgresql_psycopg2_cextensions 1018 -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_5_aggregates 2.7_postgresql_psycopg2_nocextensions 1114 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_5_aggregates 2.7_postgresql_psycopg2_cextensions 1063 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_5_aggregates 2.7_postgresql_psycopg2_nocextensions 1171 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_5_aggregates 3.2_postgresql_psycopg2_nocextensions 1120 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_5_aggregates 3.3_postgresql_psycopg2_nocextensions 1113 # TEST: test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_6_editing -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_6_editing 2.7_postgresql_psycopg2_cextensions 2614 -test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_6_editing 2.7_postgresql_psycopg2_nocextensions 2677 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_6_editing 2.7_postgresql_psycopg2_cextensions 2686 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_6_editing 2.7_postgresql_psycopg2_nocextensions 2749 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_6_editing 3.2_postgresql_psycopg2_nocextensions 2749 +test.aaa_profiling.test_zoomark_orm.ZooMarkTest.test_profile_6_editing 3.3_postgresql_psycopg2_nocextensions 2749 diff --git a/test/requirements.py b/test/requirements.py index c20bcbd0a..c09c0df99 100644 --- a/test/requirements.py +++ b/test/requirements.py @@ -556,6 +556,15 @@ class DefaultRequirements(SuiteRequirements): "cPython interpreter needed" ) + + @property + def non_broken_pickle(self): + from sqlalchemy.util import pickle + return only_if( + lambda: pickle.__name__ == 'cPickle' or sys.version_info >= (3, 2), + "Needs cPickle or newer Python 3 pickle" + ) + @property def predictable_gc(self): """target platform must remove all cycles unconditionally when diff --git a/test/sql/test_compiler.py b/test/sql/test_compiler.py index 83a02a49e..1ab45a295 100644 --- a/test/sql/test_compiler.py +++ b/test/sql/test_compiler.py @@ -20,6 +20,7 @@ from sqlalchemy import Integer, String, MetaData, Table, Column, select, \ intersect, union_all, Boolean, distinct, join, outerjoin, asc, desc,\ over, subquery, case import decimal +from sqlalchemy.util import u from sqlalchemy import exc, sql, util, types, schema from sqlalchemy.sql import table, column, label from sqlalchemy.sql.expression import ClauseList, _literal_as_text, HasPrefixes @@ -661,13 +662,13 @@ class SelectTest(fixtures.TestBase, AssertsCompiledSQL): s = select([table1.c.myid]).as_scalar() try: s.c.foo - except exc.InvalidRequestError, err: + except exc.InvalidRequestError as err: assert str(err) \ == 'Scalar Select expression has no columns; use this '\ 'object directly within a column-level expression.' try: s.columns.foo - except exc.InvalidRequestError, err: + except exc.InvalidRequestError as err: assert str(err) \ == 'Scalar Select expression has no columns; use this '\ 'object directly within a column-level expression.' @@ -734,13 +735,14 @@ class SelectTest(fixtures.TestBase, AssertsCompiledSQL): 'JOIN myothertable ON mytable.myid = ' 'myothertable.otherid') - def test_label_comparison(self): + def test_label_comparison_one(self): x = func.lala(table1.c.myid).label('foo') self.assert_compile(select([x], x == 5), 'SELECT lala(mytable.myid) AS foo FROM ' 'mytable WHERE lala(mytable.myid) = ' ':param_1') + def test_label_comparison_two(self): self.assert_compile( label('bar', column('foo', type_=String)) + 'foo', 'foo || :param_1') @@ -1184,9 +1186,9 @@ class SelectTest(fixtures.TestBase, AssertsCompiledSQL): # test unicode self.assert_compile(select( - [u"foobar(a)", u"pk_foo_bar(syslaal)"], - u"a = 12", - from_obj=[u"foobar left outer join lala on foobar.foo = lala.foo"] + ["foobar(a)", "pk_foo_bar(syslaal)"], + "a = 12", + from_obj=["foobar left outer join lala on foobar.foo = lala.foo"] ), "SELECT foobar(a), pk_foo_bar(syslaal) FROM foobar " "left outer join lala on foobar.foo = lala.foo WHERE a = 12" @@ -2313,7 +2315,7 @@ class SelectTest(fixtures.TestBase, AssertsCompiledSQL): func.lala(table1.c.name).label('gg')]) eq_( - s1.c.keys(), + list(s1.c.keys()), ['myid', 'foobar', str(f1), 'gg'] ) @@ -2341,7 +2343,7 @@ class SelectTest(fixtures.TestBase, AssertsCompiledSQL): t = table1 s1 = select([col], from_obj=t) - assert s1.c.keys() == [key], s1.c.keys() + assert list(s1.c.keys()) == [key], list(s1.c.keys()) if label: self.assert_compile(s1, @@ -2747,11 +2749,11 @@ class DDLTest(fixtures.TestBase, AssertsCompiledSQL): def test_reraise_of_column_spec_issue_unicode(self): MyType = self._illegal_type_fixture() t1 = Table('t', MetaData(), - Column(u'méil', MyType()) + Column(u('méil'), MyType()) ) assert_raises_message( exc.CompileError, - ur"\(in table 't', column 'méil'\): Couldn't compile type", + u(r"\(in table 't', column 'méil'\): Couldn't compile type"), schema.CreateTable(t1).compile ) diff --git a/test/sql/test_defaults.py b/test/sql/test_defaults.py index 79514eaf4..1508c0532 100644 --- a/test/sql/test_defaults.py +++ b/test/sql/test_defaults.py @@ -11,6 +11,10 @@ from sqlalchemy.types import TypeDecorator, TypeEngine from sqlalchemy.testing.schema import Table, Column from sqlalchemy.dialects import sqlite from sqlalchemy.testing import fixtures +from sqlalchemy.util import u, b +from sqlalchemy import util + +t = f = f2 = ts = currenttime = metadata = default_generator = None t = f = f2 = ts = currenttime = metadata = default_generator = None @@ -647,7 +651,7 @@ class SequenceExecTest(fixtures.TestBase): def _assert_seq_result(self, ret): """asserts return of next_value is an int""" - assert isinstance(ret, (int, long)) + assert isinstance(ret, util.int_types) assert ret > 0 def test_implicit_connectionless(self): @@ -779,7 +783,7 @@ class SequenceTest(fixtures.TestBase, testing.AssertsCompiledSQL): ] start = seq.start or 1 inc = seq.increment or 1 - assert values == list(xrange(start, start + inc * 3, inc)) + assert values == list(range(start, start + inc * 3, inc)) finally: seq.drop(testing.db) @@ -1157,20 +1161,12 @@ class UnicodeDefaultsTest(fixtures.TestBase): Column(Unicode(32)) def test_unicode_default(self): - # Py3K - #default = 'foo' - # Py2K - default = u'foo' - # end Py2K + default = u('foo') Column(Unicode(32), default=default) def test_nonunicode_default(self): - # Py3K - #default = b'foo' - # Py2K - default = 'foo' - # end Py2K + default = b('foo') assert_raises_message( sa.exc.SAWarning, "Unicode column received non-unicode default value.", diff --git a/test/sql/test_functions.py b/test/sql/test_functions.py index b325b7763..ee503dbb7 100644 --- a/test/sql/test_functions.py +++ b/test/sql/test_functions.py @@ -381,7 +381,7 @@ class ExecuteTest(fixtures.TestBase): assert t.select(t.c.id == id).execute().first()['value'] == 9 t.update(values={t.c.value: func.length("asdf")}).execute() assert t.select().execute().first()['value'] == 4 - print "--------------------------" + print("--------------------------") t2.insert().execute() t2.insert(values=dict(value=func.length("one"))).execute() t2.insert(values=dict(value=func.length("asfda") + -19)).\ @@ -409,7 +409,7 @@ class ExecuteTest(fixtures.TestBase): t2.update(values={t2.c.value: func.length("asfdaasdf"), t2.c.stuff: "foo"}).execute() - print "HI", select([t2.c.value, t2.c.stuff]).execute().first() + print("HI", select([t2.c.value, t2.c.stuff]).execute().first()) eq_(select([t2.c.value, t2.c.stuff]).execute().first(), (9, "foo") ) diff --git a/test/sql/test_generative.py b/test/sql/test_generative.py index 8b2abef0e..09b20d8ea 100644 --- a/test/sql/test_generative.py +++ b/test/sql/test_generative.py @@ -176,7 +176,7 @@ class BinaryEndpointTraversalTest(fixtures.TestBase): canary = [] def visit(binary, l, r): canary.append((binary.operator, l, r)) - print binary.operator, l, r + print(binary.operator, l, r) sql_util.visit_binary_product(visit, expr) eq_( canary, expected @@ -433,7 +433,7 @@ class ClauseTest(fixtures.TestBase, AssertsCompiledSQL): clause2 = Vis().traverse(clause) assert c1 == str(clause) assert str(clause2) == c1 + " SOME MODIFIER=:lala" - assert clause.bindparams.keys() == ['bar'] + assert list(clause.bindparams.keys()) == ['bar'] assert set(clause2.bindparams.keys()) == set(['bar', 'lala']) def test_select(self): @@ -446,8 +446,8 @@ class ClauseTest(fixtures.TestBase, AssertsCompiledSQL): s3 = Vis().traverse(s2) assert str(s3) == s3_assert assert str(s2) == s2_assert - print str(s2) - print str(s3) + print(str(s2)) + print(str(s3)) class Vis(ClauseVisitor): def visit_select(self, select): select.append_whereclause(t1.c.col2 == 7) @@ -459,8 +459,8 @@ class ClauseTest(fixtures.TestBase, AssertsCompiledSQL): def visit_select(self, select): select.append_whereclause(t1.c.col3 == 9) s4 = Vis().traverse(s3) - print str(s3) - print str(s4) + print(str(s3)) + print(str(s4)) assert str(s4) == s4_assert assert str(s3) == s3_assert @@ -471,8 +471,8 @@ class ClauseTest(fixtures.TestBase, AssertsCompiledSQL): binary.left = t1.c.col1 binary.right = bindparam("col1", unique=True) s5 = Vis().traverse(s4) - print str(s4) - print str(s5) + print(str(s4)) + print(str(s5)) assert str(s5) == s5_assert assert str(s4) == s4_assert diff --git a/test/sql/test_metadata.py b/test/sql/test_metadata.py index db2eaa4fa..c0873862d 100644 --- a/test/sql/test_metadata.py +++ b/test/sql/test_metadata.py @@ -89,7 +89,7 @@ class MetaDataTest(fixtures.TestBase, ComparesTables): msgs.append("attach %s.%s" % (t.name, c.name)) c1 = Column('foo', String()) m = MetaData() - for i in xrange(3): + for i in range(3): cx = c1.copy() # as of 0.7, these events no longer copy. its expected # that listeners will be re-established from the @@ -511,7 +511,7 @@ class MetaDataTest(fixtures.TestBase, ComparesTables): def _get_key(i): return [i.name, i.unique] + \ sorted(i.kwargs.items()) + \ - i.columns.keys() + list(i.columns.keys()) eq_( sorted([_get_key(i) for i in table.indexes]), diff --git a/test/sql/test_query.py b/test/sql/test_query.py index 293e629c8..ae029b11c 100644 --- a/test/sql/test_query.py +++ b/test/sql/test_query.py @@ -367,10 +367,10 @@ class QueryTest(fixtures.TestBase): ) if use_labels: eq_(result[0]['query_users_user_id'], 7) - eq_(result[0].keys(), ["query_users_user_id", "query_users_user_name"]) + eq_(list(result[0].keys()), ["query_users_user_id", "query_users_user_name"]) else: eq_(result[0]['user_id'], 7) - eq_(result[0].keys(), ["user_id", "user_name"]) + eq_(list(result[0].keys()), ["user_id", "user_name"]) eq_(result[0][0], 7) eq_(result[0][users.c.user_id], 7) @@ -523,13 +523,13 @@ class QueryTest(fixtures.TestBase): def a_eq(got, wanted): if got != wanted: - print "Wanted %s" % wanted - print "Received %s" % got + print("Wanted %s" % wanted) + print("Received %s" % got) self.assert_(got == wanted, got) a_eq(prep('select foo'), 'select foo') a_eq(prep("time='12:30:00'"), "time='12:30:00'") - a_eq(prep(u"time='12:30:00'"), u"time='12:30:00'") + a_eq(prep("time='12:30:00'"), "time='12:30:00'") a_eq(prep(":this:that"), ":this:that") a_eq(prep(":this :that"), "? ?") a_eq(prep("(:this),(:that :other)"), "(?),(? ?)") @@ -708,8 +708,6 @@ class QueryTest(fixtures.TestBase): use_labels=labels), [(3, 'a'), (2, 'b'), (1, None)]) - @testing.fails_on('mssql+pyodbc', - "pyodbc result row doesn't support slicing") def test_column_slices(self): users.insert().execute(user_id=1, user_name='john') users.insert().execute(user_id=2, user_name='jack') @@ -769,7 +767,7 @@ class QueryTest(fixtures.TestBase): ).first() eq_(r['user_id'], 1) eq_(r['user_name'], "john") - eq_(r.keys(), ["user_id", "user_name"]) + eq_(list(r.keys()), ["user_id", "user_name"]) @testing.only_on("sqlite", "sqlite specific feature") def test_column_accessor_sqlite_raw(self): @@ -784,7 +782,7 @@ class QueryTest(fixtures.TestBase): assert 'user_name' not in r eq_(r['query_users.user_id'], 1) eq_(r['query_users.user_name'], "john") - eq_(r.keys(), ["query_users.user_id", "query_users.user_name"]) + eq_(list(r.keys()), ["query_users.user_id", "query_users.user_name"]) @testing.only_on("sqlite", "sqlite specific feature") def test_column_accessor_sqlite_translated(self): @@ -799,7 +797,7 @@ class QueryTest(fixtures.TestBase): eq_(r['user_name'], "john") eq_(r['query_users.user_id'], 1) eq_(r['query_users.user_name'], "john") - eq_(r.keys(), ["user_id", "user_name"]) + eq_(list(r.keys()), ["user_id", "user_name"]) def test_column_accessor_labels_w_dots(self): users.insert().execute( @@ -812,7 +810,7 @@ class QueryTest(fixtures.TestBase): eq_(r['query_users.user_id'], 1) eq_(r['query_users.user_name'], "john") assert "user_name" not in r - eq_(r.keys(), ["query_users.user_id", "query_users.user_name"]) + eq_(list(r.keys()), ["query_users.user_id", "query_users.user_name"]) def test_column_accessor_unary(self): users.insert().execute( @@ -889,7 +887,7 @@ class QueryTest(fixtures.TestBase): ]) ).first() - eq_(row.keys(), ["case_insensitive", "CaseSensitive"]) + eq_(list(row.keys()), ["case_insensitive", "CaseSensitive"]) eq_(row["case_insensitive"], 1) eq_(row["CaseSensitive"], 2) @@ -911,7 +909,7 @@ class QueryTest(fixtures.TestBase): ]) ).first() - eq_(row.keys(), ["case_insensitive", "CaseSensitive"]) + eq_(list(row.keys()), ["case_insensitive", "CaseSensitive"]) eq_(row["case_insensitive"], 1) eq_(row["CaseSensitive"], 2) eq_(row["Case_insensitive"],1) @@ -1072,14 +1070,14 @@ class QueryTest(fixtures.TestBase): def test_keys(self): users.insert().execute(user_id=1, user_name='foo') r = users.select().execute() - eq_([x.lower() for x in r.keys()], ['user_id', 'user_name']) + eq_([x.lower() for x in list(r.keys())], ['user_id', 'user_name']) r = r.first() - eq_([x.lower() for x in r.keys()], ['user_id', 'user_name']) + eq_([x.lower() for x in list(r.keys())], ['user_id', 'user_name']) def test_items(self): users.insert().execute(user_id=1, user_name='foo') r = users.select().execute().first() - eq_([(x[0].lower(), x[1]) for x in r.items()], [('user_id', 1), ('user_name', 'foo')]) + eq_([(x[0].lower(), x[1]) for x in list(r.items())], [('user_id', 1), ('user_name', 'foo')]) def test_len(self): users.insert().execute(user_id=1, user_name='foo') @@ -1098,8 +1096,8 @@ class QueryTest(fixtures.TestBase): r = users.select(users.c.user_id==1).execute().first() eq_(r[0], 1) eq_(r[1], 'foo') - eq_([x.lower() for x in r.keys()], ['user_id', 'user_name']) - eq_(r.values(), [1, 'foo']) + eq_([x.lower() for x in list(r.keys())], ['user_id', 'user_name']) + eq_(list(r.values()), [1, 'foo']) def test_column_order_with_text_query(self): # should return values in query order @@ -1107,8 +1105,8 @@ class QueryTest(fixtures.TestBase): r = testing.db.execute('select user_name, user_id from query_users').first() eq_(r[0], 'foo') eq_(r[1], 1) - eq_([x.lower() for x in r.keys()], ['user_name', 'user_id']) - eq_(r.values(), ['foo', 1]) + eq_([x.lower() for x in list(r.keys())], ['user_name', 'user_id']) + eq_(list(r.values()), ['foo', 1]) @testing.crashes('oracle', 'FIXME: unknown, varify not fails_on()') @testing.crashes('firebird', 'An identifier must begin with a letter') @@ -1137,7 +1135,7 @@ class QueryTest(fixtures.TestBase): self.assert_(r['_parent'] == 'Hidden parent') self.assert_(r['_row'] == 'Hidden row') try: - print r._parent, r._row + print(r._parent, r._row) self.fail('Should not allow access to private attributes') except AttributeError: pass # expected @@ -2334,7 +2332,7 @@ class JoinTest(fixtures.TestBase): expr = select( [t1.c.t1_id, t2.c.t2_id, t3.c.t3_id], from_obj=[(t1.join(t2).outerjoin(t3, criteria))]) - print expr + print(expr) self.assertRows(expr, [(10, 20, 30), (11, 21, None)]) def test_mixed_where(self): @@ -2416,7 +2414,7 @@ class OperatorTest(fixtures.TestBase): select([ flds.c.intcol, func.row_number().over(order_by=flds.c.strcol) ]).execute().fetchall(), - [(13, 1L), (5, 2L)] + [(13, 1), (5, 2)] ) diff --git a/test/sql/test_quote.py b/test/sql/test_quote.py index 8b14d23a9..717f0f797 100644 --- a/test/sql/test_quote.py +++ b/test/sql/test_quote.py @@ -552,8 +552,8 @@ class PreparerTest(fixtures.TestBase): def a_eq(have, want): if have != want: - print "Wanted %s" % want - print "Received %s" % have + print("Wanted %s" % want) + print("Received %s" % have) self.assert_(have == want) a_eq(unformat('foo'), ['foo']) @@ -584,13 +584,13 @@ class PreparerTest(fixtures.TestBase): def a_eq(have, want): if have != want: - print "Wanted %s" % want - print "Received %s" % have + print("Wanted %s" % want) + print("Received %s" % have) self.assert_(have == want) a_eq(unformat('foo'), ['foo']) a_eq(unformat('`foo`'), ['foo']) - a_eq(unformat(`'foo'`), ["'foo'"]) + a_eq(unformat(repr('foo')), ["'foo'"]) a_eq(unformat('foo.bar'), ['foo', 'bar']) a_eq(unformat('`foo`.`bar`'), ['foo', 'bar']) a_eq(unformat('foo.`bar`'), ['foo', 'bar']) diff --git a/test/sql/test_rowcount.py b/test/sql/test_rowcount.py index f14f78989..2dbf4f3ea 100644 --- a/test/sql/test_rowcount.py +++ b/test/sql/test_rowcount.py @@ -53,20 +53,20 @@ class FoundRowsTest(fixtures.TestBase, AssertsExecutionResults): # WHERE matches 3, 3 rows changed department = employees_table.c.department r = employees_table.update(department=='C').execute(department='Z') - print "expecting 3, dialect reports %s" % r.rowcount + print("expecting 3, dialect reports %s" % r.rowcount) assert r.rowcount == 3 def test_update_rowcount2(self): # WHERE matches 3, 0 rows changed department = employees_table.c.department r = employees_table.update(department=='C').execute(department='C') - print "expecting 3, dialect reports %s" % r.rowcount + print("expecting 3, dialect reports %s" % r.rowcount) assert r.rowcount == 3 def test_delete_rowcount(self): # WHERE matches 3, 3 rows deleted department = employees_table.c.department r = employees_table.delete(department=='C').execute() - print "expecting 3, dialect reports %s" % r.rowcount + print("expecting 3, dialect reports %s" % r.rowcount) assert r.rowcount == 3 diff --git a/test/sql/test_selectable.py b/test/sql/test_selectable.py index 183b72111..dc0b040b0 100644 --- a/test/sql/test_selectable.py +++ b/test/sql/test_selectable.py @@ -174,11 +174,11 @@ class SelectableTest(fixtures.TestBase, AssertsExecutionResults, AssertsCompiled def test_clone_append_column(self): sel = select([literal_column('1').label('a')]) - eq_(sel.c.keys(), ['a']) + eq_(list(sel.c.keys()), ['a']) cloned = visitors.ReplacingCloningVisitor().traverse(sel) cloned.append_column(literal_column('2').label('b')) cloned.append_column(func.foo()) - eq_(cloned.c.keys(), ['a', 'b', 'foo()']) + eq_(list(cloned.c.keys()), ['a', 'b', 'foo()']) def test_append_column_after_replace_selectable(self): basesel = select([literal_column('1').label('a')]) @@ -362,10 +362,10 @@ class SelectableTest(fixtures.TestBase, AssertsExecutionResults, AssertsCompiled def test_join(self): a = join(table1, table2) - print str(a.select(use_labels=True)) + print(str(a.select(use_labels=True))) b = table2.alias('b') j = join(a, b) - print str(j) + print(str(j)) criterion = a.c.table1_col1 == b.c.col2 self.assert_(criterion.compare(j.onclause)) @@ -949,7 +949,7 @@ class PrimaryKeyTest(fixtures.TestBase, AssertsExecutionResults): primary_key=True), Column('x', Integer)) d = Table('d', meta, Column('id', Integer, ForeignKey('c.id'), primary_key=True), Column('x', Integer)) - print list(a.join(b, a.c.x == b.c.id).primary_key) + print(list(a.join(b, a.c.x == b.c.id).primary_key)) assert list(a.join(b, a.c.x == b.c.id).primary_key) == [a.c.id] assert list(b.join(c, b.c.x == c.c.id).primary_key) == [b.c.id] assert list(a.join(b).join(c, c.c.id == b.c.x).primary_key) \ @@ -1618,7 +1618,7 @@ class WithLabelsTest(fixtures.TestBase): def test_names_overlap_label(self): sel = self._names_overlap().apply_labels() eq_( - sel.c.keys(), + list(sel.c.keys()), ['t1_x', 't2_x'] ) self._assert_result_keys(sel, ['t1_x', 't2_x']) @@ -1632,7 +1632,7 @@ class WithLabelsTest(fixtures.TestBase): def test_names_overlap_keys_dont_nolabel(self): sel = self._names_overlap_keys_dont() eq_( - sel.c.keys(), + list(sel.c.keys()), ['a', 'b'] ) self._assert_result_keys(sel, ['x']) @@ -1640,7 +1640,7 @@ class WithLabelsTest(fixtures.TestBase): def test_names_overlap_keys_dont_label(self): sel = self._names_overlap_keys_dont().apply_labels() eq_( - sel.c.keys(), + list(sel.c.keys()), ['t1_a', 't2_b'] ) self._assert_result_keys(sel, ['t1_x', 't2_x']) @@ -1654,7 +1654,7 @@ class WithLabelsTest(fixtures.TestBase): def test_labels_overlap_nolabel(self): sel = self._labels_overlap() eq_( - sel.c.keys(), + list(sel.c.keys()), ['x_id', 'id'] ) self._assert_result_keys(sel, ['x_id', 'id']) @@ -1663,7 +1663,7 @@ class WithLabelsTest(fixtures.TestBase): sel = self._labels_overlap().apply_labels() t2 = sel.froms[1] eq_( - sel.c.keys(), + list(sel.c.keys()), ['t_x_id', t2.c.id.anon_label] ) self._assert_result_keys(sel, ['t_x_id', 'id_1']) @@ -1677,12 +1677,12 @@ class WithLabelsTest(fixtures.TestBase): def test_labels_overlap_keylabels_dont_nolabel(self): sel = self._labels_overlap_keylabels_dont() - eq_(sel.c.keys(), ['a', 'b']) + eq_(list(sel.c.keys()), ['a', 'b']) self._assert_result_keys(sel, ['x_id', 'id']) def test_labels_overlap_keylabels_dont_label(self): sel = self._labels_overlap_keylabels_dont().apply_labels() - eq_(sel.c.keys(), ['t_a', 't_x_b']) + eq_(list(sel.c.keys()), ['t_a', 't_x_b']) self._assert_result_keys(sel, ['t_x_id', 'id_1']) def _keylabels_overlap_labels_dont(self): @@ -1693,13 +1693,13 @@ class WithLabelsTest(fixtures.TestBase): def test_keylabels_overlap_labels_dont_nolabel(self): sel = self._keylabels_overlap_labels_dont() - eq_(sel.c.keys(), ['x_id', 'id']) + eq_(list(sel.c.keys()), ['x_id', 'id']) self._assert_result_keys(sel, ['a', 'b']) def test_keylabels_overlap_labels_dont_label(self): sel = self._keylabels_overlap_labels_dont().apply_labels() t2 = sel.froms[1] - eq_(sel.c.keys(), ['t_x_id', t2.c.id.anon_label]) + eq_(list(sel.c.keys()), ['t_x_id', t2.c.id.anon_label]) self._assert_result_keys(sel, ['t_a', 't_x_b']) self._assert_subq_result_keys(sel, ['t_a', 't_x_b']) @@ -1711,14 +1711,14 @@ class WithLabelsTest(fixtures.TestBase): def test_keylabels_overlap_labels_overlap_nolabel(self): sel = self._keylabels_overlap_labels_overlap() - eq_(sel.c.keys(), ['x_a', 'a']) + eq_(list(sel.c.keys()), ['x_a', 'a']) self._assert_result_keys(sel, ['x_id', 'id']) self._assert_subq_result_keys(sel, ['x_id', 'id']) def test_keylabels_overlap_labels_overlap_label(self): sel = self._keylabels_overlap_labels_overlap().apply_labels() t2 = sel.froms[1] - eq_(sel.c.keys(), ['t_x_a', t2.c.a.anon_label]) + eq_(list(sel.c.keys()), ['t_x_a', t2.c.a.anon_label]) self._assert_result_keys(sel, ['t_x_id', 'id_1']) self._assert_subq_result_keys(sel, ['t_x_id', 'id_1']) @@ -1736,7 +1736,7 @@ class WithLabelsTest(fixtures.TestBase): def test_keys_overlap_names_dont_label(self): sel = self._keys_overlap_names_dont().apply_labels() eq_( - sel.c.keys(), + list(sel.c.keys()), ['t1_x', 't2_x'] ) self._assert_result_keys(sel, ['t1_a', 't2_b']) diff --git a/test/sql/test_types.py b/test/sql/test_types.py index 64dbb6204..dbb475b98 100644 --- a/test/sql/test_types.py +++ b/test/sql/test_types.py @@ -141,22 +141,14 @@ class AdaptTest(fixtures.TestBase): eq_(types.Integer().python_type, int) eq_(types.Numeric().python_type, decimal.Decimal) eq_(types.Numeric(asdecimal=False).python_type, float) - # Py3K - #eq_(types.LargeBinary().python_type, bytes) - # Py2K - eq_(types.LargeBinary().python_type, str) - # end Py2K + eq_(types.LargeBinary().python_type, util.binary_type) eq_(types.Float().python_type, float) eq_(types.Interval().python_type, datetime.timedelta) eq_(types.Date().python_type, datetime.date) eq_(types.DateTime().python_type, datetime.datetime) - # Py3K - #eq_(types.String().python_type, unicode) - # Py2K eq_(types.String().python_type, str) - # end Py2K - eq_(types.Unicode().python_type, unicode) - eq_(types.String(convert_unicode=True).python_type, unicode) + eq_(types.Unicode().python_type, util.text_type) + eq_(types.String(convert_unicode=True).python_type, util.text_type) assert_raises( NotImplementedError, @@ -257,14 +249,14 @@ class UserDefinedTest(fixtures.TablesTest, AssertsCompiledSQL): def test_processing(self): users = self.tables.users users.insert().execute( - user_id=2, goofy='jack', goofy2='jack', goofy4=u'jack', - goofy7=u'jack', goofy8=12, goofy9=12) + user_id=2, goofy='jack', goofy2='jack', goofy4=util.u('jack'), + goofy7=util.u('jack'), goofy8=12, goofy9=12) users.insert().execute( - user_id=3, goofy='lala', goofy2='lala', goofy4=u'lala', - goofy7=u'lala', goofy8=15, goofy9=15) + user_id=3, goofy='lala', goofy2='lala', goofy4=util.u('lala'), + goofy7=util.u('lala'), goofy8=15, goofy9=15) users.insert().execute( - user_id=4, goofy='fred', goofy2='fred', goofy4=u'fred', - goofy7=u'fred', goofy8=9, goofy9=9) + user_id=4, goofy='fred', goofy2='fred', goofy4=util.u('fred'), + goofy7=util.u('fred'), goofy8=9, goofy9=9) l = users.select().order_by(users.c.user_id).execute().fetchall() for assertstr, assertint, assertint2, row in zip( @@ -278,7 +270,7 @@ class UserDefinedTest(fixtures.TablesTest, AssertsCompiledSQL): eq_(row[5], assertint) eq_(row[6], assertint2) for col in row[3], row[4]: - assert isinstance(col, unicode) + assert isinstance(col, util.text_type) def test_typedecorator_impl(self): for impl_, exp, kw in [ @@ -715,9 +707,9 @@ class UnicodeTest(fixtures.TestBase): expected ) - data = u"Alors vous imaginez ma surprise, au lever du jour, quand "\ - u"une drôle de petite voix m’a réveillé. "\ - u"Elle disait: « S’il vous plaît… dessine-moi un mouton! »" + data = util.u("Alors vous imaginez ma surprise, au lever du jour, quand "\ + "une drôle de petite voix m’a réveillé. "\ + "Elle disait: « S’il vous plaît… dessine-moi un mouton! »") def test_unicode_warnings_typelevel_native_unicode(self): @@ -726,13 +718,12 @@ class UnicodeTest(fixtures.TestBase): dialect = default.DefaultDialect() dialect.supports_unicode_binds = True uni = u.dialect_impl(dialect).bind_processor(dialect) - # Py3K - #assert_raises(exc.SAWarning, uni, b'x') - #assert isinstance(uni(unicodedata), str) - # Py2K - assert_raises(exc.SAWarning, uni, 'x') - assert isinstance(uni(unicodedata), unicode) - # end Py2K + if util.py3k: + assert_raises(exc.SAWarning, uni, b'x') + assert isinstance(uni(unicodedata), str) + else: + assert_raises(exc.SAWarning, uni, 'x') + assert isinstance(uni(unicodedata), unicode) def test_unicode_warnings_typelevel_sqla_unicode(self): unicodedata = self.data @@ -740,13 +731,8 @@ class UnicodeTest(fixtures.TestBase): dialect = default.DefaultDialect() dialect.supports_unicode_binds = False uni = u.dialect_impl(dialect).bind_processor(dialect) - # Py3K - #assert_raises(exc.SAWarning, uni, b'x') - #assert isinstance(uni(unicodedata), bytes) - # Py2K - assert_raises(exc.SAWarning, uni, 'x') - assert isinstance(uni(unicodedata), str) - # end Py2K + assert_raises(exc.SAWarning, uni, util.b('x')) + assert isinstance(uni(unicodedata), util.binary_type) eq_(uni(unicodedata), unicodedata.encode('utf-8')) @@ -759,14 +745,9 @@ class UnicodeTest(fixtures.TestBase): s = String() uni = s.dialect_impl(dialect).bind_processor(dialect) - # this is not the unicode type - no warning - # Py3K - #uni(b'x') - #assert isinstance(uni(unicodedata), bytes) - # Py2K - uni('x') - assert isinstance(uni(unicodedata), str) - # end Py2K + + uni(util.b('x')) + assert isinstance(uni(unicodedata), util.binary_type) eq_(uni(unicodedata), unicodedata.encode('utf-8')) @@ -1368,7 +1349,7 @@ class NumericRawSQLTest(fixtures.TestBase): metadata = self.metadata self._fixture(metadata, Integer, 45) val = testing.db.execute("select val from t").scalar() - assert isinstance(val, (int, long)) + assert isinstance(val, util.int_types) eq_(val, 45) @testing.provide_metadata diff --git a/test/sql/test_unicode.py b/test/sql/test_unicode.py index 37e44522e..ffcef903f 100644 --- a/test/sql/test_unicode.py +++ b/test/sql/test_unicode.py @@ -7,6 +7,7 @@ from sqlalchemy import testing from sqlalchemy.testing.engines import utf8_engine from sqlalchemy.sql import column from sqlalchemy.testing.schema import Table, Column +from sqlalchemy.util import u, ue class UnicodeSchemaTest(fixtures.TestBase): __requires__ = ('unicode_ddl',) @@ -18,14 +19,14 @@ class UnicodeSchemaTest(fixtures.TestBase): unicode_bind = utf8_engine() metadata = MetaData(unicode_bind) - t1 = Table('unitable1', metadata, - Column(u'méil', Integer, primary_key=True), - Column(u'\u6e2c\u8a66', Integer), + t1 = Table(u('unitable1'), metadata, + Column(u('méil'), Integer, primary_key=True), + Column(ue('\u6e2c\u8a66'), Integer), test_needs_fk=True, ) - t2 = Table(u'Unitéble2', metadata, - Column(u'méil', Integer, primary_key=True, key="a"), - Column(u'\u6e2c\u8a66', Integer, ForeignKey(u'unitable1.méil'), + t2 = Table(u('Unitéble2'), metadata, + Column(u('méil'), Integer, primary_key=True, key="a"), + Column(ue('\u6e2c\u8a66'), Integer, ForeignKey(u('unitable1.méil')), key="b" ), test_needs_fk=True, @@ -33,27 +34,27 @@ class UnicodeSchemaTest(fixtures.TestBase): # Few DBs support Unicode foreign keys if testing.against('sqlite'): - t3 = Table(u'\u6e2c\u8a66', metadata, - Column(u'\u6e2c\u8a66_id', Integer, primary_key=True, + t3 = Table(ue('\u6e2c\u8a66'), metadata, + Column(ue('\u6e2c\u8a66_id'), Integer, primary_key=True, autoincrement=False), - Column(u'unitable1_\u6e2c\u8a66', Integer, - ForeignKey(u'unitable1.\u6e2c\u8a66') + Column(ue('unitable1_\u6e2c\u8a66'), Integer, + ForeignKey(ue('unitable1.\u6e2c\u8a66')) ), - Column(u'Unitéble2_b', Integer, - ForeignKey(u'Unitéble2.b') + Column(u('Unitéble2_b'), Integer, + ForeignKey(u('Unitéble2.b')) ), - Column(u'\u6e2c\u8a66_self', Integer, - ForeignKey(u'\u6e2c\u8a66.\u6e2c\u8a66_id') + Column(ue('\u6e2c\u8a66_self'), Integer, + ForeignKey(ue('\u6e2c\u8a66.\u6e2c\u8a66_id')) ), test_needs_fk=True, ) else: - t3 = Table(u'\u6e2c\u8a66', metadata, - Column(u'\u6e2c\u8a66_id', Integer, primary_key=True, + t3 = Table(ue('\u6e2c\u8a66'), metadata, + Column(ue('\u6e2c\u8a66_id'), Integer, primary_key=True, autoincrement=False), - Column(u'unitable1_\u6e2c\u8a66', Integer), - Column(u'Unitéble2_b', Integer), - Column(u'\u6e2c\u8a66_self', Integer), + Column(ue('unitable1_\u6e2c\u8a66'), Integer), + Column(u('Unitéble2_b'), Integer), + Column(ue('\u6e2c\u8a66_self'), Integer), test_needs_fk=True, ) metadata.create_all() @@ -72,42 +73,42 @@ class UnicodeSchemaTest(fixtures.TestBase): del unicode_bind def test_insert(self): - t1.insert().execute({u'méil':1, u'\u6e2c\u8a66':5}) - t2.insert().execute({'a':1, 'b':1}) - t3.insert().execute({u'\u6e2c\u8a66_id': 1, - u'unitable1_\u6e2c\u8a66': 5, - u'Unitéble2_b': 1, - u'\u6e2c\u8a66_self': 1}) + t1.insert().execute({u('méil'):1, ue('\u6e2c\u8a66'):5}) + t2.insert().execute({u('a'):1, u('b'):1}) + t3.insert().execute({ue('\u6e2c\u8a66_id'): 1, + ue('unitable1_\u6e2c\u8a66'): 5, + u('Unitéble2_b'): 1, + ue('\u6e2c\u8a66_self'): 1}) assert t1.select().execute().fetchall() == [(1, 5)] assert t2.select().execute().fetchall() == [(1, 1)] assert t3.select().execute().fetchall() == [(1, 5, 1, 1)] def test_reflect(self): - t1.insert().execute({u'méil':2, u'\u6e2c\u8a66':7}) - t2.insert().execute({'a':2, 'b':2}) - t3.insert().execute({u'\u6e2c\u8a66_id': 2, - u'unitable1_\u6e2c\u8a66': 7, - u'Unitéble2_b': 2, - u'\u6e2c\u8a66_self': 2}) + t1.insert().execute({u('méil'):2, ue('\u6e2c\u8a66'):7}) + t2.insert().execute({u('a'):2, u('b'):2}) + t3.insert().execute({ue('\u6e2c\u8a66_id'): 2, + ue('unitable1_\u6e2c\u8a66'): 7, + u('Unitéble2_b'): 2, + ue('\u6e2c\u8a66_self'): 2}) meta = MetaData(unicode_bind) tt1 = Table(t1.name, meta, autoload=True) tt2 = Table(t2.name, meta, autoload=True) tt3 = Table(t3.name, meta, autoload=True) - tt1.insert().execute({u'méil':1, u'\u6e2c\u8a66':5}) - tt2.insert().execute({u'méil':1, u'\u6e2c\u8a66':1}) - tt3.insert().execute({u'\u6e2c\u8a66_id': 1, - u'unitable1_\u6e2c\u8a66': 5, - u'Unitéble2_b': 1, - u'\u6e2c\u8a66_self': 1}) + tt1.insert().execute({u('méil'):1, ue('\u6e2c\u8a66'):5}) + tt2.insert().execute({u('méil'):1, ue('\u6e2c\u8a66'):1}) + tt3.insert().execute({ue('\u6e2c\u8a66_id'): 1, + ue('unitable1_\u6e2c\u8a66'): 5, + u('Unitéble2_b'): 1, + ue('\u6e2c\u8a66_self'): 1}) - self.assert_(tt1.select(order_by=desc(u'méil')).execute().fetchall() == + self.assert_(tt1.select(order_by=desc(u('méil'))).execute().fetchall() == [(2, 7), (1, 5)]) - self.assert_(tt2.select(order_by=desc(u'méil')).execute().fetchall() == + self.assert_(tt2.select(order_by=desc(u('méil'))).execute().fetchall() == [(2, 2), (1, 1)]) - self.assert_(tt3.select(order_by=desc(u'\u6e2c\u8a66_id')). + self.assert_(tt3.select(order_by=desc(ue('\u6e2c\u8a66_id'))). execute().fetchall() == [(2, 7, 2, 2), (1, 5, 1, 1)]) meta.drop_all() @@ -117,7 +118,7 @@ class EscapesDefaultsTest(fixtures.TestBase): def test_default_exec(self): metadata = MetaData(testing.db) t1 = Table('t1', metadata, - Column(u'special_col', Integer, Sequence('special_col'), primary_key=True), + Column('special_col', Integer, Sequence('special_col'), primary_key=True), Column('data', String(50)) # to appease SQLite without DEFAULT VALUES ) metadata.create_all() @@ -128,8 +129,8 @@ class EscapesDefaultsTest(fixtures.TestBase): # reset the identifier preparer, so that we can force it to cache # a unicode identifier engine.dialect.identifier_preparer = engine.dialect.preparer(engine.dialect) - select([column(u'special_col')]).select_from(t1).execute().close() - assert isinstance(engine.dialect.identifier_preparer.format_sequence(Sequence('special_col')), unicode) + select([column('special_col')]).select_from(t1).execute().close() + assert isinstance(engine.dialect.identifier_preparer.format_sequence(Sequence('special_col')), str) # now execute, run the sequence. it should run in u"Special_col.nextid" or similar as # a unicode object; cx_oracle asserts that this is None or a String (postgresql lets it pass thru). diff --git a/test/sql/test_update.py b/test/sql/test_update.py index a8df86cd2..8695760fb 100644 --- a/test/sql/test_update.py +++ b/test/sql/test_update.py @@ -242,7 +242,7 @@ class UpdateFromCompileTest(_UpdateFromTestBase, fixtures.TablesTest, 'WHERE ' 'users.id = addresses.user_id AND ' 'addresses.email_address = :email_address_1', - checkparams={u'email_address_1': 'e1', 'name': 'newname'}) + checkparams={'email_address_1': 'e1', 'name': 'newname'}) def test_render_multi_table(self): users = self.tables.users @@ -250,8 +250,8 @@ class UpdateFromCompileTest(_UpdateFromTestBase, fixtures.TablesTest, dingalings = self.tables.dingalings checkparams = { - u'email_address_1': 'e1', - u'id_1': 2, + 'email_address_1': 'e1', + 'id_1': 2, 'name': 'newname' } @@ -285,15 +285,15 @@ class UpdateFromCompileTest(_UpdateFromTestBase, fixtures.TablesTest, 'WHERE ' 'users.id = addresses.user_id AND ' 'addresses.email_address = %s', - checkparams={u'email_address_1': 'e1', 'name': 'newname'}, + checkparams={'email_address_1': 'e1', 'name': 'newname'}, dialect=mysql.dialect()) def test_render_subquery(self): users, addresses = self.tables.users, self.tables.addresses checkparams = { - u'email_address_1': 'e1', - u'id_1': 7, + 'email_address_1': 'e1', + 'id_1': 7, 'name': 'newname' } |
