diff options
| author | Mike Bayer <mike_mp@zzzcomputing.com> | 2017-03-06 12:26:01 -0500 |
|---|---|---|
| committer | Mike Bayer <mike_mp@zzzcomputing.com> | 2017-03-06 17:20:06 -0500 |
| commit | f4c4f784cde8e51301b09f187d2f086bfae47453 (patch) | |
| tree | f0fd96d4b7c59dfd3a662ae2b0f4bb0e9eee2884 /test/aaa_profiling | |
| parent | 66e483ebdca8f93d3e028a201e04f2be49c71c51 (diff) | |
| download | sqlalchemy-f4c4f784cde8e51301b09f187d2f086bfae47453.tar.gz | |
Don't cache savepoint identifiers
Fixed bug in compiler where the string identifier of a savepoint would
be cached in the identifier quoting dictionary; as these identifiers
are arbitrary, a small memory leak could occur if a single
:class:`.Connection` had an unbounded number of savepoints used,
as well as if the savepoint clause constructs were used directly
with an unbounded umber of savepoint names. The memory leak does
**not** impact the vast majority of cases as normally the
:class:`.Connection`, which renders savepoint names with a simple
counter starting at "1", is used on a per-transaction or
per-fixed-number-of-transactions basis before being discarded.
The savepoint name in virtually all cases does not require quoting
at all, however to support potential third party use cases
the "check for quotes needed" logic is retained, at a small
performance cost. Uncondtionally quoting the name is another
option, but this would turn the name into a case sensitive name
which runs the risk of poor interactions with existing deployments
that may be looking at these names in other contexts.
Change-Id: I6b53c96abf7fdf1840592bbca5da81347911844c
Fixes: #3931
Diffstat (limited to 'test/aaa_profiling')
| -rw-r--r-- | test/aaa_profiling/test_memusage.py | 50 |
1 files changed, 46 insertions, 4 deletions
diff --git a/test/aaa_profiling/test_memusage.py b/test/aaa_profiling/test_memusage.py index 7013159dd..53f118e15 100644 --- a/test/aaa_profiling/test_memusage.py +++ b/test/aaa_profiling/test_memusage.py @@ -34,7 +34,8 @@ class ASub(A): pass -def profile_memory(maxtimes=50): +def profile_memory(maxtimes=50, + assert_no_sessions=True, get_num_objects=None): def decorate(func): # run the test N times. if length of gc.get_objects() # keeps growing, assert false @@ -56,15 +57,19 @@ def profile_memory(maxtimes=50): samples = [] success = False - for y in range(maxtimes // 5): + for y in range(100 // 5): for x in range(5): func(*args) gc_collect() - samples.append(len(get_objects_skipping_sqlite_issue())) + samples.append( + get_num_objects() if get_num_objects is not None + else len(get_objects_skipping_sqlite_issue()) + ) print("sample gc sizes:", samples) - assert len(_sessions) == 0 + if assert_no_sessions: + assert len(_sessions) == 0 # check for "flatline" - size is constant for # 5 iterations @@ -341,6 +346,43 @@ class MemUsageTest(EnsureZeroed): finally: metadata.drop_all() + @testing.requires.savepoints + @testing.provide_metadata + def test_savepoints(self): + metadata = self.metadata + + some_table = Table( + 't', metadata, + Column('id', Integer, primary_key=True, + test_needs_autoincrement=True) + ) + + class SomeClass(object): + pass + + mapper(SomeClass, some_table) + + metadata.create_all() + + session = Session(testing.db) + + target_strings = session.connection().\ + dialect.identifier_preparer._strings + + with session.transaction: + @profile_memory( + assert_no_sessions=False, + get_num_objects=lambda: len(target_strings)) + def go(): + + sc = SomeClass() + session.add(sc) + + with session.begin_nested(): + session.query(SomeClass).first() + + go() + @testing.crashes('mysql+cymysql', 'blocking') def test_unicode_warnings(self): metadata = MetaData(self.engine) |
