diff options
author | Chandan Singh <chandan@chandansingh.net> | 2019-10-26 19:07:03 +0100 |
---|---|---|
committer | Chandan Singh <chandan@chandansingh.net> | 2019-11-04 11:01:14 +0000 |
commit | 576be1fc19bb58d25309a911589ceb9f08f7812d (patch) | |
tree | ad72fe64328e2cedd5415dfaa6eb7dd7ecc5576d /tests/artifactcache | |
parent | 66045d9c817992dbce855d4217ff00970ae7ff79 (diff) | |
download | buildstream-576be1fc19bb58d25309a911589ceb9f08f7812d.tar.gz |
frontend: Remove tracking options from `bst build`
BREAKING CHANGE: Remove all tracking related options and flags from
`bst build` command, as discussed on mailing list previously.
See
https://mail.gnome.org/archives/buildstream-list/2019-October/msg00009.html
for the original proposal, and
https://mail.gnome.org/archives/buildstream-list/2019-October/msg00024.html
for a summary.
At the same time, remove use of the removed options from our test suite.
Diffstat (limited to 'tests/artifactcache')
-rw-r--r-- | tests/artifactcache/expiry.py | 63 |
1 files changed, 1 insertions, 62 deletions
diff --git a/tests/artifactcache/expiry.py b/tests/artifactcache/expiry.py index d33034813..9ede1a8d3 100644 --- a/tests/artifactcache/expiry.py +++ b/tests/artifactcache/expiry.py @@ -29,7 +29,7 @@ from buildstream._cas import CASCache from buildstream._exceptions import ErrorDomain, LoadErrorReason from buildstream.testing import cli # pylint: disable=unused-import -from tests.testutils import create_element_size, update_element_size, wait_for_cache_granularity +from tests.testutils import create_element_size, wait_for_cache_granularity DATA_DIR = os.path.join( @@ -257,67 +257,6 @@ def test_never_delete_required(cli, datafiles): assert states['target.bst'] != 'cached' -# Assert that we never delete a dependency required for a build tree, -# even when the artifact cache was previously populated with -# artifacts we do not require, and the new build is run with dynamic tracking. -# -@pytest.mark.datafiles(DATA_DIR) -def test_never_delete_required_track(cli, datafiles): - project = str(datafiles) - element_path = 'elements' - - cli.configure({ - 'cache': { - 'quota': 10000000 - }, - 'scheduler': { - 'fetchers': 1, - 'builders': 1 - } - }) - - # Create a linear build tree - repo_dep1 = create_element_size('dep1.bst', project, element_path, [], 2000000) - repo_dep2 = create_element_size('dep2.bst', project, element_path, ['dep1.bst'], 2000000) - repo_dep3 = create_element_size('dep3.bst', project, element_path, ['dep2.bst'], 2000000) - repo_target = create_element_size('target.bst', project, element_path, ['dep3.bst'], 2000000) - - # This should all fit into the artifact cache - res = cli.run(project=project, args=['build', 'target.bst']) - res.assert_success() - - # They should all be cached - states = cli.get_element_states(project, ['target.bst']) - assert states['dep1.bst'] == 'cached' - assert states['dep2.bst'] == 'cached' - assert states['dep3.bst'] == 'cached' - assert states['target.bst'] == 'cached' - - # Now increase the size of all the elements - # - update_element_size('dep1.bst', project, repo_dep1, 8000000) - update_element_size('dep2.bst', project, repo_dep2, 8000000) - update_element_size('dep3.bst', project, repo_dep3, 8000000) - update_element_size('target.bst', project, repo_target, 8000000) - - # Now repeat the same test we did in test_never_delete_required(), - # except this time let's add dynamic tracking - # - res = cli.run(project=project, args=['build', '--track-all', 'target.bst']) - res.assert_main_error(ErrorDomain.STREAM, None) - res.assert_task_error(ErrorDomain.CAS, 'cache-too-full') - - # Expect the almost the same result that we did in test_never_delete_required() - # As the source will be downloaded first, we will be over the limit once - # the source for dep2.bst is downloaded - # - states = cli.get_element_states(project, ['target.bst']) - assert states['dep1.bst'] == 'cached' - assert states['dep2.bst'] == 'buildable' - assert states['dep3.bst'] != 'cached' - assert states['target.bst'] != 'cached' - - # Ensure that only valid cache quotas make it through the loading # process. # |