summaryrefslogtreecommitdiff
path: root/buildstream
Commit message (Collapse)AuthorAgeFilesLines
* Misc debugging printsjmac/cas_to_casJim MacArthur2018-09-251-1/+3
|
* _casbaseddirectory.py: Descend through symlinks to directories, tooJim MacArthur2018-09-241-5/+13
|
* _casbaseddirectory: Add selectable roundtrip importJim MacArthur2018-09-191-0/+10
|
* _casbaseddirectory: List directory names in file listsJim MacArthur2018-09-191-0/+1
|
* Fixes to print statements - only use generator onceJim MacArthur2018-09-191-3/+3
|
* _casbaseddirectory: Lots of debugging print statementsJim MacArthur2018-09-191-0/+8
|
* storage/__init__.py: Add CasBasedDirectoryJim MacArthur2018-09-191-0/+1
|
* _casbaseddirectory.py: Re-add the cas-to-cas import codeJim MacArthur2018-09-191-14/+193
|
* sandbox.py: Re-add the __allow_real_directory check, to stop us using ↵Jim MacArthur2018-09-191-1/+1
| | | | CASCache on unsupported elements
* Add some more asserts to check early that we're not doing the wrong thing ↵Jim MacArthur2018-09-193-2/+7
| | | | with CAS
* _artifactcache/artifactcache.py: Changes to mark_required_elements()Tristan Van Berkom2018-09-143-32/+53
| | | | | | | | | This was previously append_required_artifacts(), which presumed that we knew at startup time what the cache keys of all elements in the loaded pipeline would be. This fixes unexpected deletions of required artifacts when dynamic tracking is enabled with `bst build --track-all target.bst`
* Handle sockets when copying trees into artifactsRichard Maw2018-09-142-0/+14
| | | | | We can't include a socket in a CAS tree, but it's mostly meaningless to do so since there can't possibly be a process serving it.
* element: Remove __cached fieldRichard Maw2018-09-131-4/+1
| | | | | This can get out of sync with other two cache states, and we can do without it.
* _artifactcache/casserver.py: Implement BatchReadBlobsJürg Billeter2018-09-101-0/+25
| | | | Fixes #632.
* _artifactcache/casserver.py: Implement Capabilities serviceJürg Billeter2018-09-101-0/+24
|
* element: validate configuration variablesTiago Gomes2018-09-102-1/+12
| | | | Ensure that protected variables are not being redefined by the user.
* docs: document protected variablesTiago Gomes2018-09-101-15/+0
| | | | | And remove then from the defaults as they are dynamically set by BuildStream.
* subprocesses: Ensure PWD is set in process environmentRichard Maw2018-09-102-3/+15
| | | | | | | | | | | | | | | Naive getcwd implementations (such as in bash 4.4) can break when bind-mounts to different paths on the same filesystem are present, since the algorithm needs to know whether it's a mount-point to know whether it can trust the inode value from the readdir result or to use stat on the directory. Less naive implementations (such as in glibc) iterate again using stat in the case of not finding the directory because the inode in readdir was wrong, though a Linux-specific implementation could use name_to_handle_at. Letting the command know what directory it is in makes it unnecessary for it to call the faulty getcwd in the first place.
* _artifactcache: There shalt be only one cache sizeTristan Van Berkom2018-09-108-86/+84
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This does a lot of house cleaning, finally bringing cache cleanup logic to a level of comprehensibility. Changes in this commit include: o _artifactcache/artifactcache.py: _cache_size, _cache_quota and _cache_lower_threshold are now all private variables. get_approximate_cache_size() is now get_cache_size() Added get_quota_exceeded() for the purpose of safely checking if we have exceeded the quota. set_cache_size() now asserts that the passed size is not None, it is not acceptable to set a None size cache anymore. o _artifactcache/cascache.py: No longer set the ArtifactCache 'cache_size' variable violently in the commit() method. Also the calculate_cache_size() method now unconditionally calculates the cache size, that is what it's for. o _scheduler/jobs/cachesizejob.py & _scheduler/jobs/cleanupjob.py: Now check the success status. Don't try to set the cache size in the case that the job was terminated. o _scheduler/jobs/elementjob.py & _scheduler/queues/queue.py: No longer passing around the cache size from child tasks, this happens only explicitly, not implicitly for all tasks. o _scheduler/queues/buildqueue.py & _scheduler/scheduler.py: Use get_quota_exceeded() accessor This is a part of #623
* element.py: Remove _get_artifact_cache() accessor.Tristan Van Berkom2018-09-104-16/+16
| | | | | | | | | | | | The artifact cache is anyway a singleton, the Element itself has an internal handle on the artifact cache because it is needed very frequently; but an artifact cache is not element specific and should not be looked up by surrounding code on a per element basis. Updated _scheduler/queues/queue.py, _scheduler/queues/buildqueue.py and _scheduler/jobs/elementjob.py to get the artifact cache directly from the Platform
* _scheduler/jobs/cleanupjob.py: Use Platform.get_platform() public accessorTristan Van Berkom2018-09-101-3/+5
| | | | | | | | This was previously poking directly at the Platform._instance. Also, use the name 'artifacts' to hold the artifact cache to be consistent with other parts of the codebase, instead of calling it 'cache'.
* _scheduler/jobs/cachesizejob.py: Use Platform.get_platform() public accessorTristan Van Berkom2018-09-101-3/+5
| | | | | | | | This was previously poking directly at the Platform._instance. Also, use the name 'artifacts' to hold the artifact cache to be consistent with other parts of the codebase, instead of calling it 'cache'.
* element.py: Remove _get_artifact_size()Tristan Van Berkom2018-09-103-36/+22
| | | | | | | | There is no justification to hold onto this state here. Instead, just make `Element._assemble()` return the size of the artifact it cached, and localize handling of that return value in the BuildQueue implementation where the value is observed.
* _artifactcache/artifactcache.py: Sealing away some the estimated sizeTristan Van Berkom2018-09-104-28/+41
| | | | | | | | | | | | | | | | | | | | | | | | | Previously, the API contract was to expose the estimated_size variable on the ArtifactCache instance for all to see, however it is only relevant to the ArtifactCache abstract class code. Subclasses were informed to update the estimated_size variable in their calculate_cache_size() implementation. To untangle this and hide away the estimated size, this commit does the following: o Introduces ArtifactCache.compute_cache_size() API for external callers o ArtifactCache.compute_cache_size() calls the abstract method for the CasCache subclass to implement o ArtifactCache.compute_cache_size() updates the private estimated_size variable o All direct callers to ArtifactCache.calculate_cache_size(), have been updated to use the ArtifactCache.compute_cache_size() method instead, which takes care of updating anything local to the ArtifactCache abstract class code (the estimated_size)
* _artifactcache/artifactcache.py: Removing unused variable.Tristan Van Berkom2018-09-101-1/+0
| | | | | | | | | The ArtifactCache._local variable used to exist in order to use a special hack to allow absolute paths to a remote artifact cache, this was all for the purpose of testing. This has all gone away with the introduction of CAS, leaving behind a stale variable.
* _scheduler: API fixes; _check_cache_size_real() -> check_cache_size()Tristan Van Berkom2018-09-103-10/+34
| | | | | | | | | | | | | | | | | | | | | | | | Here we have a very private looking _check_cache_size_real() function which no-one would ever want to call from outside of the _scheduler, especially given it's `_real()` prefix we should look for another outward facing API to use. However this is not private to the scheduler, and is intended to be called by the `Queue` implementations. o Renamed this to check_cache_size() o Moved it to the public API section of the Scheduler object o Added the missing API documenting comment o Also added the missing API documenting comment to the private `_run_cleanup()` callback which runs in response to completion of the cache size calculation job. o Also place the cleanup job logs into a cleanup subdirectory, for better symmetry with the cache_size jobs which now have their own subdirectory
* _artifactcache: Making public methods publicTristan Van Berkom2018-09-104-32/+35
| | | | | | | | | | | The artifact cache provides the following public methods for external callers, but was hiding them away as if they are private. o ArtifactCache.add_artifact_size() o ArtifactCache.set_cache_size() Mark these properly public
* element.py: Adding missing API documenting commentTristan Van Berkom2018-09-101-0/+7
|
* _scheduler/scheduler.py: Make CacheSizeJob() not require the CACHE exclusivelyTristan Van Berkom2018-09-101-1/+0
| | | | | | | | | | | This runs after every pull, and does not need the cache exclusively, only the cleanup job requires the cache exclusively. Without this, every time a cache_size job is queued, all pull and build jobs need to complete before cache_size job can run exclusively, which is not good. This is a part of #623
* _scheduler/resources.py: Set CACHE max resources to 0Tristan Van Berkom2018-09-101-1/+1
| | | | | | | | | | | This means there is no cap for shared resource requests. Together with the previous commit, this causes the cleanup job and the pull/build jobs to all require unlimited shared access to the CACHE resource; while only the cleanup job requires exclusive access to the resource. This is a part of #623
* _scheduler/queues: Mark build and pull queue as requiring shared access to ↵Tristan Van Berkom2018-09-102-2/+2
| | | | | | | | | | | the CACHE This is what the whole resource.py thing was created for, the cleanup job must have exclusive access to the cache, while the pull and build jobs which result in adding new artifacts, must only require shared access. This is a part of #623
* Upstream libtool .la file removalThomas Coldrick2018-09-081-0/+36
| | | | | | In freedesktop-sdk we add a script to our project.conf to remove libtool .la files from autotools projects after install, this seems like a sensible default, so we're attempting to send it upstream.
* qmake.py: Mark this as a BST_VIRTUAL_DIRECTORY‧pluginMartin Blanchard2018-09-071-1/+2
| | | | https://gitlab.com/BuildStream/buildstream/issues/454
* meson.py: Mark this as a BST_VIRTUAL_DIRECTORY‧pluginMartin Blanchard2018-09-071-1/+2
| | | | https://gitlab.com/BuildStream/buildstream/issues/454
* make.py: Mark this as a BST_VIRTUAL_DIRECTORY‧pluginMartin Blanchard2018-09-071-1/+2
| | | | https://gitlab.com/BuildStream/buildstream/issues/454
* cmake.py: Mark this as a BST_VIRTUAL_DIRECTORY pluginMartin Blanchard2018-09-071-1/+2
| | | | https://gitlab.com/BuildStream/buildstream/issues/454
* autotools.py: Mark this as a BST_VIRTUAL_DIRECTORY pluginJim MacArthur2018-09-071-1/+2
| | | | https://gitlab.com/BuildStream/buildstream/issues/454
* element.py: Switch to SandboxRemote if config option is setJim MacArthur2018-09-071-1/+27
| | | | https://gitlab.com/BuildStream/buildstream/issues/454
* element.py: Get the updated virtual directory after runningJim MacArthur2018-09-071-0/+2
| | | | | | | | | Executing run() on a sandbox can now replace the virtual directory, since remote execution returns a potentially different directory rather than an update to the existing one. Call get_virtual_directory() again after running to accout for this. https://gitlab.com/BuildStream/buildstream/issues/454
* sandbox/__init__.py: Add SandboxRemoteJim MacArthur2018-09-071-0/+1
| | | | https://gitlab.com/BuildStream/buildstream/issues/454
* _sandboxremote.py: Implement the REAPI clientJim MacArthur2018-09-071-0/+226
| | | | | | | | The remote execution client is implemented as a remote sandbox that sends sources and build commands to a REAPI server and fetches results once remotely executed. New file. https://gitlab.com/BuildStream/buildstream/issues/454
* Add "remote-execution" project configuration optionJim MacArthur2018-09-073-1/+16
| | | | | | This just adds one option, "remote-execution/url". Affects multiple files. https://gitlab.com/BuildStream/buildstream/issues/454
* sandbox.py: Allow setting the virtual directoryJim MacArthur2018-09-072-3/+23
| | | | | | | | This is for use after remote execution has finished, since remote execution produces a new output directory rather than modifying the initial directory. https://gitlab.com/BuildStream/buildstream/issues/454
* _casbaseddirectory.py: Add a method for hash recalculationMartin Blanchard2018-09-071-0/+9
| | | | https://gitlab.com/BuildStream/buildstream/issues/454
* cascache.py: Introduce new pull helpersMartin Blanchard2018-09-071-0/+51
| | | | | | Add a pull_tree() helper. https://gitlab.com/BuildStream/buildstream/issues/454
* cascache.py: Introduce new push helpersMartin Blanchard2018-09-071-0/+59
| | | | | | | Add push_directory() and push_message() helpers along with a verify_digest_pushed(). https://gitlab.com/BuildStream/buildstream/issues/454
* cascache.py: Preparation for remote executionJim MacArthur2018-09-071-90/+109
| | | | | | | Refactor the push() and pull() implementations so that API additions needed for remote-execution is made easier. https://gitlab.com/BuildStream/buildstream/issues/454
* jobs.py: Reduce FD leaks from queues and process objectsdanielsilverstone-ct/maybe-reduce-fd-leaksDaniel Silverstone2018-09-041-1/+6
| | | | | | | | | | | | The garbage collector can take too long to get around to cleaning up the Queue and Process instances in completed Job instances. As such, FDs tend to leak and in very large projects this can result in running out of FDs before a build, fetch, track, or other process can complete. This patch reduces the chance of that by only creating the queue when it's needed, and forcing the queue and process instances to be deleted when the parent is finished with them. Signed-off-by: Daniel Silverstone <daniel.silverstone@codethink.co.uk>
* plugins/git.py: Warn if ref is not in given trackTom Pollard2018-09-031-4/+41
| | | | | | | | | | | Add a helper function assert_ref_in_track to git.py GitMirror() which is used when staging & initing the source workspace. It checks the element's ref exists in the track (branch/tag) if it has been specified, raising a warning if necessary. The warning makes use of the warning token 'REF_NOT_IN_TRACK' from the configurable CoreWarnings. If the element has been tracked with bst, it is assumed that the value of ref exists in the track as it was generated from it & as such is not asserted.
* source.py: Track marked URLs and assert they are marked during ↵tristan/source-fetcher-changesTristan Van Berkom2018-09-021-0/+20
| | | | | | | | | | | | | | Plugin.configure() This cannot test for unaliased URLs, as those can be discovered later on outside of user provided element configuration; at least we assert that if an alias was used, we have seen it at load time. This will cause a BUG to occur for a plugin which marks an aliased URL (or attempts to translate one) outside of `Plugin.configure()`, if that URL was not previously seen. This is a part of #620