#!/usr/bin/env python # Copyright 2017 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Runs a gtest-based test on Swarming, optionally many times, collecting the output of the runs into a directory. Useful for flake checking, and faster than using trybots by avoiding repeated bot_update, compile, archive, etc. and allowing greater parallelism. To use, run in a new shell (it blocks until all Swarming jobs complete): tools/run-swarmed.py out/rel base_unittests The logs of the runs will be stored in results/ (or specify a results directory with --results=some_dir). You can then do something like `grep -L SUCCESS results/*` to find the tests that failed or otherwise process the log files. See //docs/workflow/debugging-with-swarming.md for more details. """ from __future__ import print_function import argparse import hashlib import json import multiprocessing.dummy import os import shutil import subprocess import sys INTERNAL_ERROR_EXIT_CODE = -1000 DEFAULT_ANDROID_DEVICE_TYPE = "walleye" def _Spawn(args): """Triggers a swarming job. The arguments passed are: - The index of the job; - The command line arguments object; - The hash of the isolate job used to trigger. The return value is passed to a collect-style map() and consists of: - The index of the job; - The json file created by triggering and used to collect results; - The command line arguments object. """ index, args, isolated_hash = args json_file = os.path.join(args.results, '%d.json' % index) trigger_args = [ 'tools/luci-go/swarming', 'trigger', '-S', 'https://chromium-swarm.appspot.com', '-I', 'https://isolateserver.appspot.com', '-d', 'pool=' + args.pool, '-s', isolated_hash, '-dump-json', json_file, '-d', 'os=' + args.swarming_os, '-tag=purpose:user-debug-run-swarmed', ] if args.target_os == 'fuchsia': trigger_args += [ '-d', 'kvm=1', '-d', 'gpu=none', ] if args.arch != 'detect': trigger_args += [ '-d', 'cpu=' + args.arch, ] # The aliases for device type are stored here: # luci/appengine/swarming/ui2/modules/alias.js # for example 'blueline' = 'Pixel 3' if args.target_os == 'android': if args.device_type is None and args.device_os is None: trigger_args += ['-d', 'device_type=' + DEFAULT_ANDROID_DEVICE_TYPE] if args.device_type: trigger_args += ['-d', 'device_type=' + args.device_type] if args.device_os: trigger_args += ['-d', 'device_os=' + args.device_os] runner_args = [] if not args.no_test_flags: # These flags are recognized by our test runners, but do not work # when running custom scripts. runner_args += [ '--test-launcher-summary-output=${ISOLATED_OUTDIR}/output.json', '--system-log-file=${ISOLATED_OUTDIR}/system_log' ] if args.gtest_filter: runner_args.append('--gtest_filter=' + args.gtest_filter) if args.repeat: runner_args.append('--repeat=' + args.repeat) elif args.target_os == 'fuchsia': filter_file = \ 'testing/buildbot/filters/fuchsia.' + args.target_name + '.filter' if os.path.isfile(filter_file): runner_args.append('--test-launcher-filter-file=../../' + filter_file) if runner_args: trigger_args.append('--') trigger_args.extend(runner_args) with open(os.devnull, 'w') as nul: subprocess.check_call(trigger_args, stdout=nul) return (index, json_file, args) def _Collect(spawn_result): index, json_file, args = spawn_result with open(json_file) as f: task_json = json.load(f) task_ids = [task['task_id'] for task in task_json['tasks']] for t in task_ids: print('Task {}: https://chromium-swarm.appspot.com/task?id={}'.format( index, t)) p = subprocess.Popen( [ 'tools/luci-go/swarming', 'collect', '-S', 'https://chromium-swarm.appspot.com', '--task-output-stdout=console', ] + task_ids, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) stdout = p.communicate()[0] if p.returncode != 0 and len(stdout) < 2**10 and 'Internal error!' in stdout: exit_code = INTERNAL_ERROR_EXIT_CODE file_suffix = '.INTERNAL_ERROR' else: exit_code = p.returncode file_suffix = '' if exit_code == 0 else '.FAILED' filename = '%d%s.stdout.txt' % (index, file_suffix) with open(os.path.join(args.results, filename), 'w') as f: f.write(stdout) return exit_code def main(): parser = argparse.ArgumentParser() parser.add_argument('--swarming-os', help='OS specifier for Swarming.') parser.add_argument('--target-os', default='detect', help='gn target_os') parser.add_argument('--arch', '-a', default='detect', help='CPU architecture of the test binary.') parser.add_argument('--build', dest='build', action='store_true', help='Build before isolating (default).') parser.add_argument( '--no-build', dest='build', action='store_false', help='Do not build, just isolate.') parser.add_argument('--isolate-map-file', '-i', help='path to isolate map file if not using default') parser.add_argument('--copies', '-n', type=int, default=1, help='Number of copies to spawn.') parser.add_argument( '--device-os', help='Run tests on the given version of Android.') parser.add_argument( '--device-type', help='device_type specifier for Swarming' ' from https://chromium-swarm.appspot.com/botlist .') parser.add_argument('--pool', default='chromium.tests', help='Use the given swarming pool.') parser.add_argument('--results', '-r', default='results', help='Directory in which to store results.') parser.add_argument('--gtest_filter', help='Use the given gtest_filter, rather than the ' 'default filter file, if any.') parser.add_argument( '--repeat', help='Number of times to repeat the specified set of tests.') parser.add_argument('--no-test-flags', action='store_true', help='Do not add --test-launcher-summary-output and ' '--system-log-file flags to the comment.') parser.add_argument('out_dir', type=str, help='Build directory.') parser.add_argument('target_name', type=str, help='Name of target to run.') args = parser.parse_args() if args.target_os == 'detect': with open(os.path.join(args.out_dir, 'args.gn')) as f: gn_args = {} for l in f: l = l.split('#')[0].strip() if not l: continue k, v = map(str.strip, l.split('=', 1)) gn_args[k] = v if 'target_os' in gn_args: args.target_os = gn_args['target_os'].strip('"') else: args.target_os = { 'darwin': 'mac', 'linux2': 'linux', 'win32': 'win' }[ sys.platform] if args.swarming_os is None: args.swarming_os = { 'mac': 'Mac', 'win': 'Windows', 'linux': 'Linux', 'android': 'Android', 'fuchsia': 'Linux' }[args.target_os] if args.target_os == 'win' and args.target_name.endswith('.exe'): # The machinery expects not to have a '.exe' suffix. args.target_name = os.path.splitext(args.target_name)[0] # Determine the CPU architecture of the test binary, if not specified. if args.arch == 'detect' and args.target_os not in ('android', 'mac', 'win'): executable_info = subprocess.check_output( ['file', os.path.join(args.out_dir, args.target_name)]) if 'ARM aarch64' in executable_info: args.arch = 'arm64', else: args.arch = 'x86-64' mb_cmd = [sys.executable, 'tools/mb/mb.py', 'isolate'] if not args.build: mb_cmd.append('--no-build') if args.isolate_map_file: mb_cmd += ['--isolate-map-file', args.isolate_map_file] mb_cmd += ['//' + args.out_dir, args.target_name] subprocess.check_call(mb_cmd) print('If you get authentication errors, follow:') print( ' https://www.chromium.org/developers/testing/isolated-testing/for-swes#TOC-Login-on-the-services' ) print('Uploading to isolate server, this can take a while...') isolated = os.path.join(args.out_dir, args.target_name + '.isolated') subprocess.check_output([ 'tools/luci-go/isolate', 'archive', '-I', 'https://isolateserver.appspot.com', '-i', os.path.join(args.out_dir, args.target_name + '.isolate'), '-s', isolated ]) with open(isolated) as f: isolated_hash = hashlib.sha1(f.read()).hexdigest() if os.path.isdir(args.results): shutil.rmtree(args.results) os.makedirs(args.results) try: print('Triggering %d tasks...' % args.copies) # Use dummy since threadpools give better exception messages # than process pools do, and threads work fine for what we're doing. pool = multiprocessing.dummy.Pool() spawn_args = map(lambda i: (i, args, isolated_hash), range(args.copies)) spawn_results = pool.imap_unordered(_Spawn, spawn_args) exit_codes = [] collect_results = pool.imap_unordered(_Collect, spawn_results) for result in collect_results: exit_codes.append(result) successes = sum(1 for x in exit_codes if x == 0) errors = sum(1 for x in exit_codes if x == INTERNAL_ERROR_EXIT_CODE) failures = len(exit_codes) - successes - errors clear_to_eol = '\033[K' print( '\r[%d/%d] collected: ' '%d successes, %d failures, %d bot errors...%s' % (len(exit_codes), args.copies, successes, failures, errors, clear_to_eol), end=' ') sys.stdout.flush() print() print('Results logs collected into', os.path.abspath(args.results) + '.') finally: pool.close() pool.join() return 0 if __name__ == '__main__': sys.exit(main())