blob: ef7097e35acbe2d7017078e64c1bd734a90bc514 [file] [log] [blame]
#!/usr/bin/env python
# Copyright 2017 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""Runs a Fuchsia gtest-based test on Swarming, optionally many times,
collecting the output of the runs into a directory. Useful for flake checking,
and faster than using trybots by avoiding repeated bot_update, compile, archive,
etc. and allowing greater parallelism.
To use, run in a new shell (it blocks until all Swarming jobs complete):
tools/ -t content_unittests --out-dir=out/fuch
The logs of the runs will be stored in results/ (or specify a results directory
with --results=some_dir). You can then do something like `grep -L SUCCESS
results/*` to find the tests that failed or otherwise process the log files.
import argparse
import multiprocessing
import os
import shutil
import subprocess
import sys
def _Spawn(args):
"""Triggers a swarming job. The arguments passed are:
- The index of the job;
- The command line arguments object;
- The hash of the isolate job used to trigger;
- Value of --gtest_filter arg, or empty if none.
The return value is passed to a collect-style map() and consists of:
- The index of the job;
- The json file created by triggering and used to collect results;
- The command line arguments object.
index, args, isolated_hash, gtest_filter = args
json_file = os.path.join(args.results, '%d.json' % index)
trigger_args = [
'tools/swarming_client/', 'trigger',
'-S', '',
'-I', '',
'-d', 'pool', 'Chrome',
'-s', isolated_hash,
'--dump-json', json_file,
if args.target_os == 'fuchsia':
trigger_args += [
'-d', 'os', 'Linux',
'-d', 'kvm', '1',
'-d', 'gpu', 'none',
'-d', 'cpu', args.arch,
elif args.target_os == 'win':
trigger_args += [ '-d', 'os', 'Windows' ]
trigger_args += [
if gtest_filter:
trigger_args.append('--gtest_filter=' + gtest_filter)
elif args.target_os == 'fuchsia':
filter_file = \
'testing/buildbot/filters/fuchsia.' + args.test_name + '.filter'
if os.path.isfile(filter_file):
trigger_args.append('--test-launcher-filter-file=../../' + filter_file)
with open(os.devnull, 'w') as nul:
subprocess.check_call(trigger_args, stdout=nul)
return (index, json_file, args)
def _Collect(spawn_result):
index, json_file, args = spawn_result
p = subprocess.Popen([
'tools/swarming_client/', 'collect',
'-S', '',
'--json', json_file,
stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
stdout = p.communicate()[0]
if p.returncode != 0 and len(stdout) < 2**10 and 'Internal error!' in stdout:
file_suffix = '.INTERNAL_ERROR'
exit_code = p.returncode
file_suffix = '' if exit_code == 0 else '.FAILED'
filename = '%d%s.stdout.txt' % (index, file_suffix)
with open(os.path.join(args.results, filename), 'w') as f:
return exit_code
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-C', '--out-dir', default='out/fuch',
help='Build directory.')
parser.add_argument('--target-os', default='detect', help='gn target_os')
parser.add_argument('--test-name', '-t', required=True,
help='Name of test to run.')
parser.add_argument('--arch', '-a', default='detect',
help='CPU architecture of the test binary.')
parser.add_argument('--copies', '-n', type=int, default=1,
help='Number of copies to spawn.')
parser.add_argument('--results', '-r', default='results',
help='Directory in which to store results.')
help='Use the given gtest_filter, rather than the '
'default filter file, if any.')
args = parser.parse_args()
if args.target_os == 'detect':
with open(os.path.join(args.out_dir, '')) as f:
gn_args = {}
for l in f:
l = l.split('#')[0].strip()
if not l: continue
k, v = map(str.strip, l.split('=', 1))
gn_args[k] = v
if 'target_os' in gn_args:
args.target_os = gn_args['target_os'].strip('"')
args.target_os = { 'darwin': 'mac', 'linux2': 'linux', 'win32': 'win' }[
# Determine the CPU architecture of the test binary, if not specified.
if args.arch == 'detect' and args.target_os == 'fuchsia':
executable_info = subprocess.check_output(
['file', os.path.join(args.out_dir, args.test_name)])
if 'ARM aarch64' in executable_info:
args.arch = 'arm64',
args.arch = 'x86-64'
['tools/mb/', 'isolate', '//' + args.out_dir, args.test_name])
print 'If you get authentication errors, follow:'
print ''
print 'Uploading to isolate server, this can take a while...'
archive_output = subprocess.check_output(
['tools/swarming_client/', 'archive',
'-I', '',
'-i', os.path.join(args.out_dir, args.test_name + '.isolate'),
'-s', os.path.join(args.out_dir, args.test_name + '.isolated')])
isolated_hash = archive_output.split()[0]
if os.path.isdir(args.results):
print 'Triggering %d tasks...' % args.copies
pool = multiprocessing.Pool()
spawn_args = map(lambda i: (i, args, isolated_hash, args.gtest_filter),
spawn_results = pool.imap_unordered(_Spawn, spawn_args)
exit_codes = []
collect_results = pool.imap_unordered(_Collect, spawn_results)
for result in collect_results:
successes = sum(1 for x in exit_codes if x == 0)
errors = sum(1 for x in exit_codes if x == INTERNAL_ERROR_EXIT_CODE)
failures = len(exit_codes) - successes - errors
clear_to_eol = '\033[K'
print('\r[%d/%d] collected: '
'%d successes, %d failures, %d bot errors...%s' % (len(exit_codes),
args.copies, successes, failures, errors, clear_to_eol)),
print 'Results logs collected into', os.path.abspath(args.results) + '.'
return 0
if __name__ == '__main__':