blob: 8101dcdf960c8adbf05e2c309024d7990700279e [file] [log] [blame]
#!/usr/bin/env python
# Copyright 2014 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
"""A utility script for downloading versioned Syzygy binaries."""
import hashlib
import errno
import json
import logging
import optparse
import os
import re
import shutil
import stat
import sys
import subprocess
import tempfile
import time
import zipfile
_LOGGER = logging.getLogger(os.path.basename(__file__))
# The relative path where official builds are archived in their GS bucket.
_SYZYGY_ARCHIVE_PATH = ('/builds/official/%(revision)s')
# A JSON file containing the state of the download directory. If this file and
# directory state do not agree, then the binaries will be downloaded and
# installed again.
_STATE = '.state'
# This matches an integer (an SVN revision number) or a SHA1 value (a GIT hash).
# The archive exclusively uses lowercase GIT hashes.
_REVISION_RE = re.compile('^(?:\d+|[a-f0-9]{40})$')
# This matches an MD5 hash.
_MD5_RE = re.compile('^[a-f0-9]{32}$')
# List of reources to be downloaded and installed. These are tuples with the
# following format:
# (basename, logging name, relative installation path, extraction filter)
_RESOURCES = [
('benchmark.zip', 'benchmark', '', None),
('binaries.zip', 'binaries', 'exe', None),
('symbols.zip', 'symbols', 'exe',
lambda x: x.filename.endswith('.dll.pdb'))]
# Name of the MS DIA dll that we need to copy to the binaries directory.
_DIA_DLL_NAME = "msdia140.dll"
def _LoadState(output_dir):
"""Loads the contents of the state file for a given |output_dir|, returning
None if it doesn't exist.
"""
path = os.path.join(output_dir, _STATE)
if not os.path.exists(path):
_LOGGER.debug('No state file found.')
return None
with open(path, 'rb') as f:
_LOGGER.debug('Reading state file: %s', path)
try:
return json.load(f)
except ValueError:
_LOGGER.debug('Invalid state file.')
return None
def _SaveState(output_dir, state, dry_run=False):
"""Saves the |state| dictionary to the given |output_dir| as a JSON file."""
path = os.path.join(output_dir, _STATE)
_LOGGER.debug('Writing state file: %s', path)
if dry_run:
return
with open(path, 'wb') as f:
f.write(json.dumps(state, sort_keys=True, indent=2))
def _Md5(path):
"""Returns the MD5 hash of the file at |path|, which must exist."""
return hashlib.md5(open(path, 'rb').read()).hexdigest()
def _StateIsValid(state):
"""Returns true if the given state structure is valid."""
if not isinstance(state, dict):
_LOGGER.debug('State must be a dict.')
return False
r = state.get('revision', None)
if not isinstance(r, basestring) or not _REVISION_RE.match(r):
_LOGGER.debug('State contains an invalid revision.')
return False
c = state.get('contents', None)
if not isinstance(c, dict):
_LOGGER.debug('State must contain a contents dict.')
return False
for (relpath, md5) in c.iteritems():
if not isinstance(relpath, basestring) or len(relpath) == 0:
_LOGGER.debug('State contents dict contains an invalid path.')
return False
if not isinstance(md5, basestring) or not _MD5_RE.match(md5):
_LOGGER.debug('State contents dict contains an invalid MD5 digest.')
return False
return True
def _BuildActualState(stored, revision, output_dir):
"""Builds the actual state using the provided |stored| state as a template.
Only examines files listed in the stored state, causing the script to ignore
files that have been added to the directories locally. |stored| must be a
valid state dictionary.
"""
contents = {}
state = { 'revision': revision, 'contents': contents }
for relpath, md5 in stored['contents'].iteritems():
abspath = os.path.abspath(os.path.join(output_dir, relpath))
if os.path.isfile(abspath):
m = _Md5(abspath)
contents[relpath] = m
return state
def _StatesAreConsistent(stored, actual):
"""Validates whether two state dictionaries are consistent. Both must be valid
state dictionaries. Additional entries in |actual| are ignored.
"""
if stored['revision'] != actual['revision']:
_LOGGER.debug('Mismatched revision number.')
return False
cont_stored = stored['contents']
cont_actual = actual['contents']
for relpath, md5 in cont_stored.iteritems():
if relpath not in cont_actual:
_LOGGER.debug('Missing content: %s', relpath)
return False
if md5 != cont_actual[relpath]:
_LOGGER.debug('Modified content: %s', relpath)
return False
return True
def _GetCurrentState(revision, output_dir):
"""Loads the current state and checks to see if it is consistent. Returns
a tuple (state, bool). The returned state will always be valid, even if an
invalid state is present on disk.
"""
stored = _LoadState(output_dir)
if not _StateIsValid(stored):
_LOGGER.debug('State is invalid.')
# Return a valid but empty state.
return ({'revision': '0', 'contents': {}}, False)
actual = _BuildActualState(stored, revision, output_dir)
# If the script has been modified consider the state invalid.
path = os.path.join(output_dir, _STATE)
if os.path.getmtime(__file__) > os.path.getmtime(path):
return (stored, False)
# Otherwise, explicitly validate the state.
if not _StatesAreConsistent(stored, actual):
return (stored, False)
return (stored, True)
def _DirIsEmpty(path):
"""Returns true if the given directory is empty, false otherwise."""
for root, dirs, files in os.walk(path):
return not dirs and not files
def _RmTreeHandleReadOnly(func, path, exc):
"""An error handling function for use with shutil.rmtree. This will
detect failures to remove read-only files, and will change their properties
prior to removing them. This is necessary on Windows as os.remove will return
an access error for read-only files, and git repos contain read-only
pack/index files.
"""
excvalue = exc[1]
if func in (os.rmdir, os.remove) and excvalue.errno == errno.EACCES:
_LOGGER.debug('Removing read-only path: %s', path)
os.chmod(path, stat.S_IRWXU | stat.S_IRWXG | stat.S_IRWXO)
func(path)
else:
raise
def _RmTree(path):
"""A wrapper of shutil.rmtree that handles read-only files."""
shutil.rmtree(path, ignore_errors=False, onerror=_RmTreeHandleReadOnly)
def _CleanState(output_dir, state, dry_run=False):
"""Cleans up files/directories in |output_dir| that are referenced by
the given |state|. Raises an error if there are local changes. Returns a
dictionary of files that were deleted.
"""
_LOGGER.debug('Deleting files from previous installation.')
deleted = {}
# Generate a list of files to delete, relative to |output_dir|.
contents = state['contents']
files = sorted(contents.keys())
# Try to delete the files. Keep track of directories to delete as well.
dirs = {}
for relpath in files:
fullpath = os.path.join(output_dir, relpath)
fulldir = os.path.dirname(fullpath)
dirs[fulldir] = True
if os.path.exists(fullpath):
# If somehow the file has become a directory complain about it.
if os.path.isdir(fullpath):
raise Exception('Directory exists where file expected: %s' % fullpath)
# Double check that the file doesn't have local changes. If it does
# then refuse to delete it.
if relpath in contents:
stored_md5 = contents[relpath]
actual_md5 = _Md5(fullpath)
if actual_md5 != stored_md5:
raise Exception('File has local changes: %s' % fullpath)
# The file is unchanged so it can safely be deleted.
_LOGGER.debug('Deleting file "%s".', fullpath)
deleted[relpath] = True
if not dry_run:
os.unlink(fullpath)
# Sort directories from longest name to shortest. This lets us remove empty
# directories from the most nested paths first.
dirs = sorted(dirs.keys(), key=lambda x: len(x), reverse=True)
for p in dirs:
if os.path.exists(p) and _DirIsEmpty(p):
_LOGGER.debug('Deleting empty directory "%s".', p)
if not dry_run:
_RmTree(p)
return deleted
def _FindGsUtil():
"""Looks for depot_tools and returns the absolute path to gsutil.py."""
for path in os.environ['PATH'].split(os.pathsep):
path = os.path.abspath(path)
git_cl = os.path.join(path, 'git_cl.py')
gs_util = os.path.join(path, 'gsutil.py')
if os.path.exists(git_cl) and os.path.exists(gs_util):
return gs_util
return None
def _GsUtil(*cmd):
"""Runs the given command in gsutil with exponential backoff and retries."""
gs_util = _FindGsUtil()
cmd = [sys.executable, gs_util] + list(cmd)
retries = 3
timeout = 4 # Seconds.
while True:
_LOGGER.debug('Running %s', cmd)
prog = subprocess.Popen(cmd, shell=False)
prog.communicate()
# Stop retrying on success.
if prog.returncode == 0:
return
# Raise a permanent failure if retries have been exhausted.
if retries == 0:
raise RuntimeError('Command "%s" returned %d.' % (cmd, prog.returncode))
_LOGGER.debug('Sleeping %d seconds and trying again.', timeout)
time.sleep(timeout)
retries -= 1
timeout *= 2
def _Download(resource):
"""Downloads the given GS resource to a temporary file, returning its path."""
tmp = tempfile.mkstemp(suffix='syzygy_archive')
os.close(tmp[0])
url = 'gs://syzygy-archive' + resource
_GsUtil('cp', url, tmp[1])
return tmp[1]
def _MaybeCopyDIABinaries(options, contents):
"""Try to copy the DIA DLL to the binaries exe directory."""
toolchain_data_file = os.path.join(os.path.dirname(__file__),
'win_toolchain.json')
if not os.path.exists(toolchain_data_file):
_LOGGER.debug('Toolchain JSON data file doesn\'t exist, skipping.')
return
with open(toolchain_data_file) as temp_f:
toolchain_data = json.load(temp_f)
if not os.path.isdir(toolchain_data['path']):
_LOGGER.error('The toolchain JSON file is invalid.')
return
dia_sdk_binaries_dir = os.path.join(toolchain_data['path'], 'DIA SDK', 'bin')
dia_dll = os.path.join(dia_sdk_binaries_dir, _DIA_DLL_NAME)
if not os.path.exists(dia_dll):
_LOGGER.debug('%s is missing, skipping.')
return
dia_dll_dest = os.path.join(options.output_dir, 'exe', _DIA_DLL_NAME)
_LOGGER.debug('Copying %s to %s.' % (dia_dll, dia_dll_dest))
if not options.dry_run:
shutil.copy(dia_dll, dia_dll_dest)
contents[os.path.relpath(dia_dll_dest, options.output_dir)] = (
_Md5(dia_dll_dest))
def _InstallBinaries(options, deleted={}):
"""Installs Syzygy binaries. This assumes that the output directory has
already been cleaned, as it will refuse to overwrite existing files."""
contents = {}
state = { 'revision': options.revision, 'contents': contents }
archive_path = _SYZYGY_ARCHIVE_PATH % { 'revision': options.revision }
if options.resources:
resources = [(resource, resource, '', None)
for resource in options.resources]
else:
resources = _RESOURCES
for (base, name, subdir, filt) in resources:
# Create the output directory if it doesn't exist.
fulldir = os.path.join(options.output_dir, subdir)
if os.path.isfile(fulldir):
raise Exception('File exists where a directory needs to be created: %s' %
fulldir)
if not os.path.exists(fulldir):
_LOGGER.debug('Creating directory: %s', fulldir)
if not options.dry_run:
os.makedirs(fulldir)
# Download and read the archive.
resource = archive_path + '/' + base
_LOGGER.debug('Retrieving %s archive at "%s".', name, resource)
path = _Download(resource)
_LOGGER.debug('Unzipping %s archive.', name)
with open(path, 'rb') as data:
archive = zipfile.ZipFile(data)
for entry in archive.infolist():
if not filt or filt(entry):
fullpath = os.path.normpath(os.path.join(fulldir, entry.filename))
relpath = os.path.relpath(fullpath, options.output_dir)
if os.path.exists(fullpath):
# If in a dry-run take into account the fact that the file *would*
# have been deleted.
if options.dry_run and relpath in deleted:
pass
else:
raise Exception('Path already exists: %s' % fullpath)
# Extract the file and update the state dictionary.
_LOGGER.debug('Extracting "%s".', fullpath)
if not options.dry_run:
archive.extract(entry.filename, fulldir)
md5 = _Md5(fullpath)
contents[relpath] = md5
if sys.platform == 'cygwin':
os.chmod(fullpath, os.stat(fullpath).st_mode | stat.S_IXUSR)
_LOGGER.debug('Removing temporary file "%s".', path)
os.remove(path)
if options.copy_dia_binaries:
# Try to copy the DIA binaries to the binaries directory.
_MaybeCopyDIABinaries(options, contents)
return state
def _ParseCommandLine():
"""Parses the command-line and returns an options structure."""
option_parser = optparse.OptionParser()
option_parser.add_option('--dry-run', action='store_true', default=False,
help='If true then will simply list actions that would be performed.')
option_parser.add_option('--force', action='store_true', default=False,
help='Force an installation even if the binaries are up to date.')
option_parser.add_option('--no-cleanup', action='store_true', default=False,
help='Allow installation on non-Windows platforms, and skip the forced '
'cleanup step.')
option_parser.add_option('--output-dir', type='string',
help='The path where the binaries will be replaced. Existing binaries '
'will only be overwritten if not up to date.')
option_parser.add_option('--overwrite', action='store_true', default=False,
help='If specified then the installation will happily delete and rewrite '
'the entire output directory, blasting any local changes.')
option_parser.add_option('--revision', type='string',
help='The SVN revision or GIT hash associated with the required version.')
option_parser.add_option('--revision-file', type='string',
help='A text file containing an SVN revision or GIT hash.')
option_parser.add_option('--resource', type='string', action='append',
dest='resources', help='A resource to be downloaded.')
option_parser.add_option('--verbose', dest='log_level', action='store_const',
default=logging.INFO, const=logging.DEBUG,
help='Enables verbose logging.')
option_parser.add_option('--quiet', dest='log_level', action='store_const',
default=logging.INFO, const=logging.ERROR,
help='Disables all output except for errors.')
option_parser.add_option('--copy-dia-binaries', action='store_true',
default=False, help='If true then the DIA dll will get copied into the '
'binaries directory if it\'s available.')
options, args = option_parser.parse_args()
if args:
option_parser.error('Unexpected arguments: %s' % args)
if not options.output_dir:
option_parser.error('Must specify --output-dir.')
if not options.revision and not options.revision_file:
option_parser.error('Must specify one of --revision or --revision-file.')
if options.revision and options.revision_file:
option_parser.error('Must not specify both --revision and --revision-file.')
# Configure logging.
logging.basicConfig(level=options.log_level)
# If a revision file has been specified then read it.
if options.revision_file:
options.revision = open(options.revision_file, 'rb').read().strip()
_LOGGER.debug('Parsed revision "%s" from file "%s".',
options.revision, options.revision_file)
# Ensure that the specified SVN revision or GIT hash is valid.
if not _REVISION_RE.match(options.revision):
option_parser.error('Must specify a valid SVN or GIT revision.')
# This just makes output prettier to read.
options.output_dir = os.path.normpath(options.output_dir)
return options
def _RemoveOrphanedFiles(options):
"""This is run on non-Windows systems to remove orphaned files that may have
been downloaded by a previous version of this script.
"""
# Reconfigure logging to output info messages. This will allow inspection of
# cleanup status on non-Windows buildbots.
_LOGGER.setLevel(logging.INFO)
output_dir = os.path.abspath(options.output_dir)
# We only want to clean up the folder in 'src/third_party/syzygy', and we
# expect to be called with that as an output directory. This is an attempt to
# not start deleting random things if the script is run from an alternate
# location, or not called from the gclient hooks.
expected_syzygy_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', 'third_party', 'syzygy'))
expected_output_dir = os.path.join(expected_syzygy_dir, 'binaries')
if expected_output_dir != output_dir:
_LOGGER.info('Unexpected output directory, skipping cleanup.')
return
if not os.path.isdir(expected_syzygy_dir):
_LOGGER.info('Output directory does not exist, skipping cleanup.')
return
def OnError(function, path, excinfo):
"""Logs error encountered by shutil.rmtree."""
_LOGGER.error('Error when running %s(%s)', function, path, exc_info=excinfo)
_LOGGER.info('Removing orphaned files from %s', expected_syzygy_dir)
if not options.dry_run:
shutil.rmtree(expected_syzygy_dir, True, OnError)
def main():
options = _ParseCommandLine()
if options.dry_run:
_LOGGER.debug('Performing a dry-run.')
# We only care about Windows platforms, as the Syzygy binaries aren't used
# elsewhere. However, there was a short period of time where this script
# wasn't gated on OS types, and those OSes downloaded and installed binaries.
# This will cleanup orphaned files on those operating systems.
if sys.platform not in ('win32', 'cygwin'):
if options.no_cleanup:
_LOGGER.debug('Skipping usual cleanup for non-Windows platforms.')
else:
return _RemoveOrphanedFiles(options)
# Load the current installation state, and validate it against the
# requested installation.
state, is_consistent = _GetCurrentState(options.revision, options.output_dir)
# Decide whether or not an install is necessary.
if options.force:
_LOGGER.debug('Forcing reinstall of binaries.')
elif is_consistent:
# Avoid doing any work if the contents of the directory are consistent.
_LOGGER.debug('State unchanged, no reinstall necessary.')
return
# Under normal logging this is the only only message that will be reported.
_LOGGER.info('Installing revision %s Syzygy binaries.',
options.revision[0:12])
# Clean up the old state to begin with.
deleted = []
if options.overwrite:
if os.path.exists(options.output_dir):
# If overwrite was specified then take a heavy-handed approach.
_LOGGER.debug('Deleting entire installation directory.')
if not options.dry_run:
_RmTree(options.output_dir)
else:
# Otherwise only delete things that the previous installation put in place,
# and take care to preserve any local changes.
deleted = _CleanState(options.output_dir, state, options.dry_run)
# Install the new binaries. In a dry-run this will actually download the
# archives, but it won't write anything to disk.
state = _InstallBinaries(options, deleted)
# Build and save the state for the directory.
_SaveState(options.output_dir, state, options.dry_run)
if __name__ == '__main__':
main()