|  | #!/usr/bin/env python | 
|  | # Copyright 2014 The Chromium Authors. All rights reserved. | 
|  | # Use of this source code is governed by a BSD-style license that can be | 
|  | # found in the LICENSE file. | 
|  |  | 
|  | """A utility script for downloading versioned Syzygy binaries.""" | 
|  |  | 
|  | import hashlib | 
|  | import errno | 
|  | import json | 
|  | import logging | 
|  | import optparse | 
|  | import os | 
|  | import re | 
|  | import shutil | 
|  | import stat | 
|  | import sys | 
|  | import subprocess | 
|  | import tempfile | 
|  | import time | 
|  | import zipfile | 
|  |  | 
|  |  | 
|  | _LOGGER = logging.getLogger(os.path.basename(__file__)) | 
|  |  | 
|  | # The relative path where official builds are archived in their GS bucket. | 
|  | _SYZYGY_ARCHIVE_PATH = ('/builds/official/%(revision)s') | 
|  |  | 
|  | # A JSON file containing the state of the download directory. If this file and | 
|  | # directory state do not agree, then the binaries will be downloaded and | 
|  | # installed again. | 
|  | _STATE = '.state' | 
|  |  | 
|  | # This matches an integer (an SVN revision number) or a SHA1 value (a GIT hash). | 
|  | # The archive exclusively uses lowercase GIT hashes. | 
|  | _REVISION_RE = re.compile('^(?:\d+|[a-f0-9]{40})$') | 
|  |  | 
|  | # This matches an MD5 hash. | 
|  | _MD5_RE = re.compile('^[a-f0-9]{32}$') | 
|  |  | 
|  | # List of reources to be downloaded and installed. These are tuples with the | 
|  | # following format: | 
|  | # (basename, logging name, relative installation path, extraction filter) | 
|  | _RESOURCES = [ | 
|  | ('benchmark.zip', 'benchmark', '', None), | 
|  | ('binaries.zip', 'binaries', 'exe', None), | 
|  | ('symbols.zip', 'symbols', 'exe', | 
|  | lambda x: x.filename.endswith('.dll.pdb'))] | 
|  |  | 
|  |  | 
|  | # Name of the MS DIA dll that we need to copy to the binaries directory. | 
|  | _DIA_DLL_NAME = "msdia140.dll" | 
|  |  | 
|  |  | 
|  | def _LoadState(output_dir): | 
|  | """Loads the contents of the state file for a given |output_dir|, returning | 
|  | None if it doesn't exist. | 
|  | """ | 
|  | path = os.path.join(output_dir, _STATE) | 
|  | if not os.path.exists(path): | 
|  | _LOGGER.debug('No state file found.') | 
|  | return None | 
|  | with open(path, 'rb') as f: | 
|  | _LOGGER.debug('Reading state file: %s', path) | 
|  | try: | 
|  | return json.load(f) | 
|  | except ValueError: | 
|  | _LOGGER.debug('Invalid state file.') | 
|  | return None | 
|  |  | 
|  |  | 
|  | def _SaveState(output_dir, state, dry_run=False): | 
|  | """Saves the |state| dictionary to the given |output_dir| as a JSON file.""" | 
|  | path = os.path.join(output_dir, _STATE) | 
|  | _LOGGER.debug('Writing state file: %s', path) | 
|  | if dry_run: | 
|  | return | 
|  | with open(path, 'wb') as f: | 
|  | f.write(json.dumps(state, sort_keys=True, indent=2)) | 
|  |  | 
|  |  | 
|  | def _Md5(path): | 
|  | """Returns the MD5 hash of the file at |path|, which must exist.""" | 
|  | return hashlib.md5(open(path, 'rb').read()).hexdigest() | 
|  |  | 
|  |  | 
|  | def _StateIsValid(state): | 
|  | """Returns true if the given state structure is valid.""" | 
|  | if not isinstance(state, dict): | 
|  | _LOGGER.debug('State must be a dict.') | 
|  | return False | 
|  | r = state.get('revision', None) | 
|  | if not isinstance(r, basestring) or not _REVISION_RE.match(r): | 
|  | _LOGGER.debug('State contains an invalid revision.') | 
|  | return False | 
|  | c = state.get('contents', None) | 
|  | if not isinstance(c, dict): | 
|  | _LOGGER.debug('State must contain a contents dict.') | 
|  | return False | 
|  | for (relpath, md5) in c.iteritems(): | 
|  | if not isinstance(relpath, basestring) or len(relpath) == 0: | 
|  | _LOGGER.debug('State contents dict contains an invalid path.') | 
|  | return False | 
|  | if not isinstance(md5, basestring) or not _MD5_RE.match(md5): | 
|  | _LOGGER.debug('State contents dict contains an invalid MD5 digest.') | 
|  | return False | 
|  | return True | 
|  |  | 
|  |  | 
|  | def _BuildActualState(stored, revision, output_dir): | 
|  | """Builds the actual state using the provided |stored| state as a template. | 
|  | Only examines files listed in the stored state, causing the script to ignore | 
|  | files that have been added to the directories locally. |stored| must be a | 
|  | valid state dictionary. | 
|  | """ | 
|  | contents = {} | 
|  | state = { 'revision': revision, 'contents': contents } | 
|  | for relpath, md5 in stored['contents'].iteritems(): | 
|  | abspath = os.path.abspath(os.path.join(output_dir, relpath)) | 
|  | if os.path.isfile(abspath): | 
|  | m = _Md5(abspath) | 
|  | contents[relpath] = m | 
|  |  | 
|  | return state | 
|  |  | 
|  |  | 
|  | def _StatesAreConsistent(stored, actual): | 
|  | """Validates whether two state dictionaries are consistent. Both must be valid | 
|  | state dictionaries. Additional entries in |actual| are ignored. | 
|  | """ | 
|  | if stored['revision'] != actual['revision']: | 
|  | _LOGGER.debug('Mismatched revision number.') | 
|  | return False | 
|  | cont_stored = stored['contents'] | 
|  | cont_actual = actual['contents'] | 
|  | for relpath, md5 in cont_stored.iteritems(): | 
|  | if relpath not in cont_actual: | 
|  | _LOGGER.debug('Missing content: %s', relpath) | 
|  | return False | 
|  | if md5 != cont_actual[relpath]: | 
|  | _LOGGER.debug('Modified content: %s', relpath) | 
|  | return False | 
|  | return True | 
|  |  | 
|  |  | 
|  | def _GetCurrentState(revision, output_dir): | 
|  | """Loads the current state and checks to see if it is consistent. Returns | 
|  | a tuple (state, bool). The returned state will always be valid, even if an | 
|  | invalid state is present on disk. | 
|  | """ | 
|  | stored = _LoadState(output_dir) | 
|  | if not _StateIsValid(stored): | 
|  | _LOGGER.debug('State is invalid.') | 
|  | # Return a valid but empty state. | 
|  | return ({'revision': '0', 'contents': {}}, False) | 
|  | actual = _BuildActualState(stored, revision, output_dir) | 
|  | # If the script has been modified consider the state invalid. | 
|  | path = os.path.join(output_dir, _STATE) | 
|  | if os.path.getmtime(__file__) > os.path.getmtime(path): | 
|  | return (stored, False) | 
|  | # Otherwise, explicitly validate the state. | 
|  | if not _StatesAreConsistent(stored, actual): | 
|  | return (stored, False) | 
|  | return (stored, True) | 
|  |  | 
|  |  | 
|  | def _DirIsEmpty(path): | 
|  | """Returns true if the given directory is empty, false otherwise.""" | 
|  | for root, dirs, files in os.walk(path): | 
|  | return not dirs and not files | 
|  |  | 
|  |  | 
|  | def _RmTreeHandleReadOnly(func, path, exc): | 
|  | """An error handling function for use with shutil.rmtree. This will | 
|  | detect failures to remove read-only files, and will change their properties | 
|  | prior to removing them. This is necessary on Windows as os.remove will return | 
|  | an access error for read-only files, and git repos contain read-only | 
|  | pack/index files. | 
|  | """ | 
|  | excvalue = exc[1] | 
|  | if func in (os.rmdir, os.remove) and excvalue.errno == errno.EACCES: | 
|  | _LOGGER.debug('Removing read-only path: %s', path) | 
|  | os.chmod(path, stat.S_IRWXU | stat.S_IRWXG | stat.S_IRWXO) | 
|  | func(path) | 
|  | else: | 
|  | raise | 
|  |  | 
|  |  | 
|  | def _RmTree(path): | 
|  | """A wrapper of shutil.rmtree that handles read-only files.""" | 
|  | shutil.rmtree(path, ignore_errors=False, onerror=_RmTreeHandleReadOnly) | 
|  |  | 
|  |  | 
|  | def _CleanState(output_dir, state, dry_run=False): | 
|  | """Cleans up files/directories in |output_dir| that are referenced by | 
|  | the given |state|. Raises an error if there are local changes. Returns a | 
|  | dictionary of files that were deleted. | 
|  | """ | 
|  | _LOGGER.debug('Deleting files from previous installation.') | 
|  | deleted = {} | 
|  |  | 
|  | # Generate a list of files to delete, relative to |output_dir|. | 
|  | contents = state['contents'] | 
|  | files = sorted(contents.keys()) | 
|  |  | 
|  | # Try to delete the files. Keep track of directories to delete as well. | 
|  | dirs = {} | 
|  | for relpath in files: | 
|  | fullpath = os.path.join(output_dir, relpath) | 
|  | fulldir = os.path.dirname(fullpath) | 
|  | dirs[fulldir] = True | 
|  | if os.path.exists(fullpath): | 
|  | # If somehow the file has become a directory complain about it. | 
|  | if os.path.isdir(fullpath): | 
|  | raise Exception('Directory exists where file expected: %s' % fullpath) | 
|  |  | 
|  | # Double check that the file doesn't have local changes. If it does | 
|  | # then refuse to delete it. | 
|  | if relpath in contents: | 
|  | stored_md5 = contents[relpath] | 
|  | actual_md5 = _Md5(fullpath) | 
|  | if actual_md5 != stored_md5: | 
|  | raise Exception('File has local changes: %s' % fullpath) | 
|  |  | 
|  | # The file is unchanged so it can safely be deleted. | 
|  | _LOGGER.debug('Deleting file "%s".', fullpath) | 
|  | deleted[relpath] = True | 
|  | if not dry_run: | 
|  | os.unlink(fullpath) | 
|  |  | 
|  | # Sort directories from longest name to shortest. This lets us remove empty | 
|  | # directories from the most nested paths first. | 
|  | dirs = sorted(dirs.keys(), key=lambda x: len(x), reverse=True) | 
|  | for p in dirs: | 
|  | if os.path.exists(p) and _DirIsEmpty(p): | 
|  | _LOGGER.debug('Deleting empty directory "%s".', p) | 
|  | if not dry_run: | 
|  | _RmTree(p) | 
|  |  | 
|  | return deleted | 
|  |  | 
|  |  | 
|  | def _FindGsUtil(): | 
|  | """Looks for depot_tools and returns the absolute path to gsutil.py.""" | 
|  | for path in os.environ['PATH'].split(os.pathsep): | 
|  | path = os.path.abspath(path) | 
|  | git_cl = os.path.join(path, 'git_cl.py') | 
|  | gs_util = os.path.join(path, 'gsutil.py') | 
|  | if os.path.exists(git_cl) and os.path.exists(gs_util): | 
|  | return gs_util | 
|  | return None | 
|  |  | 
|  |  | 
|  | def _GsUtil(*cmd): | 
|  | """Runs the given command in gsutil with exponential backoff and retries.""" | 
|  | gs_util = _FindGsUtil() | 
|  | cmd = [sys.executable, gs_util] + list(cmd) | 
|  |  | 
|  | retries = 3 | 
|  | timeout = 4  # Seconds. | 
|  | while True: | 
|  | _LOGGER.debug('Running %s', cmd) | 
|  | prog = subprocess.Popen(cmd, shell=False) | 
|  | prog.communicate() | 
|  |  | 
|  | # Stop retrying on success. | 
|  | if prog.returncode == 0: | 
|  | return | 
|  |  | 
|  | # Raise a permanent failure if retries have been exhausted. | 
|  | if retries == 0: | 
|  | raise RuntimeError('Command "%s" returned %d.' % (cmd, prog.returncode)) | 
|  |  | 
|  | _LOGGER.debug('Sleeping %d seconds and trying again.', timeout) | 
|  | time.sleep(timeout) | 
|  | retries -= 1 | 
|  | timeout *= 2 | 
|  |  | 
|  |  | 
|  | def _Download(resource): | 
|  | """Downloads the given GS resource to a temporary file, returning its path.""" | 
|  | tmp = tempfile.mkstemp(suffix='syzygy_archive') | 
|  | os.close(tmp[0]) | 
|  | tmp_file = tmp[1] | 
|  | url = 'gs://syzygy-archive' + resource | 
|  | if sys.platform == 'cygwin': | 
|  | # Change temporary path to Windows path for gsutil | 
|  | def winpath(path): | 
|  | return subprocess.check_output(['cygpath', '-w', path]).strip() | 
|  | tmp_file = winpath(tmp_file) | 
|  | _GsUtil('cp', url, tmp_file) | 
|  | return tmp[1] | 
|  |  | 
|  |  | 
|  | def _MaybeCopyDIABinaries(options, contents): | 
|  | """Try to copy the DIA DLL to the binaries exe directory.""" | 
|  | toolchain_data_file = os.path.join(os.path.dirname(__file__), | 
|  | 'win_toolchain.json') | 
|  | if not os.path.exists(toolchain_data_file): | 
|  | _LOGGER.debug('Toolchain JSON data file doesn\'t exist, skipping.') | 
|  | return | 
|  | with open(toolchain_data_file) as temp_f: | 
|  | toolchain_data = json.load(temp_f) | 
|  | if not os.path.isdir(toolchain_data['path']): | 
|  | _LOGGER.error('The toolchain JSON file is invalid.') | 
|  | return | 
|  | dia_sdk_binaries_dir = os.path.join(toolchain_data['path'], 'DIA SDK', 'bin') | 
|  | dia_dll = os.path.join(dia_sdk_binaries_dir, _DIA_DLL_NAME) | 
|  | if not os.path.exists(dia_dll): | 
|  | _LOGGER.debug('%s is missing, skipping.') | 
|  | return | 
|  | dia_dll_dest = os.path.join(options.output_dir, 'exe', _DIA_DLL_NAME) | 
|  | _LOGGER.debug('Copying %s to %s.' % (dia_dll, dia_dll_dest)) | 
|  | if not options.dry_run: | 
|  | shutil.copy(dia_dll, dia_dll_dest) | 
|  | contents[os.path.relpath(dia_dll_dest, options.output_dir)] = ( | 
|  | _Md5(dia_dll_dest)) | 
|  |  | 
|  |  | 
|  | def _InstallBinaries(options, deleted={}): | 
|  | """Installs Syzygy binaries. This assumes that the output directory has | 
|  | already been cleaned, as it will refuse to overwrite existing files.""" | 
|  | contents = {} | 
|  | state = { 'revision': options.revision, 'contents': contents } | 
|  | archive_path = _SYZYGY_ARCHIVE_PATH % { 'revision': options.revision } | 
|  | if options.resources: | 
|  | resources = [(resource, resource, '', None) | 
|  | for resource in options.resources] | 
|  | else: | 
|  | resources = _RESOURCES | 
|  | for (base, name, subdir, filt) in resources: | 
|  | # Create the output directory if it doesn't exist. | 
|  | fulldir = os.path.join(options.output_dir, subdir) | 
|  | if os.path.isfile(fulldir): | 
|  | raise Exception('File exists where a directory needs to be created: %s' % | 
|  | fulldir) | 
|  | if not os.path.exists(fulldir): | 
|  | _LOGGER.debug('Creating directory: %s', fulldir) | 
|  | if not options.dry_run: | 
|  | os.makedirs(fulldir) | 
|  |  | 
|  | # Download and read the archive. | 
|  | resource = archive_path + '/' + base | 
|  | _LOGGER.debug('Retrieving %s archive at "%s".', name, resource) | 
|  | path = _Download(resource) | 
|  |  | 
|  | _LOGGER.debug('Unzipping %s archive.', name) | 
|  | with open(path, 'rb') as data: | 
|  | archive = zipfile.ZipFile(data) | 
|  | for entry in archive.infolist(): | 
|  | if not filt or filt(entry): | 
|  | fullpath = os.path.normpath(os.path.join(fulldir, entry.filename)) | 
|  | relpath = os.path.relpath(fullpath, options.output_dir) | 
|  | if os.path.exists(fullpath): | 
|  | # If in a dry-run take into account the fact that the file *would* | 
|  | # have been deleted. | 
|  | if options.dry_run and relpath in deleted: | 
|  | pass | 
|  | else: | 
|  | raise Exception('Path already exists: %s' % fullpath) | 
|  |  | 
|  | # Extract the file and update the state dictionary. | 
|  | _LOGGER.debug('Extracting "%s".', fullpath) | 
|  | if not options.dry_run: | 
|  | archive.extract(entry.filename, fulldir) | 
|  | md5 = _Md5(fullpath) | 
|  | contents[relpath] = md5 | 
|  | if sys.platform == 'cygwin': | 
|  | os.chmod(fullpath, os.stat(fullpath).st_mode | stat.S_IXUSR) | 
|  |  | 
|  | _LOGGER.debug('Removing temporary file "%s".', path) | 
|  | os.remove(path) | 
|  |  | 
|  | if options.copy_dia_binaries: | 
|  | # Try to copy the DIA binaries to the binaries directory. | 
|  | _MaybeCopyDIABinaries(options, contents) | 
|  |  | 
|  | return state | 
|  |  | 
|  |  | 
|  | def _ParseCommandLine(): | 
|  | """Parses the command-line and returns an options structure.""" | 
|  | option_parser = optparse.OptionParser() | 
|  | option_parser.add_option('--dry-run', action='store_true', default=False, | 
|  | help='If true then will simply list actions that would be performed.') | 
|  | option_parser.add_option('--force', action='store_true', default=False, | 
|  | help='Force an installation even if the binaries are up to date.') | 
|  | option_parser.add_option('--no-cleanup', action='store_true', default=False, | 
|  | help='Allow installation on non-Windows platforms, and skip the forced ' | 
|  | 'cleanup step.') | 
|  | option_parser.add_option('--output-dir', type='string', | 
|  | help='The path where the binaries will be replaced. Existing binaries ' | 
|  | 'will only be overwritten if not up to date.') | 
|  | option_parser.add_option('--overwrite', action='store_true', default=False, | 
|  | help='If specified then the installation will happily delete and rewrite ' | 
|  | 'the entire output directory, blasting any local changes.') | 
|  | option_parser.add_option('--revision', type='string', | 
|  | help='The SVN revision or GIT hash associated with the required version.') | 
|  | option_parser.add_option('--revision-file', type='string', | 
|  | help='A text file containing an SVN revision or GIT hash.') | 
|  | option_parser.add_option('--resource', type='string', action='append', | 
|  | dest='resources', help='A resource to be downloaded.') | 
|  | option_parser.add_option('--verbose', dest='log_level', action='store_const', | 
|  | default=logging.INFO, const=logging.DEBUG, | 
|  | help='Enables verbose logging.') | 
|  | option_parser.add_option('--quiet', dest='log_level', action='store_const', | 
|  | default=logging.INFO, const=logging.ERROR, | 
|  | help='Disables all output except for errors.') | 
|  | option_parser.add_option('--copy-dia-binaries', action='store_true', | 
|  | default=False, help='If true then the DIA dll will get copied into the ' | 
|  | 'binaries directory if it\'s available.') | 
|  | options, args = option_parser.parse_args() | 
|  | if args: | 
|  | option_parser.error('Unexpected arguments: %s' % args) | 
|  | if not options.output_dir: | 
|  | option_parser.error('Must specify --output-dir.') | 
|  | if not options.revision and not options.revision_file: | 
|  | option_parser.error('Must specify one of --revision or --revision-file.') | 
|  | if options.revision and options.revision_file: | 
|  | option_parser.error('Must not specify both --revision and --revision-file.') | 
|  |  | 
|  | # Configure logging. | 
|  | logging.basicConfig(level=options.log_level) | 
|  |  | 
|  | # If a revision file has been specified then read it. | 
|  | if options.revision_file: | 
|  | options.revision = open(options.revision_file, 'rb').read().strip() | 
|  | _LOGGER.debug('Parsed revision "%s" from file "%s".', | 
|  | options.revision, options.revision_file) | 
|  |  | 
|  | # Ensure that the specified SVN revision or GIT hash is valid. | 
|  | if not _REVISION_RE.match(options.revision): | 
|  | option_parser.error('Must specify a valid SVN or GIT revision.') | 
|  |  | 
|  | # This just makes output prettier to read. | 
|  | options.output_dir = os.path.normpath(options.output_dir) | 
|  |  | 
|  | return options | 
|  |  | 
|  |  | 
|  | def _RemoveOrphanedFiles(options): | 
|  | """This is run on non-Windows systems to remove orphaned files that may have | 
|  | been downloaded by a previous version of this script. | 
|  | """ | 
|  | # Reconfigure logging to output info messages. This will allow inspection of | 
|  | # cleanup status on non-Windows buildbots. | 
|  | _LOGGER.setLevel(logging.INFO) | 
|  |  | 
|  | output_dir = os.path.abspath(options.output_dir) | 
|  |  | 
|  | # We only want to clean up the folder in 'src/third_party/syzygy', and we | 
|  | # expect to be called with that as an output directory. This is an attempt to | 
|  | # not start deleting random things if the script is run from an alternate | 
|  | # location, or not called from the gclient hooks. | 
|  | expected_syzygy_dir = os.path.abspath(os.path.join( | 
|  | os.path.dirname(__file__), '..', 'third_party', 'syzygy')) | 
|  | expected_output_dir = os.path.join(expected_syzygy_dir, 'binaries') | 
|  | if expected_output_dir != output_dir: | 
|  | _LOGGER.info('Unexpected output directory, skipping cleanup.') | 
|  | return | 
|  |  | 
|  | if not os.path.isdir(expected_syzygy_dir): | 
|  | _LOGGER.info('Output directory does not exist, skipping cleanup.') | 
|  | return | 
|  |  | 
|  | def OnError(function, path, excinfo): | 
|  | """Logs error encountered by shutil.rmtree.""" | 
|  | _LOGGER.error('Error when running %s(%s)', function, path, exc_info=excinfo) | 
|  |  | 
|  | _LOGGER.info('Removing orphaned files from %s', expected_syzygy_dir) | 
|  | if not options.dry_run: | 
|  | shutil.rmtree(expected_syzygy_dir, True, OnError) | 
|  |  | 
|  |  | 
|  | def main(): | 
|  | options = _ParseCommandLine() | 
|  |  | 
|  | if options.dry_run: | 
|  | _LOGGER.debug('Performing a dry-run.') | 
|  |  | 
|  | # We only care about Windows platforms, as the Syzygy binaries aren't used | 
|  | # elsewhere. However, there was a short period of time where this script | 
|  | # wasn't gated on OS types, and those OSes downloaded and installed binaries. | 
|  | # This will cleanup orphaned files on those operating systems. | 
|  | if sys.platform not in ('win32', 'cygwin'): | 
|  | if options.no_cleanup: | 
|  | _LOGGER.debug('Skipping usual cleanup for non-Windows platforms.') | 
|  | else: | 
|  | return _RemoveOrphanedFiles(options) | 
|  |  | 
|  | # Load the current installation state, and validate it against the | 
|  | # requested installation. | 
|  | state, is_consistent = _GetCurrentState(options.revision, options.output_dir) | 
|  |  | 
|  | # Decide whether or not an install is necessary. | 
|  | if options.force: | 
|  | _LOGGER.debug('Forcing reinstall of binaries.') | 
|  | elif is_consistent: | 
|  | # Avoid doing any work if the contents of the directory are consistent. | 
|  | _LOGGER.debug('State unchanged, no reinstall necessary.') | 
|  | return | 
|  |  | 
|  | # Under normal logging this is the only only message that will be reported. | 
|  | _LOGGER.info('Installing revision %s Syzygy binaries.', | 
|  | options.revision[0:12]) | 
|  |  | 
|  | # Clean up the old state to begin with. | 
|  | deleted = [] | 
|  | if options.overwrite: | 
|  | if os.path.exists(options.output_dir): | 
|  | # If overwrite was specified then take a heavy-handed approach. | 
|  | _LOGGER.debug('Deleting entire installation directory.') | 
|  | if not options.dry_run: | 
|  | _RmTree(options.output_dir) | 
|  | else: | 
|  | # Otherwise only delete things that the previous installation put in place, | 
|  | # and take care to preserve any local changes. | 
|  | deleted = _CleanState(options.output_dir, state, options.dry_run) | 
|  |  | 
|  | # Install the new binaries. In a dry-run this will actually download the | 
|  | # archives, but it won't write anything to disk. | 
|  | state = _InstallBinaries(options, deleted) | 
|  |  | 
|  | # Build and save the state for the directory. | 
|  | _SaveState(options.output_dir, state, options.dry_run) | 
|  |  | 
|  |  | 
|  | if __name__ == '__main__': | 
|  | main() |