blob: 235c33fdbc5abc84673c3379bde6bb4f4cc13fda [file] [log] [blame]
#!/usr/bin/env vpython3
# Copyright 2013 The Chromium Authors
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
""" A utility to generate an up-to-date orderfile.
The orderfile is used by the linker to order text sections such that the
sections are placed consecutively in the order specified. This allows us
to page in less code during start-up.
Example usage:
tools/cygprofile/orderfile_generator_backend.py --use-goma --target-arch=arm
"""
import argparse
import csv
import hashlib
import json
import glob
import logging
import os
import shutil
import subprocess
import sys
import tempfile
import time
import cluster
import cyglog_to_orderfile
import patch_orderfile
import process_profiles
import profile_android_startup
_SRC_PATH = os.path.join(os.path.dirname(__file__), os.pardir, os.pardir)
sys.path.append(os.path.join(_SRC_PATH, 'third_party', 'catapult', 'devil'))
from devil.android import device_utils
from devil.android.sdk import version_codes
_SRC_PATH = os.path.join(os.path.dirname(os.path.realpath(__file__)),
os.pardir, os.pardir)
sys.path.append(os.path.join(_SRC_PATH, 'build', 'android'))
import devil_chromium
from pylib import constants
# Needs to happen early for GetBuildType()/GetOutDirectory() to work correctly
constants.SetBuildType('Release')
# Architecture specific GN args. Trying to build an orderfile for an
# architecture not listed here will eventually throw.
_ARCH_GN_ARGS = {
'arm': ['target_cpu="arm"'],
'arm64': ['target_cpu="arm64"'],
'x86': ['target_cpu="x86"'],
'x64': ['target_cpu="x64"'],
}
class CommandError(Exception):
"""Indicates that a dispatched shell command exited with a non-zero status."""
def __init__(self, value):
super().__init__()
self.value = value
def __str__(self):
return repr(self.value)
def _GenerateHash(file_path):
"""Calculates and returns the hash of the file at file_path."""
sha1 = hashlib.sha1()
with open(file_path, 'rb') as f:
while True:
# Read in 1mb chunks, so it doesn't all have to be loaded into memory.
chunk = f.read(1024 * 1024)
if not chunk:
break
sha1.update(chunk)
return sha1.hexdigest()
def _GetFileExtension(file_name):
"""Calculates the file extension from a file name.
Args:
file_name: The source file name.
Returns:
The part of file_name after the dot (.) or None if the file has no
extension.
Examples: /home/user/foo.bar -> bar
/home/user.name/foo -> None
/home/user/.foo -> None
/home/user/foo.bar.baz -> baz
"""
file_name_parts = os.path.basename(file_name).split('.')
if len(file_name_parts) > 1:
return file_name_parts[-1]
return None
def _StashOutputDirectory(buildpath):
"""Takes the output directory and stashes it in the default output directory.
This allows it to be used for incremental builds next time (after unstashing)
by keeping it in a place that isn't deleted normally, while also ensuring
that it is properly clobbered when appropriate.
This is a dirty hack to deal with the needs of clobbering while also handling
incremental builds and the hardcoded relative paths used in some of the
project files.
Args:
buildpath: The path where the building happens. If this corresponds to the
default output directory, no action is taken.
"""
if os.path.abspath(buildpath) == os.path.abspath(os.path.dirname(
constants.GetOutDirectory())):
return
name = os.path.basename(buildpath)
stashpath = os.path.join(constants.GetOutDirectory(), name)
if not os.path.exists(buildpath):
return
if os.path.exists(stashpath):
shutil.rmtree(stashpath, ignore_errors=True)
shutil.move(buildpath, stashpath)
def _UnstashOutputDirectory(buildpath):
"""Inverse of _StashOutputDirectory.
Moves the output directory stashed within the default output directory
(out/Release) to the position where the builds can actually happen.
This is a dirty hack to deal with the needs of clobbering while also handling
incremental builds and the hardcoded relative paths used in some of the
project files.
Args:
buildpath: The path where the building happens. If this corresponds to the
default output directory, no action is taken.
"""
if os.path.abspath(buildpath) == os.path.abspath(os.path.dirname(
constants.GetOutDirectory())):
return
name = os.path.basename(buildpath)
stashpath = os.path.join(constants.GetOutDirectory(), name)
if not os.path.exists(stashpath):
return
if os.path.exists(buildpath):
shutil.rmtree(buildpath, ignore_errors=True)
shutil.move(stashpath, buildpath)
class StepRecorder:
"""Records steps and timings."""
def __init__(self, buildbot):
self.timings = []
self._previous_step = ('', 0.0)
self._buildbot = buildbot
self._error_recorded = False
def BeginStep(self, name):
"""Marks a beginning of the next step in the script.
On buildbot, this prints a specially formatted name that will show up
in the waterfall. Otherwise, just prints the step name.
Args:
name: The name of the step.
"""
self.EndStep()
self._previous_step = (name, time.time())
print('Running step: ', name)
def EndStep(self):
"""Records successful completion of the current step.
This is optional if the step is immediately followed by another BeginStep.
"""
if self._previous_step[0]:
elapsed = time.time() - self._previous_step[1]
print('Step %s took %f seconds' % (self._previous_step[0], elapsed))
self.timings.append((self._previous_step[0], elapsed))
self._previous_step = ('', 0.0)
def FailStep(self, message=None):
"""Marks that a particular step has failed.
On buildbot, this will mark the current step as failed on the waterfall.
Otherwise we will just print an optional failure message.
Args:
message: An optional explanation as to why the step failed.
"""
print('STEP FAILED!!')
if message:
print(message)
self._error_recorded = True
self.EndStep()
def ErrorRecorded(self):
"""True if FailStep has been called."""
return self._error_recorded
def RunCommand(self, cmd, cwd=constants.DIR_SOURCE_ROOT, raise_on_error=True,
stdout=None):
"""Execute a shell command.
Args:
cmd: A list of command strings.
cwd: Directory in which the command should be executed, defaults to build
root of script's location if not specified.
raise_on_error: If true will raise a CommandError if the call doesn't
succeed and mark the step as failed.
stdout: A file to redirect stdout for the command to.
Returns:
The process's return code.
Raises:
CommandError: An error executing the specified command.
"""
print('Executing %s in %s' % (' '.join(cmd), cwd))
process = subprocess.Popen(cmd, stdout=stdout, cwd=cwd, env=os.environ)
process.wait()
if raise_on_error and process.returncode != 0:
self.FailStep()
raise CommandError('Exception executing command %s' % ' '.join(cmd))
return process.returncode
class ClankCompiler:
"""Handles compilation of clank."""
def __init__(self, out_dir, step_recorder, arch, use_goma, goma_dir,
use_remoteexec, ninja_command, system_health_profiling,
monochrome, public, orderfile_location):
self._out_dir = out_dir
self._step_recorder = step_recorder
self._arch = arch
# TODO(b/236070141): remove goma config.
self._use_goma = use_goma
self._goma_dir = goma_dir
self._use_remoteexec = use_remoteexec
self._ninja_command = ninja_command
self._system_health_profiling = system_health_profiling
self._public = public
self._orderfile_location = orderfile_location
if monochrome:
self._apk = 'Monochrome.apk'
self._apk_target = 'monochrome_apk'
if '64' in self._arch:
# Monochrome has a _64 suffix for arm64 and x64 builds.
self._libname = 'libmonochrome_64'
self._libchrome_target = 'libmonochrome_64'
else:
self._libname = 'libmonochrome'
self._libchrome_target = 'libmonochrome'
else:
self._apk = 'Chrome.apk'
self._apk_target = 'chrome_apk'
self._libname = 'libchrome'
self._libchrome_target = 'libchrome'
if public:
self._apk = self._apk.replace('.apk', 'Public.apk')
self._apk_target = self._apk_target.replace('_apk', '_public_apk')
self.obj_dir = os.path.join(self._out_dir, 'Release', 'obj')
self.lib_chrome_so = os.path.join(
self._out_dir, 'Release', 'lib.unstripped',
'{}.so'.format(self._libname))
self.chrome_apk = os.path.join(self._out_dir, 'Release', 'apks', self._apk)
def Build(self, instrumented, use_call_graph, target):
"""Builds the provided ninja target with or without order_profiling on.
Args:
instrumented: (bool) Whether we want to build an instrumented binary.
use_call_graph: (bool) Whether to use the call graph instrumentation.
target: (str) The name of the ninja target to build.
"""
self._step_recorder.BeginStep('Compile %s' % target)
assert not use_call_graph or instrumented, ('You can not enable call graph '
'without instrumentation!')
# Set the "Release Official" flavor, the parts affecting performance.
args = [
'enable_resource_allowlist_generation=false',
'is_chrome_branded=' + str(not self._public).lower(),
'is_debug=false',
'is_official_build=true',
'symbol_level=1', # to fit 30 GiB RAM on the bot when LLD is running
'target_os="android"',
'use_goma=' + str(self._use_goma).lower(),
'use_remoteexec=' + str(self._use_remoteexec).lower(),
'use_order_profiling=' + str(instrumented).lower(),
'use_call_graph=' + str(use_call_graph).lower(),
]
args += _ARCH_GN_ARGS[self._arch]
if self._goma_dir:
args += ['goma_dir="%s"' % self._goma_dir]
if self._system_health_profiling:
args += ['devtools_instrumentation_dumping = ' +
str(instrumented).lower()]
if self._public and os.path.exists(self._orderfile_location):
# GN needs the orderfile path to be source-absolute.
src_abs_orderfile = os.path.relpath(self._orderfile_location,
constants.DIR_SOURCE_ROOT)
args += ['chrome_orderfile="//{}"'.format(src_abs_orderfile)]
self._step_recorder.RunCommand(
['gn', 'gen', os.path.join(self._out_dir, 'Release'),
'--args=' + ' '.join(args)])
self._step_recorder.RunCommand(
self._ninja_command + [os.path.join(self._out_dir, 'Release'), target])
def ForceRelink(self):
"""Forces libchrome.so or libmonochrome.so to be re-linked.
With partitioned libraries enabled, deleting these library files does not
guarantee they'll be recreated by the linker (they may simply be
re-extracted from a combined library). To be safe, touch a source file
instead. See http://crbug.com/972701 for more explanation.
"""
file_to_touch = os.path.join(constants.DIR_SOURCE_ROOT, 'chrome', 'browser',
'chrome_browser_main_android.cc')
assert os.path.exists(file_to_touch)
self._step_recorder.RunCommand(['touch', file_to_touch])
def CompileChromeApk(self, instrumented, use_call_graph, force_relink=False):
"""Builds a Chrome.apk either with or without order_profiling on.
Args:
instrumented: (bool) Whether to build an instrumented apk.
use_call_graph: (bool) Whether to use the call graph instrumentation.
force_relink: Whether libchromeview.so should be re-created.
"""
if force_relink:
self.ForceRelink()
self.Build(instrumented, use_call_graph, self._apk_target)
def CompileLibchrome(self, instrumented, use_call_graph, force_relink=False):
"""Builds a libchrome.so either with or without order_profiling on.
Args:
instrumented: (bool) Whether to build an instrumented apk.
use_call_graph: (bool) Whether to use the call graph instrumentation.
force_relink: (bool) Whether libchrome.so should be re-created.
"""
if force_relink:
self.ForceRelink()
self.Build(instrumented, use_call_graph, self._libchrome_target)
class OrderfileUpdater:
"""Handles uploading and committing a new orderfile in the repository.
Only used for testing or on a bot.
"""
_CLOUD_STORAGE_BUCKET_FOR_DEBUG = None
_CLOUD_STORAGE_BUCKET = None
_UPLOAD_TO_CLOUD_COMMAND = 'upload_to_google_storage.py'
def __init__(self, repository_root, step_recorder):
"""Constructor.
Args:
repository_root: (str) Root of the target repository.
step_recorder: (StepRecorder) Step recorder, for logging.
"""
self._repository_root = repository_root
self._step_recorder = step_recorder
def CommitStashedFileHashes(self, files):
"""Commits unpatched and patched orderfiles hashes if changed.
The files are committed only if their associated sha1 hash files match, and
are modified in git. In normal operations the hash files are changed only
when a file is uploaded to cloud storage. If the hash file is not modified
in git, the file is skipped.
Args:
files: [str or None] specifies file paths. None items are ignored.
Raises:
Exception if the hash file does not match the file.
NotImplementedError when the commit logic hasn't been overridden.
"""
files_to_commit = [_f for _f in files if _f]
if files_to_commit:
self._CommitStashedFiles(files_to_commit)
def UploadToCloudStorage(self, filename, use_debug_location):
"""Uploads a file to cloud storage.
Args:
filename: (str) File to upload.
use_debug_location: (bool) Whether to use the debug location.
"""
bucket = (self._CLOUD_STORAGE_BUCKET_FOR_DEBUG if use_debug_location
else self._CLOUD_STORAGE_BUCKET)
extension = _GetFileExtension(filename)
cmd = [self._UPLOAD_TO_CLOUD_COMMAND, '--bucket', bucket]
if extension:
cmd.extend(['-z', extension])
cmd.append(filename)
self._step_recorder.RunCommand(cmd)
print('Download: https://sandbox.google.com/storage/%s/%s' %
(bucket, _GenerateHash(filename)))
def _GetHashFilePathAndContents(self, filename):
"""Gets the name and content of the hash file created from uploading the
given file.
Args:
filename: (str) The file that was uploaded to cloud storage.
Returns:
A tuple of the hash file name, relative to the reository root, and the
content, which should be the sha1 hash of the file
('base_file.sha1', hash)
"""
abs_hash_filename = filename + '.sha1'
rel_hash_filename = os.path.relpath(
abs_hash_filename, self._repository_root)
with open(abs_hash_filename, 'r') as f:
return (rel_hash_filename, f.read())
def _GitStash(self):
"""Git stash the current clank tree.
Raises:
NotImplementedError when the stash logic hasn't been overridden.
"""
raise NotImplementedError
def _CommitStashedFiles(self, expected_files_in_stash):
"""Commits stashed files.
The local repository is updated and then the files to commit are taken from
modified files from the git stash. The modified files should be a subset of
|expected_files_in_stash|. If there are unexpected modified files, this
function may raise. This is meant to be paired with _GitStash().
Args:
expected_files_in_stash: [str] paths to a possible superset of files
expected to be stashed & committed.
Raises:
NotImplementedError when the commit logic hasn't been overridden.
"""
raise NotImplementedError
class OrderfileGenerator:
"""A utility for generating a new orderfile for Clank.
Builds an instrumented binary, profiles a run of the application, and
generates an updated orderfile.
"""
_CHECK_ORDERFILE_SCRIPT = os.path.join(
constants.DIR_SOURCE_ROOT, 'tools', 'cygprofile', 'check_orderfile.py')
_BUILD_ROOT = os.path.abspath(os.path.dirname(os.path.dirname(
constants.GetOutDirectory()))) # Normally /path/to/src
# Previous orderfile_generator debug files would be overwritten.
_DIRECTORY_FOR_DEBUG_FILES = '/tmp/orderfile_generator_debug_files'
_CLOUD_STORAGE_BUCKET_FOR_DEBUG = None
def _PrepareOrderfilePaths(self):
if self._options.public:
self._clank_dir = os.path.join(constants.DIR_SOURCE_ROOT,
'')
if not os.path.exists(os.path.join(self._clank_dir, 'orderfiles')):
os.makedirs(os.path.join(self._clank_dir, 'orderfiles'))
else:
self._clank_dir = os.path.join(constants.DIR_SOURCE_ROOT,
'clank')
self._unpatched_orderfile_filename = os.path.join(
self._clank_dir, 'orderfiles', 'unpatched_orderfile.%s')
self._path_to_orderfile = os.path.join(
self._clank_dir, 'orderfiles', 'orderfile.%s.out')
def _GetPathToOrderfile(self):
"""Gets the path to the architecture-specific orderfile."""
# TODO(https://crbug.com/1517659): We are testing if arm64 can improve perf
# while not regressing arm32 memory or perf by too much. For now we are
# keeping the fake arch as 'arm' to avoid needing to change the path. In
# the future we should consider either generating multiple orderfiles,
# one per architecture, or remove the fake arch as it would no longer be
# accurate.
# Build GN files use the ".arm" orderfile irrespective of the actual
# architecture. Fake it, otherwise the orderfile we generate here is not
# going to be picked up by builds.
orderfile_fake_arch = 'arm'
return self._path_to_orderfile % orderfile_fake_arch
def _GetUnpatchedOrderfileFilename(self):
"""Gets the path to the architecture-specific unpatched orderfile."""
return self._unpatched_orderfile_filename % self._options.arch
def _SetDevice(self):
""" Selects the device to be used by the script.
Returns:
(Device with given serial ID) : if the --device flag is set.
(Device running Android[K,L]) : if --use-legacy-chrome-apk flag is set or
no device running Android N+ was found.
(Device running Android N+) : Otherwise.
Raises Error:
If no device meeting the requirements has been found.
"""
devices = None
if self._options.device:
devices = [device_utils.DeviceUtils(self._options.device)]
else:
devices = device_utils.DeviceUtils.HealthyDevices()
assert devices, 'Expected at least one connected device'
if self._options.use_legacy_chrome_apk:
self._monochrome = False
for device in devices:
device_version = device.build_version_sdk
if (version_codes.KITKAT <= device_version <=
version_codes.LOLLIPOP_MR1):
return device
assert not self._options.use_legacy_chrome_apk, \
'No device found running suitable android version for Chrome.apk.'
preferred_device = None
for device in devices:
if device.build_version_sdk >= version_codes.NOUGAT:
preferred_device = device
break
self._monochrome = preferred_device is not None
return preferred_device if preferred_device else devices[0]
def __init__(self, options, orderfile_updater_class):
self._options = options
self._ninja_command = ['autoninja']
if self._options.ninja_path:
self._ninja_command = [self._options.ninja_path]
if self._options.ninja_j:
self._ninja_command += ['-j', self._options.ninja_j]
self._ninja_command += ['-C']
self._instrumented_out_dir = os.path.join(
self._BUILD_ROOT, self._options.arch + '_instrumented_out')
if self._options.use_call_graph:
self._instrumented_out_dir += '_call_graph'
self._uninstrumented_out_dir = os.path.join(
self._BUILD_ROOT, self._options.arch + '_uninstrumented_out')
self._no_orderfile_out_dir = os.path.join(
self._BUILD_ROOT, self._options.arch + '_no_orderfile_out')
self._PrepareOrderfilePaths()
if options.profile:
output_directory = os.path.join(self._instrumented_out_dir, 'Release')
host_profile_dir = os.path.join(output_directory, 'profile_data')
urls = [profile_android_startup.AndroidProfileTool.TEST_URL]
use_wpr = True
simulate_user = False
urls = options.urls
use_wpr = not options.no_wpr
simulate_user = options.simulate_user
device = self._SetDevice()
self._profiler = profile_android_startup.AndroidProfileTool(
output_directory, host_profile_dir, use_wpr, urls, simulate_user,
device, debug=self._options.streamline_for_debugging)
if options.pregenerated_profiles:
self._profiler.SetPregeneratedProfiles(
glob.glob(options.pregenerated_profiles))
else:
assert not options.pregenerated_profiles, (
'--pregenerated-profiles cannot be used with --skip-profile')
assert not options.profile_save_dir, (
'--profile-save-dir cannot be used with --skip-profile')
self._monochrome = not self._options.use_legacy_chrome_apk
# Outlined function handling enabled by default for all architectures.
self._order_outlined_functions = not options.noorder_outlined_functions
self._output_data = {}
self._step_recorder = StepRecorder(options.buildbot)
self._compiler = None
if orderfile_updater_class is None:
orderfile_updater_class = OrderfileUpdater
assert issubclass(orderfile_updater_class, OrderfileUpdater)
self._orderfile_updater = orderfile_updater_class(self._clank_dir,
self._step_recorder)
assert os.path.isdir(constants.DIR_SOURCE_ROOT), 'No src directory found'
@staticmethod
def _RemoveBlanks(src_file, dest_file):
"""A utility to remove blank lines from a file.
Args:
src_file: The name of the file to remove the blanks from.
dest_file: The name of the file to write the output without blanks.
"""
assert src_file != dest_file, 'Source and destination need to be distinct'
try:
src = open(src_file, 'r')
dest = open(dest_file, 'w')
for line in src:
if line and not line.isspace():
dest.write(line)
finally:
src.close()
dest.close()
def _GenerateAndProcessProfile(self):
"""Invokes a script to merge the per-thread traces into one file.
The produced list of offsets is saved in
self._GetUnpatchedOrderfileFilename().
"""
self._step_recorder.BeginStep('Generate Profile Data')
files = []
logging.getLogger().setLevel(logging.DEBUG)
if self._options.profile_save_dir:
# The directory must not preexist, to ensure purity of data. Check
# before profiling to save time.
if os.path.exists(self._options.profile_save_dir):
raise Exception('Profile save directory must not pre-exist')
os.makedirs(self._options.profile_save_dir)
if self._options.system_health_orderfile:
files = self._profiler.CollectSystemHealthProfile(
self._compiler.chrome_apk)
self._MaybeSaveProfile(files)
try:
self._ProcessPhasedOrderfile(files)
except Exception:
for f in files:
self._SaveForDebugging(f)
self._SaveForDebugging(self._compiler.lib_chrome_so)
raise
finally:
self._profiler.Cleanup()
else:
self._CollectLegacyProfile()
logging.getLogger().setLevel(logging.INFO)
def _ProcessPhasedOrderfile(self, files):
"""Process the phased orderfiles produced by system health benchmarks.
The offsets will be placed in _GetUnpatchedOrderfileFilename().
Args:
file: Profile files pulled locally.
"""
self._step_recorder.BeginStep('Process Phased Orderfile')
profiles = process_profiles.ProfileManager(files)
processor = process_profiles.SymbolOffsetProcessor(
self._compiler.lib_chrome_so)
ordered_symbols = cluster.ClusterOffsets(profiles, processor,
call_graph=self._options.use_call_graph)
if not ordered_symbols:
raise Exception('Failed to get ordered symbols')
for sym in ordered_symbols:
assert not sym.startswith('OUTLINED_FUNCTION_'), (
'Outlined function found in instrumented function, very likely '
'something has gone very wrong!')
self._output_data['offsets_kib'] = processor.SymbolsSize(
ordered_symbols) / 1024
with open(self._GetUnpatchedOrderfileFilename(), 'w') as orderfile:
orderfile.write('\n'.join(ordered_symbols))
def _CollectLegacyProfile(self):
files = []
try:
files = self._profiler.CollectProfile(
self._compiler.chrome_apk,
constants.PACKAGE_INFO['chrome'])
self._MaybeSaveProfile(files)
self._step_recorder.BeginStep('Process profile')
assert os.path.exists(self._compiler.lib_chrome_so)
offsets = process_profiles.GetReachedOffsetsFromDumpFiles(
files, self._compiler.lib_chrome_so)
if not offsets:
raise Exception('No profiler offsets found in {}'.format(
'\n'.join(files)))
processor = process_profiles.SymbolOffsetProcessor(
self._compiler.lib_chrome_so)
ordered_symbols = processor.GetOrderedSymbols(offsets)
if not ordered_symbols:
raise Exception('No symbol names from offsets found in {}'.format(
'\n'.join(files)))
with open(self._GetUnpatchedOrderfileFilename(), 'w') as orderfile:
orderfile.write('\n'.join(ordered_symbols))
except Exception:
for f in files:
self._SaveForDebugging(f)
raise
finally:
self._profiler.Cleanup()
def _MaybeSaveProfile(self, files):
if self._options.profile_save_dir:
logging.info('Saving profiles to %s', self._options.profile_save_dir)
for f in files:
shutil.copy(f, self._options.profile_save_dir)
logging.info('Saved profile %s', f)
def _PatchOrderfile(self):
"""Patches the orderfile using clean version of libchrome.so."""
self._step_recorder.BeginStep('Patch Orderfile')
patch_orderfile.GeneratePatchedOrderfile(
self._GetUnpatchedOrderfileFilename(), self._compiler.lib_chrome_so,
self._GetPathToOrderfile(), self._order_outlined_functions)
def _VerifySymbolOrder(self):
self._step_recorder.BeginStep('Verify Symbol Order')
return_code = self._step_recorder.RunCommand([
self._CHECK_ORDERFILE_SCRIPT, self._compiler.lib_chrome_so,
self._GetPathToOrderfile()
],
constants.DIR_SOURCE_ROOT,
raise_on_error=False)
if return_code:
self._step_recorder.FailStep('Orderfile check returned %d.' % return_code)
def _RecordHash(self, file_name):
"""Records the hash of the file into the output_data dictionary."""
self._output_data[os.path.basename(file_name) + '.sha1'] = _GenerateHash(
file_name)
def _SaveFileLocally(self, file_name, file_sha1):
"""Saves the file to a temporary location and prints the sha1sum."""
if not os.path.exists(self._DIRECTORY_FOR_DEBUG_FILES):
os.makedirs(self._DIRECTORY_FOR_DEBUG_FILES)
shutil.copy(file_name, self._DIRECTORY_FOR_DEBUG_FILES)
print('File: %s, saved in: %s, sha1sum: %s' %
(file_name, self._DIRECTORY_FOR_DEBUG_FILES, file_sha1))
def _SaveForDebugging(self, filename):
"""Uploads the file to cloud storage or saves to a temporary location."""
file_sha1 = _GenerateHash(filename)
if not self._options.buildbot:
self._SaveFileLocally(filename, file_sha1)
else:
print('Uploading file for debugging: ' + filename)
self._orderfile_updater.UploadToCloudStorage(
filename, use_debug_location=True)
def _SaveForDebuggingWithOverwrite(self, file_name):
"""Uploads and overwrites the file in cloud storage or copies locally.
Should be used for large binaries like lib_chrome_so.
Args:
file_name: (str) File to upload.
"""
file_sha1 = _GenerateHash(file_name)
if not self._options.buildbot:
self._SaveFileLocally(file_name, file_sha1)
else:
print('Uploading file for debugging: %s, sha1sum: %s' % (file_name,
file_sha1))
upload_location = '%s/%s' % (
self._CLOUD_STORAGE_BUCKET_FOR_DEBUG, os.path.basename(file_name))
self._step_recorder.RunCommand([
'gsutil.py', 'cp', file_name, 'gs://' + upload_location])
print('Uploaded to: https://sandbox.google.com/storage/' +
upload_location)
def _MaybeArchiveOrderfile(self, filename):
"""In buildbot configuration, uploads the generated orderfile to
Google Cloud Storage.
Args:
filename: (str) Orderfile to upload.
"""
# First compute hashes so that we can download them later if we need to.
self._step_recorder.BeginStep('Compute hash for ' + filename)
self._RecordHash(filename)
if self._options.buildbot:
self._step_recorder.BeginStep('Archive ' + filename)
self._orderfile_updater.UploadToCloudStorage(
filename, use_debug_location=False)
def UploadReadyOrderfiles(self):
self._step_recorder.BeginStep('Upload Ready Orderfiles')
for file_name in [self._GetUnpatchedOrderfileFilename(),
self._GetPathToOrderfile()]:
self._orderfile_updater.UploadToCloudStorage(
file_name, use_debug_location=False)
def _NativeCodeMemoryBenchmark(self, apk):
"""Runs system_health.memory_mobile to assess native code memory footprint.
Args:
apk: (str) Path to the apk.
Returns:
results: ([int]) Values of native code memory footprint in bytes from the
benchmark results.
"""
self._step_recorder.BeginStep("Running orderfile.memory_mobile")
try:
out_dir = tempfile.mkdtemp()
self._profiler._RunCommand(['tools/perf/run_benchmark',
'--device={}'.format(
self._profiler._device.serial),
'--browser=exact',
'--output-format=csv',
'--output-dir={}'.format(out_dir),
'--reset-results',
'--browser-executable={}'.format(apk),
'orderfile.memory_mobile'])
out_file_path = os.path.join(out_dir, 'results.csv')
if not os.path.exists(out_file_path):
raise Exception('Results file not found!')
results = {}
with open(out_file_path, 'r') as f:
reader = csv.DictReader(f)
for row in reader:
if not row['name'].endswith('NativeCodeResidentMemory'):
continue
# Note: NativeCodeResidentMemory records a single sample from each
# story run, so this average (reported as 'avg') is exactly the value
# of that one sample. Each story is run multiple times, so this loop
# will accumulate into a list all values for all runs of each story.
results.setdefault(row['name'], {}).setdefault(
row['stories'], []).append(row['avg'])
if not results:
raise Exception('Could not find relevant results')
return results
except Exception as e:
return 'Error: ' + str(e)
finally:
shutil.rmtree(out_dir)
def _PerformanceBenchmark(self, apk):
"""Runs Speedometer2.0 to assess performance.
Args:
apk: (str) Path to the apk.
Returns:
results: ([float]) Speedometer2.0 results samples in milliseconds.
"""
self._step_recorder.BeginStep("Running Speedometer2.0.")
try:
out_dir = tempfile.mkdtemp()
self._profiler._RunCommand(['tools/perf/run_benchmark',
'--device={}'.format(
self._profiler._device.serial),
'--browser=exact',
'--output-format=histograms',
'--output-dir={}'.format(out_dir),
'--reset-results',
'--browser-executable={}'.format(apk),
'speedometer2'])
out_file_path = os.path.join(out_dir, 'histograms.json')
if not os.path.exists(out_file_path):
raise Exception('Results file not found!')
with open(out_file_path, 'r') as f:
results = json.load(f)
if not results:
raise Exception('Results file is empty.')
for el in results:
if 'name' in el and el['name'] == 'Total' and 'sampleValues' in el:
return el['sampleValues']
raise Exception('Unexpected results format.')
except Exception as e:
return 'Error: ' + str(e)
finally:
shutil.rmtree(out_dir)
def RunBenchmark(self, out_directory, no_orderfile=False):
"""Builds chrome apk and runs performance and memory benchmarks.
Builds a non-instrumented version of chrome.
Installs chrome apk on the device.
Runs Speedometer2.0 benchmark to assess performance.
Runs system_health.memory_mobile to evaluate memory footprint.
Args:
out_directory: (str) Path to out directory for this build.
no_orderfile: (bool) True if chrome to be built without orderfile.
Returns:
benchmark_results: (dict) Results extracted from benchmarks.
"""
benchmark_results = {}
try:
_UnstashOutputDirectory(out_directory)
self._compiler = ClankCompiler(out_directory, self._step_recorder,
self._options.arch, self._options.use_goma,
self._options.goma_dir,
self._options.use_remoteexec,
self._ninja_command,
self._options.system_health_orderfile,
self._monochrome, self._options.public,
self._GetPathToOrderfile())
if no_orderfile:
orderfile_path = self._GetPathToOrderfile()
backup_orderfile = orderfile_path + '.backup'
shutil.move(orderfile_path, backup_orderfile)
open(orderfile_path, 'w').close()
# Build APK to be installed on the device.
self._compiler.CompileChromeApk(instrumented=False,
use_call_graph=False,
force_relink=True)
benchmark_results['Speedometer2.0'] = self._PerformanceBenchmark(
self._compiler.chrome_apk)
benchmark_results['orderfile.memory_mobile'] = (
self._NativeCodeMemoryBenchmark(self._compiler.chrome_apk))
except Exception as e:
benchmark_results['Error'] = str(e)
finally:
if no_orderfile and os.path.exists(backup_orderfile):
shutil.move(backup_orderfile, orderfile_path)
_StashOutputDirectory(out_directory)
return benchmark_results
def Generate(self):
"""Generates and maybe upload an order."""
assert (bool(self._options.profile) ^
bool(self._options.manual_symbol_offsets))
if self._options.system_health_orderfile and not self._options.profile:
raise AssertionError('--system_health_orderfile must be not be used '
'with --skip-profile')
if (self._options.manual_symbol_offsets and
not self._options.system_health_orderfile):
raise AssertionError('--manual-symbol-offsets must be used with '
'--system_health_orderfile.')
if self._options.profile:
try:
_UnstashOutputDirectory(self._instrumented_out_dir)
self._compiler = ClankCompiler(
self._instrumented_out_dir, self._step_recorder, self._options.arch,
self._options.use_goma, self._options.goma_dir,
self._options.use_remoteexec, self._ninja_command,
self._options.system_health_orderfile, self._monochrome,
self._options.public, self._GetPathToOrderfile())
if not self._options.pregenerated_profiles:
# If there are pregenerated profiles, the instrumented build should
# not be changed to avoid invalidating the pregenerated profile
# offsets.
self._compiler.CompileChromeApk(instrumented=True,
use_call_graph=
self._options.use_call_graph)
self._GenerateAndProcessProfile()
self._MaybeArchiveOrderfile(self._GetUnpatchedOrderfileFilename())
finally:
_StashOutputDirectory(self._instrumented_out_dir)
elif self._options.manual_symbol_offsets:
assert self._options.manual_libname
assert self._options.manual_objdir
with open(self._options.manual_symbol_offsets) as f:
symbol_offsets = [int(x) for x in f]
processor = process_profiles.SymbolOffsetProcessor(
self._options.manual_libname)
generator = cyglog_to_orderfile.OffsetOrderfileGenerator(
processor, cyglog_to_orderfile.ObjectFileProcessor(
self._options.manual_objdir))
ordered_sections = generator.GetOrderedSections(symbol_offsets)
if not ordered_sections: # Either None or empty is a problem.
raise Exception('Failed to get ordered sections')
with open(self._GetUnpatchedOrderfileFilename(), 'w') as orderfile:
orderfile.write('\n'.join(ordered_sections))
if self._options.patch:
if self._options.profile:
self._RemoveBlanks(self._GetUnpatchedOrderfileFilename(),
self._GetPathToOrderfile())
try:
_UnstashOutputDirectory(self._uninstrumented_out_dir)
self._compiler = ClankCompiler(
self._uninstrumented_out_dir, self._step_recorder,
self._options.arch, self._options.use_goma, self._options.goma_dir,
self._options.use_remoteexec, self._ninja_command,
self._options.system_health_orderfile, self._monochrome,
self._options.public, self._GetPathToOrderfile())
self._compiler.CompileLibchrome(instrumented=False,
use_call_graph=False)
self._PatchOrderfile()
# Because identical code folding is a bit different with and without
# the orderfile build, we need to re-patch the orderfile with code
# folding as close to the final version as possible.
self._compiler.CompileLibchrome(instrumented=False,
use_call_graph=False, force_relink=True)
self._PatchOrderfile()
self._compiler.CompileLibchrome(instrumented=False,
use_call_graph=False, force_relink=True)
self._VerifySymbolOrder()
self._MaybeArchiveOrderfile(self._GetPathToOrderfile())
finally:
_StashOutputDirectory(self._uninstrumented_out_dir)
if self._options.benchmark:
self._output_data['orderfile_benchmark_results'] = self.RunBenchmark(
self._uninstrumented_out_dir)
self._output_data['no_orderfile_benchmark_results'] = self.RunBenchmark(
self._no_orderfile_out_dir, no_orderfile=True)
if self._options.buildbot:
self._orderfile_updater._GitStash()
self._step_recorder.EndStep()
return not self._step_recorder.ErrorRecorded()
def GetReportingData(self):
"""Get a dictionary of reporting data (timings, output hashes)"""
self._output_data['timings'] = self._step_recorder.timings
return self._output_data
def CommitStashedOrderfileHashes(self):
"""Commit any orderfile hash files in the current checkout.
Only possible if running on the buildbot.
Returns: true on success.
"""
if not self._options.buildbot:
logging.error('Trying to commit when not running on the buildbot')
return False
self._orderfile_updater._CommitStashedFiles([
filename + '.sha1'
for filename in (self._GetUnpatchedOrderfileFilename(),
self._GetPathToOrderfile())])
return True
def CreateArgumentParser():
"""Creates and returns the argument parser."""
parser = argparse.ArgumentParser()
parser.add_argument('--no-benchmark', action='store_false', dest='benchmark',
default=True, help='Disables running benchmarks.')
parser.add_argument(
'--buildbot', action='store_true',
help='If true, the script expects to be run on a buildbot')
parser.add_argument(
'--device', default=None, type=str,
help='Device serial number on which to run profiling.')
parser.add_argument(
'--verify', action='store_true',
help='If true, the script only verifies the current orderfile')
parser.add_argument('--target-arch',
action='store',
dest='arch',
default='arm',
choices=list(_ARCH_GN_ARGS.keys()),
help='The target architecture for which to build.')
parser.add_argument('--output-json', action='store', dest='json_file',
help='Location to save stats in json format')
parser.add_argument(
'--skip-profile', action='store_false', dest='profile', default=True,
help='Don\'t generate a profile on the device. Only patch from the '
'existing profile.')
parser.add_argument(
'--skip-patch', action='store_false', dest='patch', default=True,
help='Only generate the raw (unpatched) orderfile, don\'t patch it.')
parser.add_argument('--goma-dir', help='GOMA directory.')
parser.add_argument(
'--use-goma', action='store_true', help='Enable GOMA.', default=False)
parser.add_argument('--use-remoteexec',
action='store_true',
help='Enable remoteexec. see //build/toolchain/rbe.gni.',
default=False)
parser.add_argument('--ninja-path',
help='Path to the ninja binary. If given, use this'
'instead of autoninja.')
parser.add_argument('--ninja-j',
help='-j value passed to ninja.'
'pass -j to ninja. no need to set this when '
'--ninja-path is not specified.')
parser.add_argument('--adb-path', help='Path to the adb binary.')
parser.add_argument('--public',
action='store_true',
help='Build non-internal APK and change the orderfile '
'location. Required if your checkout is non-internal.',
default=False)
parser.add_argument('--nosystem-health-orderfile', action='store_false',
dest='system_health_orderfile', default=True,
help=('Create an orderfile based on an about:blank '
'startup benchmark instead of system health '
'benchmarks.'))
parser.add_argument(
'--use-legacy-chrome-apk', action='store_true', default=False,
help=('Compile and instrument chrome for [L, K] devices.'))
parser.add_argument('--manual-symbol-offsets', default=None, type=str,
help=('File of list of ordered symbol offsets generated '
'by manual profiling. Must set other --manual* '
'flags if this is used, and must --skip-profile.'))
parser.add_argument('--manual-libname', default=None, type=str,
help=('Library filename corresponding to '
'--manual-symbol-offsets.'))
parser.add_argument('--manual-objdir', default=None, type=str,
help=('Root of object file directory corresponding to '
'--manual-symbol-offsets.'))
parser.add_argument('--noorder-outlined-functions', action='store_true',
help='Disable outlined functions in the orderfile.')
parser.add_argument('--pregenerated-profiles', default=None, type=str,
help=('Pregenerated profiles to use instead of running '
'profile step. Cannot be used with '
'--skip-profiles.'))
parser.add_argument('--profile-save-dir', default=None, type=str,
help=('Directory to save any profiles created. These can '
'be used with --pregenerated-profiles. Cannot be '
'used with --skip-profiles.'))
parser.add_argument('--upload-ready-orderfiles', action='store_true',
help=('Skip orderfile generation and manually upload '
'orderfiles (both patched and unpatched) from '
'their normal location in the tree to the cloud '
'storage. DANGEROUS! USE WITH CARE!'))
parser.add_argument('--streamline-for-debugging', action='store_true',
help=('Streamline where possible the run for faster '
'iteration while debugging. The orderfile '
'generated will be valid and nontrivial, but '
'may not be based on a representative profile '
'or other such considerations. Use with caution.'))
parser.add_argument('--commit-hashes', action='store_true',
help=('Commit any orderfile hash files in the current '
'checkout; performs no other action'))
parser.add_argument('--use-call-graph', action='store_true', default=False,
help='Use call graph instrumentation.')
profile_android_startup.AddProfileCollectionArguments(parser)
return parser
def CreateOrderfile(options, orderfile_updater_class=None):
"""Creates an orderfile.
Args:
options: As returned from optparse.OptionParser.parse_args()
orderfile_updater_class: (OrderfileUpdater) subclass of OrderfileUpdater.
Returns:
True iff success.
"""
logging.basicConfig(level=logging.INFO)
devil_chromium.Initialize(adb_path=options.adb_path)
generator = OrderfileGenerator(options, orderfile_updater_class)
try:
if options.verify:
generator._VerifySymbolOrder()
elif options.commit_hashes:
return generator.CommitStashedOrderfileHashes()
elif options.upload_ready_orderfiles:
return generator.UploadReadyOrderfiles()
else:
return generator.Generate()
finally:
json_output = json.dumps(generator.GetReportingData(),
indent=2) + '\n'
if options.json_file:
with open(options.json_file, 'w') as f:
f.write(json_output)
print(json_output)
return False
def main():
parser = CreateArgumentParser()
options = parser.parse_args()
return 0 if CreateOrderfile(options) else 1
if __name__ == '__main__':
sys.exit(main())