| # Copyright (c) 2012 The Chromium Authors. All rights reserved. |
| # Use of this source code is governed by a BSD-style license that can be |
| # found in the LICENSE file. |
| |
| |
| import logging |
| import re |
| import os |
| import pexpect |
| |
| from perf_tests_helper import PrintPerfResult |
| from test_result import BaseTestResult, TestResults |
| from valgrind_tools import CreateTool |
| |
| |
| # TODO(bulach): TestPackage, TestPackageExecutable and |
| # TestPackageApk are a work in progress related to making the native tests |
| # run as a NDK-app from an APK rather than a stand-alone executable. |
| class TestPackage(object): |
| """A helper base class for both APK and stand-alone executables. |
| |
| Args: |
| adb: ADB interface the tests are using. |
| device: Device to run the tests. |
| test_suite: A specific test suite to run, empty to run all. |
| timeout: Timeout for each test. |
| rebaseline: Whether or not to run tests in isolation and update the filter. |
| performance_test: Whether or not performance test(s). |
| cleanup_test_files: Whether or not to cleanup test files on device. |
| tool: Name of the Valgrind tool. |
| dump_debug_info: A debug_info object. |
| """ |
| |
| def __init__(self, adb, device, test_suite, timeout, rebaseline, |
| performance_test, cleanup_test_files, tool, dump_debug_info): |
| self.adb = adb |
| self.device = device |
| self.test_suite_full = test_suite |
| self.test_suite = os.path.splitext(test_suite)[0] |
| self.test_suite_basename = os.path.basename(self.test_suite) |
| self.test_suite_dirname = os.path.dirname(self.test_suite) |
| self.rebaseline = rebaseline |
| self.performance_test = performance_test |
| self.cleanup_test_files = cleanup_test_files |
| self.tool = CreateTool(tool, self.adb) |
| if timeout == 0: |
| if self.test_suite_basename == 'page_cycler_tests': |
| timeout = 900 |
| else: |
| timeout = 60 |
| # On a VM (e.g. chromium buildbots), this timeout is way too small. |
| if os.environ.get('BUILDBOT_SLAVENAME'): |
| timeout = timeout * 2 |
| self.timeout = timeout * self.tool.GetTimeoutScale() |
| self.dump_debug_info = dump_debug_info |
| |
| def _BeginGetIOStats(self): |
| """Gets I/O statistics before running test. |
| |
| Return: |
| Tuple of (I/O stats object, flag of ready to continue). When encountering |
| error, ready-to-continue flag is False, True otherwise. The I/O stats |
| object may be None if the test is not performance test. |
| """ |
| initial_io_stats = None |
| # Try to get the disk I/O statistics for all performance tests. |
| if self.performance_test and not self.rebaseline: |
| initial_io_stats = self.adb.GetIoStats() |
| # Get rid of the noise introduced by launching Chrome for page cycler. |
| if self.test_suite_basename == 'page_cycler_tests': |
| try: |
| chrome_launch_done_re = re.compile( |
| re.escape('Finish waiting for browser launch!')) |
| self.adb.WaitForLogMatch(chrome_launch_done_re) |
| initial_io_stats = self.adb.GetIoStats() |
| except pexpect.TIMEOUT: |
| logging.error('Test terminated because Chrome launcher has no' |
| 'response after 120 second.') |
| return (None, False) |
| finally: |
| if self.dump_debug_info: |
| self.dump_debug_info.TakeScreenshot('_Launch_Chrome_') |
| return (initial_io_stats, True) |
| |
| def _EndGetIOStats(self, initial_io_stats): |
| """Gets I/O statistics after running test and calcuate the I/O delta. |
| |
| Args: |
| initial_io_stats: I/O stats object got from _BeginGetIOStats. |
| |
| Return: |
| String for formated diso I/O statistics. |
| """ |
| disk_io = '' |
| if self.performance_test and initial_io_stats: |
| final_io_stats = self.adb.GetIoStats() |
| for stat in final_io_stats: |
| disk_io += '\n' + PrintPerfResult(stat, stat, |
| [final_io_stats[stat] - |
| initial_io_stats[stat]], |
| stat.split('_')[1], True, False) |
| logging.info(disk_io) |
| return disk_io |
| |
| def GetDisabledPrefixes(self): |
| return ['DISABLED_', 'FLAKY_', 'FAILS_'] |
| |
| def _ParseGTestListTests(self, all_tests): |
| ret = [] |
| current = '' |
| disabled_prefixes = self.GetDisabledPrefixes() |
| for test in all_tests: |
| if not test: |
| continue |
| if test[0] != ' ': |
| current = test |
| continue |
| if 'YOU HAVE' in test: |
| break |
| test_name = test[2:] |
| if not any([test_name.startswith(x) for x in disabled_prefixes]): |
| ret += [current + test_name] |
| return ret |
| |
| def _WatchTestOutput(self, p): |
| """Watches the test output. |
| Args: |
| p: the process generating output as created by pexpect.spawn. |
| """ |
| ok_tests = [] |
| failed_tests = [] |
| timed_out = False |
| overall_fail = False |
| re_run = re.compile('\[ RUN \] ?(.*)\r\n') |
| re_fail = re.compile('\[ FAILED \] ?(.*)\r\n') |
| re_runner_fail = re.compile('\[ RUNNER_FAILED \] ?(.*)\r\n') |
| re_ok = re.compile('\[ OK \] ?(.*)\r\n') |
| (io_stats_before, ready_to_continue) = self._BeginGetIOStats() |
| while ready_to_continue: |
| found = p.expect([re_run, pexpect.EOF, re_runner_fail], |
| timeout=self.timeout) |
| if found == 1: # matched pexpect.EOF |
| break |
| if found == 2: # RUNNER_FAILED |
| logging.error('RUNNER_FAILED') |
| overall_fail = True |
| break |
| if self.dump_debug_info: |
| self.dump_debug_info.TakeScreenshot('_Test_Start_Run_') |
| full_test_name = p.match.group(1) |
| found = p.expect([re_ok, re_fail, pexpect.EOF, pexpect.TIMEOUT], |
| timeout=self.timeout) |
| if found == 0: # re_ok |
| ok_tests += [BaseTestResult(full_test_name.replace('\r', ''), |
| p.before)] |
| continue |
| failed_tests += [BaseTestResult(full_test_name.replace('\r', ''), |
| p.before)] |
| if found >= 2: |
| # The test crashed / bailed out (i.e., didn't print OK or FAIL). |
| if found == 3: # pexpect.TIMEOUT |
| logging.error('Test terminated after %d second timeout.', |
| self.timeout) |
| timed_out = True |
| break |
| p.close() |
| if not self.rebaseline and ready_to_continue: |
| ok_tests += self._EndGetIOStats(io_stats_before) |
| ret_code = self._GetGTestReturnCode() |
| if ret_code: |
| failed_tests += [BaseTestResult('gtest exit code: %d' % ret_code, |
| 'pexpect.before: %s' |
| '\npexpect.after: %s' |
| % (p.before, |
| p.after))] |
| return TestResults.FromOkAndFailed(ok_tests, failed_tests, |
| timed_out, overall_fail) |