breadcrumbs: Testing Home > page_name: test-suites title: Test Suites
This document describes how to write, update, run, disable, and enumerate tests that are in a Test Suite.
A test suite is a collection of tests, client and server side, that are run on a particular build to validate a larger system.
Test Suites build on the Autotest control file specification.
In order to create a valid suite the following two requirements need to be met:
We have a mechanism that allows suites of tests to be defined dynamically by a control file at runtime. [test_suites/control.bvt-inline](https://chromium.googlesource.com/chromiumos/third_party/autotest/+/master/test_suites/control.bvt-inline)
in the autotest repo is a simple example of such a suite, and login_LoginSuccess is one test that's included in this suite. Looking at the control file for login_LoginSuccess, note the line ATTRIBUTES = “suite:bvt-inline, suite:push_to_prod, suite:smoke”. This is used in control.bvt-inline
like this:
...
args_dict['name'] = 'bvt-inline'
dynamic_suite.reimage_and_run(**args_dict)
The infrastructure finds login_LoginSuccess by matching the contents of the suite tags in ATTRIBUTES field in its control file against the suite name passed into dynamic_suite.reimage_and_run()
If one test should be in multiple suites, the ATTRIBUTES field in the control file can be a comma delimited list of suite tags.
ATTRIBUTES = "suite:bvt-inline, suite:push_to_prod, suite:smoke"
For an end-to-end demonstration of the flow of creating a new suite, see Dynamic Suites Code Lab
This section will go over adding, removing, and marking tests as experimental.
Some suites have restrictions when it comes to adding tests to them. To read about the requirements of a particular test suite refer to its suite control under autotest/files/test_suites/control.suite_name.
For example: control.bvt-inline
Add the ‘suite:<your_suite_name>’ tag to the ATTRIBUTES field in the control file of the test you want to be run in the suite. If this test is new add it as an experimental test first, see below.
To disable a test remove the ‘suite:<your_suite_name>’ tag from the ATTRIBUTES field, if it is the only variable in the ATTRIBUTES field, remove the entire ATTRIBUTES line.
You may consider marking a test as experimental to avoid closing of the tree but continue running tests.
Some times you may want to add tests that are experimental to start gathering some data but you do not want the tree to be closed. To do this use the EXPERIMENTAL control file variable.
AUTHOR = "Autotest Team" NAME = "Sleeptest" TIME = "SHORT" TEST_CATEGORY = "Functional" TEST_CLASS = "General" TEST_TYPE = "client" ATTRIBUTES = "suite:bvt-inline" EXPERIMENTAL = "True"
An example of a test marked experimental.
These are options that you can use to tweak the behavior in your suite control file.
check_hosts
Boolean, if the hosts required are not currently in good health fail the suite. You may want to let the suite sit around for a while and wait for the hosts to be repaired. Default: True
add_experimental
If you only want the suite to include non-experimental tests. Default: True
file_bugs
Whether or not bugs should be filed when tests fail (See below for a more detailed example)
file_expermental_bugs
Whether or not bugs should be filed when experimental tests fail
max_runtime_mins
How long each job should be allowed to run in a suite. Default is 24 hours. (See below for an example)
suite_dependencies
If you want to apply a dependency across the whole suite (See below for a more detailed example)
version_prefix
What version_prefix to use. Currently we have two types of version_prefix CROS_VERSION_PREFIX and FW_VERSION_PREFIX, defined in autotest_lib.server.cros.provision
This tells our infrastructure how you would like to provision the allocateviarabled machines. CROS_VERSION_PREFIX tells the infrastructure to provision the machines to the correct OS version.
FW_VERSION_PREFIX tells the infrastructure to provision the machines to the correct firmware version.
deprecated option: skip_reimage
This option has been deprecated. If you specify a build string when creating the suite, our infrastructure will automatically provision the allocated machines to the requested build.
In addition to DEPENDENCIES specified on a per-test level, a set of suite dependencies can be specified at the suite level. These dependencies will be appended to the dependency list of all tests when they are run through that suite. To add suite-level dependencies, pass in your desired dependencies to dynamic_suite.reimage_and_run with the suite_dependencies argument. For example:
... args_dict['suite_dependencies'] = 'carrier:tmobile' dynamic_suite.reimage_and_run(**args_dict)
Multiple suite-level dependencies can be given as a comma separated string.
If your suite is already being run via suite_scheduler, you only need to edit the suite control file.
... _BUG_TEMPLATE = { 'labels': ['OS-Chrome','Type-Bug', 'your-label'], 'owner': '', 'status': None, 'summary': None, 'title': None, 'cc': ['a@chromium.org','b@chromium.org', 'c@chromium.org'] } args_dict['bug_template'] = _BUG_TEMPLATE dynamic_suite.reimage_and_run(**args_dict)
If you want to sign up (being owner or being cc-ed), or apply certain label to the bug when test failed, you only need to add an optional BUG_TEMPLATE dictionary into your test control file:
BUG_TEMPLATE = {
'labels': ['your-label'],
'owner': 'your_email@chromium.org',
'status': None,
'summary': None,
'title': None,
'cc': ['a@chromium.org','b@chromium.org', 'c@chromium.org']
}
The settings in this bug template will be merged to the bug template defined in suite control file. That is, labels and cc defined in both test and suite control files will be merged and applied to the bug filed.
In your suite control file pass the max_runtime_mins option:
An example to allow tests to run for 6 hours.
... args_dict['max_runtime_mins'] = 60*6 dynamic_suite.reimage_and_run(**args_dict)
There is a tool under the autotest repo to enumerate tests that will run in a suite.
From within the chroot the following command can be executed:
-a option is to specify what Autotest directory to look in.
The argument bvt-inline is the SUITE you are intersted in.
$~/trunk/src/third_party/autotest/files/site_utils/suite_enumerator.py -a ~/trunk/src/third_party/autotest/files -l
(chroot)$ test_that -b ${board} ${host} suite:bvt-inline
Note that if your suite contains multiple control files that are associated with the same autotest, you may get an error that says something like this:
TestError: <AUTOTEST_NAME> already exists; multiple tests cannot run with the same subdirectory
To fix this, add a “tag” parameter to the job.run_test() in the associated control files (e.g., job.run_test(‘foo’, tag=‘some_tag’)) to distinguish them and make the results get written into uniquely-named result directories.
Suite scheduler is the piece of software that is responsible for kicking off suites out of band based on a set of events as opposed to in a waterfall. If you are interested in running suites in a waterfall refer to the HW_Test section below. Below is a run down on how this works and what needs to happen to add new suites to this configuration.
Suites are scheduled based off of build artifacts, the autotest tarball generated at build time is used to enumerate what control files belong to what suites and what version of those particular tests should be run. If updated a test on TOT it will need to be back ported to every branch that needs those updates just like any other change.
If your suite is a part of bvt, kaen_bvt, or smoke it will automatically run and no further action must be taken.
An out of band process runs on the Autotest server to schedule suite runs based on a number of events. The code lives under autotest/files/site_utils/suite_scheduler and it's configuration (a Python config file) can be created in autotest/files/suite_scheduler.ini.
For the test lab, the config file is located in chromeos-admin repo as suite_scheduler.ini (sample CL).
To add your suite send a code review with a new entry as outlined below.
[WeeklyKernelRegressions] run_on: weekly suite: kernel_weekly_regression day: 0 branch_specs: >=tot-1 pool: suites boards: first_board,second_board
There are 7 parts to a configuration entry.
The top level unique name marker is part of the Python config format and must be unique and it should reflect what suite is being configured.
Pending implementation: new_chrome crbug.com/212678
*If there are other events you would like to have available please file a bug.
The suite line can only be used to describe one individual suite. The name used pertains to the suite added under autotest/files/test_suites/control.suite_name. The above example would invoke the control.kernel_weekly_regression suite file.
There are three different branch specs that can be used to ensure your suite runs on a particular branch:
A pool is a label on a machine like pool:bvt. We are using pools extensively currently as our dependency support is not quite there yet. Unless you are doing something out of the ordinary with extra hardware required to be attached to the DUT you should always use pool: suites which is a general pool for suite runs.
If you are doing things that require extra hardware chances are there is already a precedent for this in the suite_scheduler.ini file and you should reference those. When in doubt ask chromeos-lab-infrastructure@google.com.
Bug filing is optional, and defaults to False. If file_bugs is set to true in the ini section of the suite each test failure will either result in a new bug or a comment on an existing open bug, on the chromium bug tracker. Deduplication of bugs happens based on the test name, the suite name and the reason for failure. You can specify more bug filing options through the bug_template in the suite control file.
Suites can also be run as a step in the buildbot waterfall. This assumes you are running your builds via cbuildbot, if you are not please file a bug with your need and the Lab team can look at how to best direct you.
*Note when running the HW_Test step you are blocking your waterfall on actual hardware tests. If you do not want your builders to block on this consider using Suite Scheduler as described above.
*Note if your build client is not on the master2 vlan you will need to file a bug with the Chrome OS lab team so a proxy connection can be set up.
Adding a HW_Test step to your cbuildbot config is as easy as editing cbuildbot_config.py in the chromite repo and adding a hw_test entry like the following:
Example running the bvt suite for a lumpy build:
You need to specify upload_hw_test_artifacts so that Autotest has access to the image and the build artifacts via Google Storage and specify what suite(s) you want to run via a list called hw_tests.
For suite jobs:
$build-test_suites/control.$suite_name
i.e.
x86-alex-release/R20-2264.0.0-test_suites/control.kernel.per_build.benchmarks
For sub-jobs run from the suite:
$build/$suite_name/$test_name
i.e.
x86-alex-release/R20-2264.0.0/kernel.per_build.benchmarks/Hackbench
If a particular test is set with EXPERIMENTAL = True it is prefaced with experimental_test_name.
i.e.
x86-alex-release/R20-2264.0.0/kernel.per_build.benchmarks/experimental_platform_AesThroughput
*This is not a typical use case for users.
Suite scheduler lives under autotest/files/site_utils/suite_scheduler/suite_scheduler.py
An example of forcing all nightly suites for stumpy to be kicked off.
site_utils/suite_scheduler/suite_scheduler.py -i stumpy-release/R21-2270.0.0 -e nightly
If you need to run an individual suite for a particular build but not all of the suites that would be kicked off by a particular event you can use the same command the buildbots use, run_suite.py.
site_utils/run_suite.py -b parrot -p parrot_2gb -s kernel_weekly_regression -i parrot-release/R23-1913.156.1 -u 1
Other than the BVT suite no other suite results are displayed on the buildbot waterfall. To see your results you need to refer to the Wmatrix dashboard at https://wmatrix.googleplex.com
From there, select “Suite list” --> select your suite name, e.g. “bvt”. You will see a matrix where the column represents boards and the rows represents builds. Each cell shows the test results available for that (build, board).