tree: e1bc9cdaec7d8a63cdd4608633880bddfda5ef37
  1. animations/
  2. application/
  3. assertion/
  4. changes/
  5. console/
  6. coverage/
  7. cross_tool_integration/
  8. css_overview/
  9. elements/
  10. emulation/
  11. extensions/
  12. helpers/
  13. host/
  14. inline_editor/
  15. issues/
  16. layers/
  17. lighthouse/
  18. media/
  19. memory/
  20. network/
  21. performance/
  22. puppeteer/
  23. quick_open/
  24. recorder/
  25. rendering/
  26. resources/
  27. search/
  28. security/
  29. sensors/
  30. settings/
  31. snippets/
  32. sources/
  33. targets/
  34. webaudio/
  35. .mocharc.js
  36. BUILD.gn
  37. mocharc.ts
  38. README.md
  39. test-runner-config.json
  40. tsconfig.json
test/e2e/README.md

Guide on end-to-end testing

This directory hosts the end-to-end tests we run on DevTools. These tests open a target page and a DevTools frontend page, for which the DevTools frontend connects to the target page over CDP. We use Puppeteer to talk over CDP and all functionality of Puppeteer is available to you as well when writing end-to-end tests. We use Mocha as testing framework.

The goal of these end-to-end tests is to implement core user journeys throughout the application. As such, the tests you write should read like a little story that you can read, even if you don't know how it is implemented.

The tests therefore have a dual purpose:

  1. Verify that core user stories are working as intended and are not broken by a particular DevTools frontend change.
  2. Serve as documentation and reference point for how DevTools is intended to be used.

Skipping tests

You can disable a test for all platforms using it.skip. If you are disabling a flaky test, consider disabling it only on the affected platforms. For example,

it.skipOnPlatforms(['mac', 'win32'], '[crbug.com/xxx] ...', () => {...});
it.skipOnPlatforms(['linux'], '[crbug.com/xxx] ...', () => {...});

To use skipOnPlatforms, you need to import it from test/shared/mocha-extensions.ts.

Debugging flaky tests

The it.repeat helper is useful for reproducing a flaky test failure. e.g.

it.repeat(20, 'find element', async () => {...});

it.repeat behaves like it.only in that it will cause just that single test to be run.

Debugging flaky tests

To see if certain tests are flaky you can use the E2E stressor bots. Open a CL with your test changes and run the following command specifying your test file:

git cl try -B devtools-frontend/try -b e2e_stressor_linux -b e2e_stressor_win64 -b e2e_stressor_mac -p e2e_env='{"TEST_PATTERNS":"network/network-datagrid_test.ts","ITERATIONS":20}'

or multiple test files:

git cl try -B devtools-frontend/try -b e2e_stressor_linux -b e2e_stressor_win64 -b e2e_stressor_mac -p e2e_env='{"TEST_PATTERNS":"network/network-datagrid_test.ts,network/network_test.ts","ITERATIONS":20}'

This will run the specified tests on dedicated bots with the specified number of iterations. Note that in order for iterations to work the test should be using it from mocha_extensions.ts.

The following command runs the stressor bot on all files of the latest commit with reasonable settings:

git cl try -B devtools-frontend/try \
  -b e2e_stressor_linux \
  -p e2e_env="{ \
    \"TEST_PATTERNS\": \"$(git diff-tree --no-commit-id --name-only HEAD -r | grep test/e2e | cut -c 10- - | tr "\n" ",")_\", \
    \"STRESS\": true, \
    \"LATE_PROMISES\": true, \
    \"ITERATIONS\": 20 \
  }"

By default, tests are run using the debug build. To run it with the release build, append -p builder_config=Release to the end of the command.

Please use a reasonable number of iterations and include the minimal amount of test files to avoid overloading the bots. This bot is experimental and the parameters might change in the future.

General implementation details

To that end, the “what” from the “how” are separate in end-to-end tests. The “what” is the actual end-to-end test. The “how” are functions in helpers that implement the interaction with DevTools to perform a particular action.

For example, an end-to-end test might read like this:

it("can show newly created snippets show up in command menu", async () => {
  await openSourcesPanel();
  await openSnippetsSubPane();
  await createNewSnippet("New snippet");

  await openCommandMenu();
  await showSnippetsAutocompletion();

  assert.deepEqual(await getAvailableSnippets(), ["New snippet\u200B"]);
});

The test describes the user journey and then describes what actions the user takes to fulfill that journey. For example, when the user wants to “open the command menu”, that is an action performed. The implementation is separated in a helper that implements “how” the user performs that action.

The separation of the two concepts allows us to change the underlying implementation of the action while making sure that the user journey remains intact.

For example, this is the implementation of openCommandMenu():

export const openCommandMenu = async () => {
  const { frontend } = getBrowserAndPages();

  switch (platform) {
    case "mac":
      await frontend.keyboard.down("Meta");
      await frontend.keyboard.down("Shift");
      break;

    case "linux":
    case "win32":
      await frontend.keyboard.down("Control");
      await frontend.keyboard.down("Shift");
      break;
  }

  await frontend.keyboard.press("P");

  switch (platform) {
    case "mac":
      await frontend.keyboard.up("Meta");
      await frontend.keyboard.up("Shift");
      break;

    case "linux":
    case "win32":
      await frontend.keyboard.up("Control");
      await frontend.keyboard.up("Shift");
      break;
  }

  await waitFor(QUICK_OPEN_SELECTOR);
};

As you can see, the way the user opens the command menu is via key-bindings. We don‘t “bypass” the functionality by calling functions on components or on our models directly; we instruct Puppeteer to do exactly what a user would do. Doing so, we are certain that we don’t test our component abstractions and potentially lose track of integration issues.

Secondly, the function has a waitFor, which waits for the command menu (found by the QUICK_OPEN_SELECTOR) to appear. For every action that is performed in DevTools, there must be a corresponding user-visible change in the UI. This means that you always have to wait for something to happen and you can't assume that, as soon as you have performed an action, the UI has updated accordingly.

Note: Because of the async rendering of DevTools, content might not be strictly visible when DOM Nodes are appended to the DOM. As such, be aware of the functionality you are testing and relying on, as it could render differently than you originally assumed.

To summarize:

  1. Separate the “what” from the “how”.
  2. Use real actions (clicking, using key-bindings, typing) instead of “bypassing” via components/models.
  3. Every action must be observed by a change in the UI and must be waited for.
  4. Be aware of the async rendering of DevTools

Helpers

There are two kinds of helpers:

  1. Helpers written as part of the end-to-end test, implementing the “how” of user actions.
  2. Helpers supplied to interact with a page and abstract away the way the tests are run.

The former are implemented in helpers, written by the DevTools maintainers and are specific to the implementation of the DevTools frontend. The latter are implemented in ../shared, written by the Puppeteer maintainers and are predominantly DevTools-agnostic (apart from the handling of a target page + CDP connection).

In general, the e2e/helpers make use of the shared helpers. See ../shared/README.md for more documentation on the shared helpers.