Skip to content

Tests Waterfall & Telemetry

Tests Waterfall showing parallel test execution across workers with system telemetry overlay

Certain test runners expose information about test execution details, like which parallel worker executed the test, when it started and how long did it take. In addition, certain reporters embed system telemetry to overlay the test execution.

The lower section shows one horizontal lane per parallel worker (thread/process). Each test appears as a colored bar:

ColorMeaning
GreenTest passed
RedTest failed
GrayTest was skipped

The bar’s width represents the test’s duration. Gaps between bars show idle time on that worker. This makes it easy to spot:

  • Uneven distribution — one worker doing most of the work while others sit idle
  • Long-running tests — single tests that dominate the timeline
  • Bottlenecks — workers waiting because a slow test is blocking

Hover over a test bar to see its name, outcome, and retry number. Click a test bar to navigate to that test’s detailed results; other tests are dimmed to highlight the selection.

When available, the upper section overlays CPU and memory usage synchronized with the test execution timeline, helping you correlate test failures with resource pressure.

  • CPU — average load across all CPU cores, rendered with an orange-to-red gradient fill
  • Memory — rendered with a blue-to-purple gradient fill

Hover over any point to see the exact CPU percentage and memory usage (both as a percentage and in absolute bytes, e.g. “2.4 GB (54%)”) at that moment in the test run.

Telemetry depends on the test runner reporter. For example, @flakiness/playwright captures CPU and memory telemetry automatically, while @flakiness/pytest-flakiness does not. If no telemetry data is present in the report, the telemetry section is simply not shown.

When tests fail, the Test Report shows error details with full context. Each failed test shows its errors inline. An error includes:

  • Error message — the assertion or exception message
  • Stack trace — the full call stack with ANSI colors preserved from the terminal
  • Code snippet — when the reporter embeds source files, the viewer shows the relevant lines of code around the failure location

Test Report includes an Errors tab that aggregates all unique errors across the report into a summary table.

The table sorts by impacted test count (most common errors first), making it easy to identify systemic issues — like a shared service being down or a common assertion pattern failing across many tests.

Clicking an error filters the report to show only tests with that specific error.

Use FQL to search for tests by error text:

$timeout # tests with "timeout" in the error
$undefined # tests with "undefined" in the error
$"connection refused" f:api # connection errors in API test files

A single test can produce multiple errors — for example, when using soft assertions that continue execution after a failure. Test Report displays all errors for the test, not just the first one.

Infrastructure errors that occur outside of a specific test (e.g., during setup or teardown) are shown separately at the Test Report header.