Read more about flaky tests here.
- Implements the DeFlaker algorithm for pytest
- Implements two traceback-matching classifiers from Alshammari et al. (2024).
- Implements a novel coverage-independence classifier that classifies tests as flaky if they fail independently of passing test cases that exercise overlapping code.
- Optionally rerun or suppress flaky failures
- Output results to JSON, HTML, or JUnitXML
- Save test outcome history to a remote or local database
Flakefighters is a pytest plugin developed as part of the TestFLARE project. The plugin provides a "Swiss army knife" of techniques (called flakefighters) to detect flaky tests. Where existing flaky test plugins such as pytest-rerunfailures and pytest-flaky are primarily focused on rerunning (potentially) flaky tests until they pass, our main aim is to identify flaky tests by classifying test failures as genuine or flaky. The pytest-flakefinder plugin does this by simply rerunning tests multiple times and observing the result.
By contrast, Flakefighters incorporates several cutting edge flaky test detection techniques from research to automatically classify test failures as either genuine: indicating either a fault in the code or a mis-specified test case, or flaky: indicating a test with a nondeterministic outcome. Flaky tests are then reported separately in the test report, and can be optionally rerun or suppressed so they don't block CI/CD pipelines.
| Feature | pytest-flakefighters | pytest-rerunfailures | pytest-flaky | pytest-flakefinder | pytest-replay |
|---|---|---|---|---|---|
| Purpose | Classify test failures as genuine or flaky | Rerun failing tests in case they are flaky | Decorator-based reruns | Copy tests to observe nondeterministic outcomes | Reproduce flaky failures from CI when running with xdist |
| Detection Method | DeFlaker algorithm + coverage analysis | None | None | Reruns | None |
| Reporting | Terminal, HTML, JSON, JUnitXML | Terminal | Terminal | Terminal | Terminal |
| History Tracking | Database of test outcomes over commits | None | None | None | None |
| Rerun Option | Optional | Required | Required | Required | Required |
| Suppression Option | Optional | None | None | None | None |
| Debugging support | Insight into why tests are flaky | None | None | None | Reliable reproduction of flaky failures |
Use pytest-flakefighters when you want to:
- Understand WHY tests are flaky, not just hide the symptoms
- Classify flaky tests by root cause (coverage-independent, traceback-matched, etc.)
- Track test flakiness over time and across commits
- Make informed decisions about whether failures are legitimate
- pytest-rerunfailures: Quick fix for CI builds
- pytest-flaky: A few tests are known to be flaky
- pytest-flakefinder: Brute force search for flaky tests
- pytest-replay: Debugging specific flaky failures
Yes! pytest-flakefighters can be combined with other flaky test plugins:
- Use pytest-flakefighters to identify and classify flaky tests
- Use pytest-rerunfailures or pytest-flaky as a temporary measure while fixing them
- Use pytest-replay to debug specific instances identified by flakefighters
- Use pytest-xdist to randomise the order of your test cases
For more information on flaky test management best practices, see the pytest documentation.
You can install the extension by running pip install pytest-flakefighters from within your project's virtual environment.
If you use uv for Python package management, you can install pytest-flakefighters with uv add pytest-flakefighters.
This will add the plugin to your main dependencies.
dependencies = [
"pytest-flakefighters>=x.y.z",
]
However, pytest is typically a development dependency, and so should be added with uv add --dev pytest-flakefighters.
[dependency-groups]
dev = [
"pytest-flakefighters>=x.y.z",
]
You can install "pytest-flakefighters" by cloning this repo and running pip install . from the root directory.
If you intend to develop the plugin, run pip install -e .[dev] instead.
If you use uv, you can install pytest-flakefighters with:
# Install with uv
uv pip install .
# For development
uv pip install -e .[dev]FlakeFighter is intended to run on git repositories that have test suites runnable with pytest.
Once you have installed FlakeFighter, you can run it from the root directory of your repo simply by running pytest in your usual way.
FlakeFighter has the following arguments.
--target-commit=TARGET_COMMIT
The target (newer) commit hash. Defaults to HEAD (the most recent commit).
--source-commit=SOURCE_COMMIT
The source (older) commit hash. Defaults to HEAD^ (the previous commit to target).
--repo=REPO_ROOT The commit hash to compare against.
--suppress-flaky-failures-exit-code
Return OK exit code if the only failures are flaky failures.
--no-save Do not save this run to the database of previous flakefighters runs.
-M LOAD_MAX_RUNS, --load-max-runs=LOAD_MAX_RUNS
The maximum number of previous runs to consider.
-D DATABASE_URL, --database-url=DATABASE_URL
The database URL. Defaults to 'flakefighter.db' in current working directory.
--store-max-runs=STORE_MAX_RUNS
The maximum number of previous flakefighters runs to store. Default is to store all.
--time-immemorial=TIME_IMMEMORIAL
How long to store flakefighters runs for, specified as `days:hours:minutes`. E.g. to store
tests for one week, use 7:0:0.
By default, pytest-flakefighters runs whenever it is installed. To disable it for a specific test run, use:
pytest --no-flakefightersThis is useful when you have the plugin installed but want to run quick tests without flaky test detection.
You can also configure this in your pyproject.toml:
[tool.pytest.ini_options]
addopts = "--no-flakefighters"Contributions are very welcome. Tests can be run with pytest, please ensure the coverage at least stays the same before you submit a pull request.
Our plugin is made up of a collection of heuristics that come together to help inform whether a test failure is genuine or flaky.
These come in two "flavours": those which run live after each test, and those which run at the end of the entire test suite.
Both extend the base class FlakeFighter and implement the flaky_failure method, which returns True if the test is deemed to be flaky.
If you encounter any problems, please file an issue along with a detailed description.
This pytest plugin was generated with Cookiecutter along with @hackebrot's cookiecutter-pytest-plugin template.