Awesome pytest speedup
A checklist of best practices to speed up your pytest suite.
Tick them off, one by one:
- Hardware is fast
- Collection is fast
- PYTHONDONTWRITEBYTECODE=1 is set
- Built-in pytest plugins are disabled
- Only a subset of tests are executed
- Network access is disabled
- Disk access is disabled
- Database access is optimized
- Tests run in parallel
There's a recording of my talk going through some of the above at https://www.youtube.com/watch?v=uvkSOaFYsLo.
Finally, read the general guidelines before you start and later on go through the extra tips.
Let's start!
Measure first!
Before you do any of the changes below, remember the old measure twice, optimize once mantra. Instead of changing some pytest config and blindly believing it will help make your test suite faster, always measure.
In other words:
- Change one thing.
- Measure locally and on CI.
- Commit and push if there is a difference.
- Leave it running for a few days.
- Rinse & repeat.
For timing the entire test suite, hyperfine
is a fantastic tool.
For timing single tests, pytest --durations 10
will print out the ten slowest tests.
For measuring CPU usage and memory consumption, look at pytest-monitor
.
For detailed per-function-call profiling of tests, pytest-profiling
is a great start. Or you can try using the cProfile module directly.
Another popular profiler, pyinstrument
, provides examples of how to use it with pytest.
How do you run your tests?
Hardware
Modern laptops are incredibly fast. Why spend hours of time waiting for tests if you can throw money at the problem? Get a faster machine!
Using a CI that gets very expensive as you increase CPU cores? Most of them allow you to use self-hosted runners for cheap. Buy a few Mac Minis and have them run your tests, using all of their modern fast cores.
Collection
Collection is the first part of running a test suite: finding all tests that need to be run. It has to be super fast because it happens even if you run a single test.
Use the -βcollect-only
flag to see how fast (or slow) the collection step is. On a modern laptop, the collection should be take around 1s per 1000 tests, maybe up to 2s or 3s per 1000 tests for large codebases.
If it is slower, you can try the following:
-
Tell pytest not to look into certain folders:
# pytest.ini [pytest] norecursedirs = docs *.egg-info .git .tox var/large_ml_model/
-
Tell pytest exactly where the tests are so it doesn't look anywhere else:
# pytest.ini [pytest] testpaths = tests
or
pytest src/my_app/tests
-
Maybe collection is slow because of some code in
conftest.py
? Try runningpytest --collect-only --noconftest
to see if there is a difference. Much faster? Then itβs something inconftest.py
that is causing the slowdown. -
Maybe imports are making collection slow. Try this:
* python -X importtime -m pytest
- Paste the output into https://kmichel.github.io/python-importtime-graph/.
- See who the big offenders are.
- Try moving top-level imports into function imports.
- This is especially likely to be the culprit if your code imports large libraries such as pytorch, opencv, Django, Plone, etc.
PYTHONDONTWRITEBYTECODE
As per The Hitchhiker's Guide to Python there is no need to generate bytecode files on development machines.
On CIs it makes even less sense to do so. And can potentially even slow down the execution of your test suite.
Most people won't benefit a ton from this, but it's an easy thing to try: add export PYTHONDONTWRITEBYTECODE=1
to your ~/.profile
and your CI configuration.
Builtin plugins
Did you know that pytest comes with over 30 builtin plugins? You probably donβt need all of them.
- List them with
pytest --trace-config
- Disable with
pytest -p no:doctest
There is not much speedup to gain here, so I usually only disable the legacy ones: -p no:pastebin -p no:nose -p no:doctest
Be picky!
Hot take: you don't always have to run all tests:
pytest-skip-slow
Skip known slow tests by adding the@pytest.mark.slow
decorator. Do it in local dev and potentially in CI branch runs. Run all tests in the main CI run with--slow
.pytest-incremental
analyses your codebase and file modifications between test runs to decide which tests need to be run. Useful for local development, but also in CI: only run the full suite after diff-suite is green and save some CPU minutes.pytest-testmon
does the same, but using a smarter algorithm that includes looking into coverage report to decide which tests should run for a certain line change.
How do you write your tests?
Network access
Unit tests rarely need to access the Internet. Any network traffic is usually mocked to be able to test various responses and to ensure tests are fast by not having to wait for responses.
However, often we don't even realize that the code being tested is using the network. Maybe someone added support for loading profile pics from Gravatar and now a bunch of tests are pinging the Gravatar API under the hood.
pytest-socket
is a great plugin to prevent inadvertent internet access. Straightforward to use and with a bunch of escape hatches for those rare cases when you do in fact need network access in your test code.
Disk access
Unit tests should ideally not rely on the current filesystem to run successfully. If they do, they are more error prone as the filesystem can change, and are also slower if they write stuff to the disk.
One alternative is mocking disk I/O calls.
Another is a temporary in-memory filesystem, provided by pyfakefs
, making your tests both safer and faster.
Database access
Testing web apps usually requires some database access in tests, and that's OK. There are still several optimizations available.
Do all tests need a database?
Some tests could potentially be rewritten in a way that does not require a database to do the testing. For example, if calculating a user's age, there is no need to fetch a birth date from a DB. It's better to provide a fake birth date to the calculation function and remove the test's dependency on the database fixture.
Do all tests need the entire database?
Populating the database with test data takes time. Do all tests need the entire dummy dataset? If a group of tests is testing some user profile feature, they potentially don't need to populate non-relevant tables in the database.
Only prepare the database once
Preparing, using, and then destroying the dummy database for every test is wasteful. There are better approaches.
Truncate
Instead of destroying the database after a test run, rather TRUNCATE
its tables. This saves you from recreating the database for every test that relies on it.
- Create your test database once.
- Populate it with data for the test.
- Execute a test over it.
TRUNCATE
your tables, i.e. fast remove all data.- Populate with data for the next test, run the test, rinse & repeat.
Rollback
But data population is slow too! Can we save/cache that as well? We sure can! Instead of TRUNCATE
-ing all tables, we could just rollback the transaction the test used. As a result, no data needs to be prepared for the next test and it can just be run.
- Create your test database once.
- Populate it with dummy data once.
- Execute a test over it, making sure the test does not commit anything.
- Rollback the transaction.
Note that this approach requires you to be a bit more careful when writing tests, as they shouldn't do any database commits. If they do, you need to manually revert their changes.
Parallelization
By default, pytest uses a single CPU core. Your laptop likely has multiple cores. CI runners also come with multiple cores. It's just a waste of everyone's time not to use them all!
pytest-xdist
The most popular tool to help you use all cores is pytest-xdist
. It supports running across multiple cores and CPUs, even on remote machines!
It usually doesn't work out-of-the-box in a real-world project with complex fixtures, databases involved, etc. The main reason is that session
-scoped fixtures run on all workers, not just once. There are a couple of workarounds, but they are not trivial.
For example, you can create a separate database for each pytext-xdist
worker process and use the worker name/number as a suffix. pytest-django
does this by default.
pytest-split
Compared to pytest-xdist
, pytest-split
is easier to use. It does not really help with local development, but it can greatly decrease the speed of your CI runs without much or any changes to your tests.
The way pytest-split
works is that it splits the test suite to equally sized sub-suites based on test execution time. These sub-suites can be run in parallel in as many CI workers as your budget allows.
Caveats:
- To ensure 100% test coverage, you need to run an additional CI runner after all test runners have finished that collects and merges coverage reports.
pytest-randomly
seed needs to be pre-computed and fed into all CI runners.
Extra tips
Keep 'em fast!
It is really annoying to invest time into speeding up your test suite, only to come back to the codebase a few months later to discover that the tests are slow again. No more! Check out BlueRacer.io, a simple GitHub App that blocks a Pull Request from merging if tests have become slower: https://github.com/apps/blueracer-io.
pytest --lf
pytest --lf
, or pytest --last-failed
, tells pytest to only run tests that have failed during the last run. Handy in local development to decrease iteration time.
Handy pytest plugins
There are some pytest plugins that I tend to use in every project. I've listed them on https://niteo.co/blog/indispensable-pytest-plugins.
config.scan()
If you are using Pyramid's config.scan()
, then it's a potential bottleneck in large codebases. You could speed it up by telling it to ignore folders with tests.
config.scan(ignore=[".tests"])
config.scan("myapp", ignore=re.compile("tests?$").search)