You're welcome to come and procrastinate with us :)
$ source ./dev-env
Of course, feel free to read the script before launching it.
This script is intended to be a one-liner that sets up everything you need. It makes the following assumptions:
- You're using
MacOS
orLinux
, andbash
orzsh
. - You already have
python3
available - Either:
- you already have
poetry
,pre-commit
andnox
installed - or you have
pipx
installed and you're ok installing those 3 tools withpipx
- or you don't have
pipx
installed but it's ok if we install it for you
- you already have
- Either:
- you've already setup a PostgreSQL database and environment variables (
PG*
) are set - or you have
docker compose
available and port 5432 is free.
- you've already setup a PostgreSQL database and environment variables (
- Either:
psql
and otherlibpq
executables are available in thePATH
- or they are located in
usr/local/opt/libpq/bin
(Homebrew
).
The dev-env
script will add the scripts
folder to your $PATH
for the current
shell, so in the following documentation, if you see scripts/foo
, you're welcome
to call foo
directly.
Alternatively, you can utilize our development container setup. In VSCode, select
Dev Containers: Reopen in Container
from the command palette. This action sets up a
container preconfigured with all required dependencies and automatically provisions a
database. The virtual environment is created and activated seamlessly within the container.
The export
command below will be necessary whenever you want to interact
with the database (using the project locally, launching tests, ...). These are
standard libpq environment variables environment variables, and the values
used below correspond to the Docker setup. Feel free to adjust them as
necessary (The dev-env
script will set them for you).
$ export PGDATABASE=procrastinate PGHOST=localhost PGUSER=postgres PGPASSWORD=password
The development database can be launched using Docker with a single command. The PostgreSQL database we used is a fresh standard out-of-the-box database on the latest stable version.
$ docker compose up -d postgres
If you want to try out the project locally, it's useful to have postgresql-client
installed. It will give you both a PostgreSQL console (psql
) and specialized
commands like createdb
we use below.
$ # Ubuntu
$ sudo apt install postgresql-client
$ createdb
$ # MacOS
$ brew install libpq
$ /usr/local/opt/libpq/bin/createdb
The development environment is managed by poetry. It's a tool that manages dependencies and virtual environments. We also use pre-commit to keep the code clean.
If you don't already have poetry
or pre-commit
installed, you can
install them with:
$ scripts/bootstrap
This will install pipx if necessary and use it to install poetry
and
pre-commit
.
Then, install Procrastinate with development dependencies in a virtual environment:
$ poetry env use 3.{x} # Select the Python version you want to use (replace {x})
$ poetry install
$ poetry shell # Activate the virtual environment
You can check that your Python environment is properly activated:
(venv) $ which python
/path/to/current/folder/.venv/bin/python
With a running database:
(venv) $ pytest # Test the code with the current interpreter
If you're not familiar with Pytest, do yourself a treat and look into this fabulous tool.
To look at coverage in the browser after launching the tests, use:
$ scripts/htmlcov
This project uses pre-commit to keep the code clean. It's a tool that runs automated checks on your code before you commit it. Install the pre-commit hooks with:
$ pre-commit install
This will keep you from creating a commit if there's a linting problem.
In addition, an editorconfig file will help your favorite editor to respect procrastinate coding style. It is automatically used by most famous IDEs, such as Pycharm and VS Code.
The documentation is written in Markdown
and built with Sphinx
and MyST
.
Docstrings are written in reStructuredText
(because writing them in MyST
is
still a bit complicated, it seems).
Build with:
$ scripts/docs # build the html doc
$ scripts/htmldoc # browse the doc in you browser
If Sphinx's console output is localized and you would rather have it in English,
(which make google-based debugging much easier), use the environment variable
export LC_ALL=C.utf-8
If you make changes to the database structure (procrastinate/sql/schema.sql
) you
also need to create a corresponding migration script in the
procrastinate/sql/migrations
directory.
For example, let's say you want to add a column named extra
to the
procrastinate_jobs
table. You will first edit procrastinate/sql/schema.sql
and
change the definition of the table to add that column. That would be sufficient for new
Procrastinate users, but existing users, whose database already includes Procrastinate
objects (tables, indexes, ...), need to be able to migrate their existing schema into
the new one. For that reason, as a Procrastinate developer, you'll also need to create
a migration script, whose content would look like this:
-- add a column extra to the procrastinate_jobs table
ALTER TABLE procrastinate_jobs ADD COLUMN extra TEXT;
The name of migration scripts must follow a specific pattern:
xx.yy.zz_ab_{pre|post}_very_short_description_of_your_changes.sql
xx.yy.zz
is the number of the latest released version of Procrastinate. (The
latest release is the one marked Latest release
on the Procrastinate
releases page.) xx
, yy
and zz
must be 2-digit numbers, with leading
zeros if necessary. ab
is the 2-digit migration script's serial number, the
first number for each release being 01
for pre-migrations and 50
for
post-migrations. pre
is if the migration should be applied before upgrading
the code, post
is if the migration should be applied after upgrading the
code. And, finally, very_short_description_of_your_changes
is a very short
description of the changes (wow). It is important to use underscores between
the different parts, and between words in the short description.
For example, let's say the latest released version of Procrastinate is 1.0.1
,
that the migrations
directory already includes a post-migration script whose
serial number for that release number and your migration should be
applied after deploying the corresponding python code. In that case, if you
need to add a migration script, its name will start with 01.00.01_51_post_
.
As a Procrastinate developer, you must ensure you use pre-migrations and post-migrations to maintain backward compatibility with previous versions of Procrastinate.
For example, let's say that the current Procrastinate database schema includes an SQL function
procrastinate_func(arg1 integer, arg2 text, arg3 timestamp)
that you want to change to
procrastinate_func(arg1 integer, arg2 text)
The straightforward way to do that would be to edit the schema.sql
file and just
replace the old function by the new one, and add a migration script that removes the old
function and adds the new one:
DROP FUNCTION procrastinate_func_v3(integer, text, timestamp);
CREATE FUNCTION procrastinate_func_v3(arg1 integer, arg2 text)
RETURNS INT
...
But if you do that you will break the Procrastinate Python code that uses the old
version of the procrastinate_func
function. The direct consequence of that is
that Procrastinate users won't be able to upgrade Procrastinate without incurring
a service outage.
So when you make changes to the Procrastinate database schema you must ensure that the new schema still works with old versions of the Procrastinate Python code.
Going back to our procrastinate_func
example. Instead of replacing the old function
by the new one in schema.sql
, you add a new function in pre-migrations and remove the
old function in post-migrations:
-- xx_xx_xx_01_pre_add_new_version_procrastinate_func.sql
CREATE FUNCTION procrastinate_func_v4(arg1 integer, arg2 text)
RETURNS INT
...
-- xx_xx_xx_50_post_remove_old_version_procrastinate_func.sql
DROP FUNCTION procrastinate_func(integer, text, timestamp);
...
The continuous integration contains tests that will check that the schema and the migrations succeed in producing the same database structure. The migration tests are included in the normal test suite, but you can run them specifically with:
(venv) $ pytest tests/migration
We run the acceptance
tests on 3 different configurations:
- Without the post-migrations applied and with the last released version of Procrastinate
- Without the post-migrations applied and with the current checked out code
- With all migrations applied and with the current checked out code (this is just part of the normal test suite)
This is to ensure that the migrations are backward-compatible and that the database schema can be upgraded without downtime. We simulate all stages of the upgrade process:
- (the initial situation being that Procrastinate is running with the last released version of the code and all migrations of the last released version have been applied)
- First, the user would apply pre-migrations while the old version of the code is still running.
- Then, the user would upgrade the code to the new version.
- Finally, the user would apply post-migrations.
There are cases where new acceptance tests cannot work on the last released version.
In that case, the tests can be skipped by adding @pytest.mark.skip_before_version("x.y.z")
,
where x.y.z
is the version of Procrastinate where the test would start running.
See the demos page for instructions on how to run the demos ({doc}demos
).
In the development setup described above, Procrastinate, its dependencies, and the
development tools (tox
, black
, pytest
, etc.) are installed in a virtual
Python environment on the host system. Alternatively, they can be installed in a Docker
image, and Procrastinate and all the development tools can be run in Docker containers.
Docker is useful when you can't, or don't want to, install system requirements.
This section shows, through docker compose
command examples, how to test and run
Procrastinate in Docker.
Build the procrastinate
Docker image:
$ export UID GID
$ docker compose build procrastinate
Run the automated tests:
$ docker compose run --rm procrastinate pytest
Docker Compose is configured (in docker compose.yml
) to mount the local directory on
the host system onto /src
in the container. This means that local
changes made to the Procrastinate code are visible in Procrastinate containers.
The UID
and GID
environment variables are set and exported for the Procrastinate
container to be run with the current user id and group id. If not set or exported, the
Procrastinate container will run as root, and files owned by root may be created in the
developer's working directory.
In the definition of the procrastinate
service in docker compose.yml
the
PROCRASTINATE_APP
variable is set to procrastinate_demo.app.app
(the
Procrastinate demo application). So procrastinate
commands run in Procrastinate
containers are always run as if they were passed --app procrastinate_demo.app.app
.
Run the procrastinate
command :
$ docker compose run --rm procrastinate procrastinate -h
Apply the Procrastinate database schema:
$ docker compose run --rm procrastinate procrastinate schema --apply
Run the Procrastinate healthchecks:
$ docker compose run --rm procrastinate procrastinate healthchecks
Start a Procrastinate worker (-d
used to start the container in detached mode):
$ docker compose up -d procrastinate
Run a command (bash
here) in the Procrastinate worker container just started:
$ docker compose exec procrastinate bash
Watch the Procrastinate worker logs:
$ docker compose logs -ft procrastinate
Use the procrastinate defer
command to create a job:
$ docker compose run --rm procrastinate procrastinate defer procrastinate_demo.tasks.sum '{"a":3, "b": 5}'
Or run the demo main file:
$ docker compose run --rm procrastinate python -m procrastinate_demo
Stop and remove all the containers (including the postgres
container):
$ docker compose down
Yes, in order to provide both a synchronous and asynchronous API, Procrastinate needs to be asynchronous at core.
When possible, we're trying to avoid duplicating code, with designs such as "hoisting the I/O".
Dependencies for the package are handled by Poetry in
pyproject.toml
.
Whenever possible, we avoid pinning or putting any kind of limits on the
requirements. We'll typically only do that if we know that there's a
known conflict with specific versions. Typically, even if we support a subset of
Django versions, we won't restrict the Procrastinate package to those versions,
and if users use procrastinate with unsupported Django version and it works for
them, everyone is happy.
Dependencies for the development environment are kept in
poetry.lock
.
Those are updated regularily by Renovate which
merges their own PRs.
The versions in pre-commit-config.yaml
are kept in sync with poetry.lock
by the pre-commit
hook
poetry-to-pre-commit.
If you need to recompute the lockfile in your PR, you can use:
$ # Update all the pinned dependencies in pyproject.toml & all versions in poetry.lock
$ # (there are actually no pinned dependencies in pyproject.toml, so this only updates the
$ # lockfile).
$ poetry update
$ # Similarly, update dependencies in the lockfile. In procrastinate, it's equivalent
$ # to the command above
$ poetry lock
$ # Recompute the lockfile (e.g. after the pyproject.toml was updated) without trying
$ # to update anything
$ poetry lock --no-update
Please remember to tag Issues with appropriate labels.
PR labels help pre-filling the next release draft. They're not mandatory, but releasing will be easier if they're present.
Draft a new release, Generate release notes, add ## Migrations
. Make relevant
edits to the changelog, (see TODO
). Click on Release, that's it, the rest is
automated. This works with pre-release too.
When creating the release, GitHub will save the release info and create a tag
with the provided version. The new tag will be seen by GitHub Actions, which
will then create a wheel (using the tag as version number, thanks to
poetry-dynamic-versioning
), and push it to PyPI (using Trusted publishing).
That tag should also trigger a ReadTheDocs build, which will read GitHub
releases (thanks to our changelog
extension) which will write the changelog
in the published documentation (transformed from Markdown
to
RestructuredText
).
After a new major version is released (e.g. 2.0.0
), in preparation for the next
minor release (2.1.0
), the migration scripts in the future_migrations
directory
that remove the SQL compatibility code must be moved to the migrations
directory.
And the schema.sql
file must be updated accordingly.
:::{note} If you need to edit the name or body of a release in the GitHub UI, don't forget to also rebuild the stable and latest doc on readthedocs. :::