This repository is intended to be a base template, a cookiecutter for a new Python package project while keeping PEP518 in mind. Because it’s hosted on Github it already utilizes a few Github Actions that enforce repository-side checks for continuous integration and that implement a semantic release setup. And while this package is a starting point for a Python project with good engineering practices, it’s intended to be improved and added to in various ways — see the Wiki for more suggestions.
Features
Typing
Quality assurance
Unit testing
Documentation
Versioning and publishing
Dependency analysis
Security analysis
Package or application?
How to use this repository
Updating dependent packages
Git hooks
Testing
Generating documentation
Synchronizing with this template repo
Versioning, publishing and changelog
Build integrity using SLSA framework
Cleaning up
Frequently asked questions
The badges above give you an idea of what this project template provides. It’s work in progress, and I try to enable as much engineering goodness as is possible and is sensibly bearable using git hooks (see below) and Github Actions.
The package requires a minimum of Python 3.10, and it supports Python 3.11, Python 3.12 and Python 3.13 (default). All code requires comprehensive typing. The mypy static type checker and the flake8-pyi plugin are invoked by git hooks and through a Github Action to enforce continuous type checks on Python source and stub files. Make sure to add type hints to your code or to use stub files for types, to ensure that users of your package can import and type-check your code (see also PEP 561).
A number of git hooks are invoked before and after a commit, and before push. These hooks are all managed by the pre-commit tool and enforce a number of software quality assurance measures (see below).
Comprehensive unit testing is enabled using pytest combined with doctest and Hypothesis (to support property-based testing), and both code and branch coverage are measured using coverage (see below).
Documentation is important, and Sphinx is already set up to produce standard documentation in HTML and Markdown formats for the package, assuming that code contains docstrings with reStructuredText; the generated Markdown documentation can also optionally be pushed to the repository’s Github Wiki (see below).
Automatic package versioning and tagging, publishing to PyPI, and Changelog generation are enabled using Github Actions. Furthermore, an optional Release Notification Action allows Github to push an update notification to a Slack bot of your choice. For setup instructions, please see below.
Dependabot is enabled to scan the dependencies and automatically create pull requests when an updated version is available.
CodeQL is enabled to scan the Python code for security vulnerabilities. You can adjust the GitHub Actions workflow at .github/workflows/codeql-analysis.yaml and the configuration file at .github/codeql/codeql-config.yaml to add more languages, change the default paths, scan schedule, and queries.
OSSF Security Scorecards is enabled as a GitHub Actions workflow to give the consumers information about the supply-chain security posture of this project, assigning a score of 0–10. We upload the results as a SARIF (Static Analysis Results Interchange Format) artifact after each run and the results can be found at the Security tab of this GitHub project. We also allow publishing the data at OpenSSF. We use this data to continuously improve the security posture of this project. Note that this configuration supports the main (default) branch and requires the repository to be public and not forked.
pip-audit is part of the default Python virtual environment, and can be used to check all installed packages for documented CVE by querying the Python Packaging Advisory Database. The _build.yaml workflow always runs a package audit before the artifacts are being built. In automated production environments it may, on rare occasions, be necessary to suspend package auditing in which case you can add a repository variable DISABLE_PIP_AUDIT with value true to your repository to explicitly disable running pip-audit.
A shared package or library is intended to be imported by another package or application; an application is a self-contained, standalone, runnable package. Unfortunately, Python’s packaging ecosystem is mostly focused on packaging shared packages (libraries), and packaging Python applications is not as well-supported (discussion). This template, however, supports both scenarios.
Shared package: this template works out of the box as a shared package. Direct dependencies on other packages are declared in pyproject.toml (see the dependencies field) and should allow for as wide a version range as possible to ensure that this package and its dependencies can be installed by and coexist with other packages and applications without version conflicts.
Application: the __main__.py file ensures an entry point to run this package as a standalone application using Python’s -m command-line option. A wrapper script named something is also generated as an entry point into this package by make setup or make upgrade. In addition to specifying directly dependent packages and their version ranges in pyproject.toml, an application should pin its entire environment using the requirements.txt. Use the make requirements command to generate that file if you’re building an application.
The generated requirements.txt file with its integrity hash for every dependent package is used to generate a Software Bill of Materials (SBOM) in CycloneDX format. This is an important provenance material to provide transparency in the packaging process (see also SBOM + SLSA). That requirements.txt file, in addition to the SBOM, is also stored as a build artifact for every package release.
If you’d like to contribute to the project template, please open an issue for discussion or submit a pull request.
If you’d like to start your own Python project from scratch, you can either copy the content of this repository into your new project folder or fork this repository. Either way, consider making the following adjustments to your local copy:
-
Change the
LICENSE.mdfile and the license badge according to your needs, and adjust theSECURITY.mdfile to your needs (more details here). Remove all content from theCHANGELOG.mdfile. -
Rename the
src/package/folder to whatever your own package’s name will be, adjust the Github Actions in.github/workflows/, and review theMakefile,pyproject.toml,.pre-commit-config.yamlfiles as well as the unit tests accordingly. Note: by default all Actions run on three different host types (Linux, MacOS, and Windows) whose rates vary widely, so make sure that you disable or budget accordingly if you’re in a private repository! -
A new protected
releasebranch, should be created if it doesn't already exist. This branch should be configured with appropriate security policies and essential checks to ensure the integrity and stability of the release process. -
Adjust the content of the
pyproject.tomlfile according to your needs, and make sure to fill in the project URL, maintainer and author information too. Don’t forget to reset the package’s version number insrc/package/__init__.py. -
If you import packages that do not provide type hints into your new repository, then
mypyneeds to be configured accordingly: add these packages to thepyproject.tomlfile using theignore_missing_importsoption. -
If you’d like to publish your package to PyPI then uncomment the code in the
release.yamlAction, and add the appropriate environment variables. -
Adjust the Dependabot settings in
.github/dependabot.yamlto your desired target branch that you’d like to have monitored by Dependabot. -
Create the following Personal Access Tokens (PAT) with certain scopes for your Github user account and then create secrets for the new Github repository whose values are these new PATs:
- one PAT with
workflowandreposcope (including all of therepopermissions) for the secret namedREPO_ACCESS_TOKEN; this secret is used by the Release Action to push the release commit and attach assets to the generated Github release. - one PAT with
public_repo,read:discussion,read:org, andread:repo_hookscopes (detailed docs) for the secret namedSCORECARD_READ_TOKEN; this secret is used by the Scorecard Action to analyze the code and add its results to your repository. - one PAT with
reposcope for the secret namedDEPENDABOT_AUTOMERGE_TOKEN; this secret is used by the Dependabot Automerge Action to comment on Dependabot PRs.
- one PAT with
-
Create a Wiki and a first empty Wiki page for your new repository. Using the Wiki Documentation Action the repository’s Wiki will be automatically updated as part of publishing a new release.
To develop your new package, first create a virtual environment by either using the Makefile:
make venv # Create a new virtual environment in .venv folder using Python 3.13.or for a specific version of Python:
PYTHON=python3.12 make venv # Same virtual environment for a different Python version.or manually:
python3.13 -m venv .venv # Or use .venv313 for more than one local virtual environments.When working with this Makefile it is important to always activate the virtual environment because some of the git hooks (see below) depend on that:
. .venv/bin/activate # Where . is a bash shortcut for the source command.Finally, set up the new package with all of its extras and initialize the local git hooks:
make setupWith that in place, you’re ready to build your own package!
It’s likely that during development you’ll add or update dependent packages in the pyproject.toml file, which requires an update to the virtual environment:
make upgradeUsing the pre-commit tool and its .pre-commit-config.yaml configuration, the following git hooks are active in this repository:
- When committing code, a number of pre-commit hooks ensure that your code is formatted according to PEP 8 using the
blacktool, and they’ll invokeflake8(and various plugins),pylintandmypyto check for lint and correct types. There are more checks, but those two are the important ones. You can adjust the settings for these tools in thepyproject.tomlor.flake8configuration files. - The commit message hook enforces conventional commit messages and that, in turn, enables a semantic release of this package on the Github side: upon merging changes into the
releasebranch, the release action uses the Commitizen tool to produce a changelog and it computes the next version of this package and publishes a release — all based on the commit messages of a release. - Using a pre-push hook this package is also set up to run
pytest; in addition, thecoverageplugin makes sure that all of your package’s code is covered by tests and Hypothesis is already installed to help with generating test payloads. - The
actionlinthook is set up to lint GitHub Actions workflows. Ifshellcheckis installed on the system,actionlintrunsshellcheckto lint therunsteps in GitHub Actions. Note thatshellcheckis available on Ubuntu GitHub Actions runners by default.
You can also run these hooks manually, which comes in very handy during daily development tasks. For example
make check-coderuns all the code checks (i.e. bandit, flake8, pylint, mypy, actionlint), whereas
make checkruns all installed git hooks over your code. For more control over the code checks, the Makefile also implements the check-bandit, check-flake8, check-lint, check-mypy, and check-actionlint goals.
As mentioned above, this repository is set up to use pytest either standalone or as a pre-push git hook. Tests are stored in the tests/ folder, and you can run them manually like so:
make testwhich runs all tests in both your local Python virtual environment. For more options, see the pytest command-line flags. Also note that pytest includes doctest, which means that module and function docstrings, as well as the documentation, may contain test code that executes as part of the unit tests.
Both statement and branch coverage are being tracked using coverage and the pytest-cov plugin for pytest, and it measures how much code in the src/package/ folder is covered by tests:
Run unit tests...........................................................Passed
- hook id: pytest
- duration: 0.76s
============================= test session starts ==============================
platform darwin -- Python 3.13.1, pytest-8.3.4, pluggy-1.5.0 -- /path/to/python-package-template/.venv/bin/python
cachedir: .pytest_cache
hypothesis profile 'default-with-verbose-verbosity' -> max_examples=500, verbosity=Verbosity.verbose, database=DirectoryBasedExampleDatabase(PosixPath('/path/to/python-package-template/.hypothesis/examples'))
rootdir: /path/to/python-package-template
configfile: pyproject.toml
plugins: cov-6.0.0, hypothesis-6.122.7, env-1.1.5, custom-exit-code-0.3.0, doctestplus-1.3.0
collected 3 items
src/package/something.py::package.something.Something.do_something PASSED [ 33%]
tests/test_something.py::test_something PASSED [ 66%]
docs/source/index.rst::index.rst PASSED [100%]
---------- coverage: platform darwin, python 3.13.1-final-0 ----------
Name Stmts Miss Branch BrPart Cover Missing
----------------------------------------------------------------------
src/package/__init__.py 1 0 0 0 100%
src/package/something.py 4 0 0 0 100%
----------------------------------------------------------------------
TOTAL 5 0 0 0 100%
Required test coverage of 100.0% reached. Total coverage: 100.00%
============================ Hypothesis Statistics =============================
tests/test_something.py::test_something:
- during generate phase (0.00 seconds):
- Typical runtimes: ~ 0-1 ms, of which < 1ms in data generation
- 2 passing examples, 0 failing examples, 0 invalid examples
- Stopped because nothing left to do
============================== slowest durations ===============================
0.01s call tests/test_something.py::test_something
0.00s call src/package/something.py::package.something.Something.do_something
0.00s setup src/package/something.py::package.something.Something.do_something
0.00s call docs/source/index.rst::index.rst
0.00s teardown src/package/something.py::package.something.Something.do_something
0.00s setup docs/source/index.rst::index.rst
0.00s teardown docs/source/index.rst::index.rst
0.00s setup tests/test_something.py::test_something
0.00s teardown tests/test_something.py::test_something
============================== 3 passed in 0.09s ===============================
Note that code that’s not covered by tests is listed under the Missing column, and branches not taken too. The net effect of enforcing 100% code and branch coverage is that every new major and minor feature, every code change, and every fix are being tested (keeping in mind that high coverage does not imply comprehensive, meaningful test data).
Hypothesis is a package that implements property based testing and that provides payload generation for your tests based on strategy descriptions (more). Using its pytest plugin Hypothesis is ready to be used for this package.
As mentioned above, all package code should make use of Python docstrings in reStructured text format. Using these docstrings and the documentation template in the docs/source/ folder, you can then generate proper documentation in different formats using the Sphinx tool:
make docsThis example generates documentation in HTML, which can then be found here:
open docs/_build/html/index.htmlIn addition to the default HTML, Sphinx also generates Markdown documentation compatible with Github Wiki, and the Wiki Documentation Action automatically updates the project repository’s Wiki.
The sync-with-upstream.yaml GitHub Acions workflow checks this template repo daily and automatically creates a pull request in the downstream repo if there is a new release. Make sure to set up the GitHub username and email address in this workflow accordingly.
To enable automation for semantic versioning, package publishing, and changelog generation it is important to use meaningful conventional commit messages! This package template already has a built-in semantic release support enabled which is set up to take care of all three of these aspects — every time changes are pushed to the release branch.
With every package release, a new bump: commit is pushed to the release branch and tagged with the package’s new version. In addition, the main branch (which this repository uses to stage merged pull requests into for the next release) is rebased on top of the updated release branch automatically, so that subsequent pull requests can be merged while keeping a linear history.
If you’d like to receive Slack notifications whenever a new release is published, follow the comments in the Release Notification Action and set up a Slack bot by following the instructions here.
In order to build a distribution of your package locally instead of publishing it through the Github Actions workflow, you can simply call:
make distThis builds a source package and a binary distribution, and stores the files in your local dist/ folder.
You can also generate a changelog and bump the version manually and locally using commitizen (already installed as a dev dependency), for example:
cz changelog
cz bumpThe build process in this repository follows the requirements in the SLSA framework to be compliant at level 3. An important aspect of SLSA to improve the supply chain security posture is to generate a verifiable provenance for the build pipeline. Such a provenance can be used to verify the builder and let the consumers check the materials and configurations used while building an artifact. In this repository we use the generic provenance generator reusable workflow to generate a provenance that can attest to the following artifacts in every release:
- Binary dist (wheel)
- Source dist (tarball)
- SBOM (CycloneDx format)
- HTML and Markdown Docs
- A UNIX epoch timestamp file of the build time for reproducible builds
To verify the artifact using the provenance follow the instructions in the SLSA verifier project to install the verifier tool. After downloading the artifacts and provenance, verify each artifact individually, e.g.,:
slsa-verifier -artifact-path ~/Downloads/package-2.2.0.tar.gz -provenance attestation.intoto.jsonl -source github.com/jenstroeger/python-package-templateWhich should pass and provide the verification details.
On occasion it’s useful (and perhaps necessary) to clean up stale files, caches that tools like mypy leave behind, or even to nuke the complete virtual environment:
- Remove distribution artifacts:
make dist-clean - In addition, remove tool caches and documentation:
make clean - In addition, remove Python code caches and git hooks:
make nuke-caches - In addition and to reset everything, to restore a clean package to start over fresh:
make nuke
Please be careful when nuking your environment, and make sure you know what you’re doing.
- Question: Why don’t you use tools like tox or nox to orchestrate testing?
Answer: We’ve removedtoxbased on a discussion in issue #100 and PR #102. In short: we want to run tests inside the development venv usingpytest, and run more tests using an extensive test matrix using Github Actions.