In this tutorial we will go through in detail the steps required to set-up a Python package using the UCL-ARC/python-tooling
🍪 cookiecutter template as well as some follow on steps illustrating how to use the newly created package with some of the included tools.
Tip
Some of the commands and URLs in the instructions contain placeholders within curly braces such as {project_slug}
. You will need to replace these placeholders (including the curly braces) with the relevant values for your particular package / project - the text should give details of what should be subsituted for each placeholder.
Click to expand...
To use the template you will need to install the following software tools
The instructions for installing these tools can be accessed by clicking the individual links above.
Additionally, cookiecutter will be required to generate your package from the template, which can be installed using the instructions below. For the follow on exercises you will also need an account set up on GitHub if you don't already have one.
An easy way to install and manage Python packages is using uv
, which can create a dedicated environment with the necessary packages installed using the step by step instructions below
-
Open a terminal window
- On Windows: open the Start menu from the taskbar, type
cmd
in the search field, then click command prompt from the results. - On MacOS: click the Launchpad icon in the Dock, type
Terminal
in the search field, then clickTerminal
from the results. - On Linux: open the default terminal application installed in your distribution.
- On Windows: open the Start menu from the taskbar, type
-
In the terminal window, type
uv tool install cookiecutter
to install the "tools" necessary for creating and using a package using the
UCL-ARC/python-tooling
cookiecutter template. -
To check that all the dependencies have installed correctly, run each of the following commands in turn, one at a time
cookiecutter --version gh --version git --version
For each command you should see some text outputted to the terminal giving details of the installed versions of the applications - the output itself is not important as long as you do not see any error messages.
-
If you also want to try out creating a GitHub repository for the package you will need to sign-up for a free GitHub account if you don't already have one. Once you have a GitHub account, open a terminal window - you can either use the same one as previously if you still have it open, or open a new terminal window as described in step 2 and then run
gh auth login
to authenticate the GitHub command line interface tool
gh
with your GitHub account credentials. The tool will ask you a series of question, for most of which you can select the default options by just hitting theEnter
key. Specifically select:GitHub.com
for account to log into,HTTPS
for preferred protocol,Y
to authenticate Git with your GitHub credentials,Login with a web browser
as the method to authenticate.
Once you have selected all these options, a one-time code will be printed to the terminal. You need to copy this code and then hit the
Enter
key to open a page to complete the authentication in your default browser. Once you have entered and submitted the code in the authenticatication page, you should see aAuthentication complete
message appear in the terminal window.
We will first go through the steps for creating a new package using the UCL-ARC/python-tooling
cookiecutter template.
-
Open a new terminal/command prompt window and change the working directory to the path you wish to create the package in using the
cd
(change directory) commandcd path/to/directory
-
To begin creating the package run
cookiecutter gh:ucl-arc/python-tooling --checkout latest
You will then be shown a series of prompts at the command line asking for details of the project and package. You can choose to use the default placeholder value (shown in parenthesis
()
in prompt) for any question by hittingEnter
. If you already have a specific project in mind you want to set up a package for using the template you can use this project's details, otherwise you can just use the placeholder values. You should chooseY
(yes) to the questions on whether to initialise Git repository and automatically deploy HTML documentation to GitHub Pages to allow you to complete the follow on exercises which rely on these options being enabled. For the prompt asking for the GitHub user or organization name to be owner of repository you should supply your GitHub user name. -
Once you have completed all the cookiecutter prompts some additional instructions will be printed to screen (which we will come back to in the next sections) and your new package will be generated in a directory named
{project_slug}
in the current working directory (where{project_slug}
is the value entered for the'Slugified' project name...
1 prompt, this will bepython-template
if you used the default placeholder values). You can see the directory tree of files generated by runningtree {project_slug}
or on Windows
tree /F {project_slug}
in both cases replacing
{project_slug}
with the relevant value you used (python-template
if you used the default values). Some of the key files and directories are- The
README.md
file which is a Markdown file describing the project and acting as a landing page for first-time users. - The
LICENSE.md
file which contains the terms of the open-source license the code is released under. - The
src
directory which will contain the Python package source code. - The
docs
directory which will contain the Markdown files used to generate the documentation pages for the package. - The
tests
directory which will contain the Python modules defining tests to check for the correctness of the package code.
Try viewing the content of some of the files generated by running
cat {path_to_file}
on macOS/Linux or
type {path_to_file}
on Windows, replacing
{path_to_file}
with the relevant path. - The
The package you created in the previous section will have been initialised locally as a Git repository (providing you answered 'Y' to the relevant prompt!) to allow keeping the package under version control, but some extra steps are required to have the local repository synchronised to a remote repository on GitHub. The advantages of doing this are that makes it easy for other people to both download and use the package but also contribute to your project. The package has also been set up to take advantage of GitHub's continuous integration and deployment service GitHub Actions (which is free to use for public repositories) with workflows included for automatically running the package tests, performing linting checks and building the project documentation, on every pull-request and on updates to the default main
branch for the repository.
When you completed setting up the package using the cookiecutter
command you should have seen some additional instructions printed to screen including, providing you have the GitHub CLI gh
installed, a message of the form
GitHub CLI detected, you can create a repo with the following:
gh repo create {github_owner}/{project_slug} -d "{project_description}" --public -r origin --source {project_slug}
where {github_owner}
, {project_slug}
and {project_description}
are replaced with the relevant GitHub repository owner (user name), project slug and description that you entered when setting up the package using cookiecutter
. You should now copy-paste and run the gh repo create ...
command that was printed out in the message - if you no longer have access to the message you can use the example above, being careful to subsitute the placeholders in braces {}
with the relevant values.
If you get an error message at this point it may be because you have not installed the GitHub CLI or set up the authorisation for the GitHub CLI as described in the setup instructions above.
If the command runs successfully you should see a message of the form
✓ Created repository {github_owner}/{project_slug} on GitHub
https://github.com/{github_owner}/{project_slug}
✓ Added remote https://github.com/{github_owner}/{project_slug}.git
A repository should have been created at the printed URL which you should be able to navigate to in your browser (either copy-paste in to browser or depending on terminal application you may be able to hold Ctrl
and click on URL directly). Currently both the remote and local repositories are empty as we have not made any commits.
To commit the package files generated by cookiecutter
locally on the default main
branch and also push this commit to the repository on GitHub, run each of the commands below in turn, replacing the {project_slug}
placeholder with the relevant value you used when creating package (python-template
if you used default)
cd {project_slug}
git add .
git commit -m "Initial commit"
git push --set-upstream origin main
If you now navigate to the GitHub repository URL in your browser you should see the package files present, with the content of the top-level README.md
being displayed as the repository landing page.
The push of the initial commit to the default main
branch will have also triggered runs of the GitHub Actions workflows set up in the repository. You can view the status of the workflow job runs by opening the Actions
pane from the top navigation bar on the repository landing page, or going to the URL https://github.com/{github_owner}/{project_slug}/actions
.
Depending on your account settings, you may find that the Documentation
workflow shows as failed (❌). If this is the case this is likely because the repository has been set by default to only allow GitHub Actions workflows read permissions on the repository, which prevents the step in the Documentation
workflow which pushes the built HTML documentation to a branch gh-pages
from completing. In the next section we will look at how to configure the repository to ensure the permissions are set correctly and have the HTML documentation deployed to a GitHub Pages website.
Now that the repository is synchronised to GitHub, a few additional steps are required to configure the repository for the HTML documentation to be automatically deployed to GitHub Pages. On completion of the cookiecutter
command to create the package, a message of the form below should have been displayed
The 'Documentation' GitHub Actions workflow has been set up to push the built
HTML documentation to a branch gh-pages on pushes to main for deploying as a
GitHub Pages website. To allow the GitHub Actions bot to push to the gh-pages
branch you need to enable 'Read and write permissions' under 'Workflow
permissions' at
https://github.com/{github_owner}/{project_slug}/settings/actions
After the 'Documentation' workflow has successfully completed at least once
you will also need to configure the repository to deploy a GitHub pages site
from the content on the gh-pages branch by going to
https://github.com/{github_owner}/{project_slug}/settings/pages
and under 'Built and deployment' selecting 'Deploy from a branch' for the
'Source' drop-down and 'gh-pages' for the 'Branch' drop-down, leaving the
branch path drop-down with its default value of '/ (root)'.
where as before {github_owner}
and {project_slug}
are replaced with the relevant GitHub repository owner and project slug that you entered when setting up the package.
The first part of the instructions gives details of how to set the permissions for GitHub Actions workflows in the repository to allow the GitHub Actions bot to push the built HTML documentation to a branch gh-pages
on updates to the default main
branch on the repository, so that the documentation is automatically updated whenever changes are merged in to main
. After going to the Actions settings page at the URL indicated (which can also be reached by going to the Settings
tab in the top navigation bar of the repository and then selecting Action > General
in the left hand navigation menu) and ensuring Workflow permissions
is set to allow Read and write permissions
, if you have changed the permissions you will also need to re-run the documentation workflow with the updated permissions. This can be done by going to the Actions pane in the repository (accessible from top navigation bar or by going to https://github.com/{github_owner}/{project_slug}/actions
), clicking on the (top / latest if there are multiple) entry for the Documentation workflow, which should be showing as failing, and then from the resulting page click the Re-run all jobs
button.
Once the Documentation workflow has successfully completed, a new branch gh-pages
should have been created in the repository containing the HTML documentation. The second part of the instructions printed in the message output by the cookiecutter
command indicate to go to a URL https://github.com/{github_owner}/{project_slug}/settings/pages
to set this branch as the source for a GitHub Pages website for the repository. If you now follow those instructions, setting the Source
to Deploy from branch
and Branch
to gh-pages
, the a new Actions workflow pages-build-deployment
will be automatically triggered. Once this workflow has completed, you should be able to view the HTML documentation for the repository at a URL
https://{github_owner}.github.io/{project_slug}
The index page of the documentation reproduces the content from the repository README file. The documentation site also importantly includes application programming interface (API) reference documentation built from the docstrings in the package source code; this API documentation is accessible from the API reference
link in the navigation.
It is recommended to set up a project-specific virtual environment whenever working on Python project. This allows you to install the versions of third-party packages the project requires without conflicting with the requirements of other projects you are developing.
There are a variety of virtual environment management tools available for Python. One option is uv, which you will have installed if you followed our detailed set-up instructions above, or may already have installed previously. uv
is an extremely fast Python package and virtual environment manager that has API compatiblility with pip
and the built-in virtual environment manager - venv
.
Once you are in your {project_slug}
directory, a virtual environment can be created by running the following in a terminal/command prompt window
uv venv .venv
This will create a new environment with name .venv
, which can be activated for installing packages using
source .venv/bin/activate
or if on Windows
.venv\Scripts\activate
You should see (.venv)
on the left of your terminal window. Once you have activated the environment you can use the following command to install any packages within the .venv
environment
uv pip install numpy
The environment can be deactivated using
deactivate
You can even use an arbitrary Python version to create uv
environments, and uv
will fetch that Python version if it is not available locally
uv venv --python 3.11.6
Once you have created and activated a virtual environment for the project, you can install the package in editable mode, along with both its required dependencies and optional sets of dependencies for development (dev
), documentation (docs
) and testing (test
) by running
uv pip install --editable ".[dev,docs,test]"
from the root of the project repository.
Using venv as an alternative to uv
Alternatively, you can use the venv
module, which is slower and has fewer features, when compared to uv
, but is built-in to the Python standard library. venv
has the advantage of being available in any Python (3.3+) environment, but unlike uv
will not by itself allow you to use a different Python version from the system level install. A common pattern is to store the virtual environment files in a directory .venv
within the root directory of the project repository. This can be achieved by running
python -m venv .venv
from the root of the project repository. To activate the new virtual environment, if on Linux or MacOS run from the root of the project repository
source .venv/bin/activate
or if on Windows
.venv\Scripts\activate
Once you have activated the environment you should make sure the version of the Python package manager pip
installed in the environment is up to date by running
python -m pip install --upgrade pip
Similar to uv
, once the the environment is active, you can install the package in editable mode, along with both its required dependencies and optional sets of dependencies by running
python -m pip install --editable ".[dev,docs,test]"
The package template includes support for running tests using the pytest
testing framework. Tests are defined in modules in the tests
directory. By default a single test module tests/test_dummy.py
is created with a placeholder test. You can run the test locally using pytest
directly (you will need to have the virtual environment active in which you installed the package and its dependencies) by running
pytest
from the root of the project repository.
The package template also sets up tox
, an automation tool which allows easily running tests in isolated environments and under multiple Python versions. You can run the tests across all compatible Python versions currently installed using tox
by running
tox
from the root of the project repository.
It can sometimes be useful when editing docstrings or adding additional pages in the docs
directory to be able to render the HTML documentation locally. The package is set up to use MkDocs to build the documentation with the package API documentation generated using the mkdocstrings
plug-in. These packages will have been installed in to you local development environment providing you installed the package with optional docs
dependencies as recommended above. To build and serve the documentation locally run
mkdocs serve
from the root of the project repository and navigate to http://127.0.0.1:8000/
in your browser. The development server used here supports auto-reloading, meaning the content will be automatically refreshed in your browser if you make any edits to source.
A tox
environment docs
to build the documentation is also available. This will be build the documentation in an isolated environment and is also used for building the documentation in the GitHub Actions Documentation workflow so can be useful to run locally when debugging issues with the workflow - it can be executed by running
tox -e docs
from the root of the project repository. The built documentation will be output to a directory site
.
The package is set-up to use pre-commit, a framework for running Git hook scripts on each commit to the repository. In particular the a .pre-commit-config.yaml
configuration file is provided which will run a series of linters, checks and formatters on the repository such as ruff, mypy and prettier on every commit. These Git hook scripts can installed locally by running
pre-commit install
from the root of the project repository. Once installed the hook scripts will be called to inspect the changes each time git commit
is run, with any failures of the checks needing to be resolved before the changes can be commited. Some of the pre-commit
hooks include support for auto-fixing some problems - in this case you will be alerted that a file has been changed by a hook and these changes need to be staged using git add
before recommitting.
The pre-commit
hooks can be run against all files in the repository by running
pre-commit run --all-files
from the root of the project repository. As the hooks are typically only run on the files changed in a commit, this can be useful to check the hooks will pass on all files when the pre-commit
configuration is updated by, for example, adding a new plug-in.