From 48b946c1d85f0a707ce796bd8b27c39eb6eeb8dc Mon Sep 17 00:00:00 2001 From: Theodore Kisner Date: Sun, 5 Jan 2025 22:28:37 -0800 Subject: [PATCH] Improvements in use of python SharedMemory (#26) * Improvements in use of python SharedMemory: - Remove previous patch and instead ship a backport of the changes which are in python 3.13. This allows the "track=False" option when creating SharedMemory segments. - In order to better handle cleanup, keep a registry of all memory segments allocated on every process. Register a cleanup handler for python atexit and also register a cleanup handler for signals such as SIGTERM. - Expand unit tests - Convert build to pyproject.toml with minimal setup.py backend. - Improve test and deployment workflows. * Fix test.yml branch * Remove some typing info in backport, for compatibility with 3.9 * Use mpich for running CI tests * Remove stale scripts * Add multinode test script for perlmutter. * Remove stale scripts and fix typo in readme --- .github/workflows/deploy.yml | 24 +- .github/workflows/test.yml | 149 +- README.md | 20 +- pshmem/__init__.py | 2 +- pshmem/locking.py | 2 +- pshmem/registry.py | 60 + pshmem/shmem.py | 47 +- pshmem/test.py | 347 +-- pshmem/utils.py | 100 +- pyproject.toml | 45 + setup.cfg | 12 - setup.py | 43 +- .../install_deps_cibuildwheel_linux.sh | 13 - .../install_deps_cibuildwheel_macos.sh | 12 - test_scripts/install_deps_github_macos.sh | 19 - test_scripts/install_deps_github_ubuntu.sh | 20 - test_scripts/run_cori_haswell.slurm | 32 - ...un_cori_knl.slurm => run_perlmutter.slurm} | 16 +- test_scripts/test_cibuild.sh | 15 - versioneer.py | 2277 ----------------- 20 files changed, 544 insertions(+), 2711 deletions(-) create mode 100644 pshmem/registry.py create mode 100644 pyproject.toml delete mode 100644 setup.cfg delete mode 100755 test_scripts/install_deps_cibuildwheel_linux.sh delete mode 100755 test_scripts/install_deps_cibuildwheel_macos.sh delete mode 100755 test_scripts/install_deps_github_macos.sh delete mode 100755 test_scripts/install_deps_github_ubuntu.sh delete mode 100644 test_scripts/run_cori_haswell.slurm rename test_scripts/{run_cori_knl.slurm => run_perlmutter.slurm} (66%) delete mode 100755 test_scripts/test_cibuild.sh delete mode 100644 versioneer.py diff --git a/.github/workflows/deploy.yml b/.github/workflows/deploy.yml index 6e36ece..a7efe90 100644 --- a/.github/workflows/deploy.yml +++ b/.github/workflows/deploy.yml @@ -29,32 +29,26 @@ jobs: - name: Checkout uses: actions/checkout@v4 - - name: Set up Python 3.10 + - name: Set up Python 3.12 uses: actions/setup-python@v5 with: - python-version: "3.10" + python-version: "3.12" - name: Install Dependencies - run: ./test_scripts/install_deps_github_ubuntu.sh && pip install twine + run: python3 -m pip install build twine numpy - name: Install Package - run: pip install . + run: python3 -m pip install . - - name: Run Serial Test - run: MPI_DISABLE=1 python3 -c 'import pshmem.test; pshmem.test.run()' - - - name: Run MPI Test on 1 Process - run: mpirun -np 1 python3 -c 'import pshmem.test; pshmem.test.run()' - - - name: Run MPI Test on 2 Processes - run: mpirun -np 2 python3 -c 'import pshmem.test; pshmem.test.run()' + - name: Run Simple Test + run: python3 -c 'import pshmem.test; pshmem.test.run()' - name: Build source package - run: rm -rf dist && python setup.py sdist + run: rm -rf dist && python3 -m build --sdist - name: Build wheels - run: mkdir -p wheelhouse && pip wheel -w wheelhouse . + run: python3 -m build --wheel - name: Upload to PyPI run: | - python -m twine upload dist/*.tar.gz && python -m twine upload wheelhouse/pshmem*.whl + python -m twine upload dist/*.tar.gz && python -m twine upload dist/pshmem*.whl diff --git a/.github/workflows/test.yml b/.github/workflows/test.yml index 6b4bfdc..0d25bc2 100644 --- a/.github/workflows/test.yml +++ b/.github/workflows/test.yml @@ -1,72 +1,105 @@ -# Use pre-built docker containers to run our unit tests on different python versions. +# In general, we try to run on: +# - The oldest supported python +# - The latest stable python that is the common default on most systems and conda +# - (During transitions) The newly released bleeding edge python -name: Tests +name: Run Test Suite on: push: - branches: [ master ] + branches: + - master pull_request: - branches: [ master ] + branches: + - master concurrency: group: ${{ github.workflow }}-${{ github.ref }} cancel-in-progress: true jobs: - linux: - runs-on: ubuntu-latest + test: + name: Tests on ${{ matrix.arch }} with Conda Python-${{ matrix.python }} + runs-on: ${{ matrix.os }} + defaults: + run: + shell: bash -l {0} strategy: + fail-fast: false matrix: - python-version: ["3.8", "3.10", "3.11", "3.12"] + include: + - os: ubuntu-latest + python: "3.9" + arch: Linux-x86_64 + - os: ubuntu-latest + python: "3.11" + arch: Linux-x86_64 + - os: ubuntu-latest + python: "3.13" + arch: Linux-x86_64 + - os: macos-latest + python: "3.10" + arch: MacOSX-x86_64 + - os: macos-latest + python: "3.13" + arch: MacOSX-x86_64 + - os: macos-latest + python: "3.10" + arch: MacOSX-arm64 + - os: macos-latest + python: "3.13" + arch: MacOSX-arm64 steps: - - name: Checkout - uses: actions/checkout@v4 - - - name: Set up Python ${{ matrix.python-version }} - uses: actions/setup-python@v5 - with: - python-version: ${{ matrix.python-version }} - - - name: Install Dependencies - run: ./test_scripts/install_deps_github_ubuntu.sh - - - name: Install Package - run: pip3 install . - - - name: Run Serial Test - run: MPI_DISABLE=1 python3 -c 'import pshmem.test; pshmem.test.run()' - - - name: Run MPI Test on 1 Process - run: mpirun -np 1 python3 -c 'import pshmem.test; pshmem.test.run()' - - - name: Run MPI Test on 2 Processes - run: mpirun -np 2 python3 -c 'import pshmem.test; pshmem.test.run()' - - macos: - runs-on: macos-latest - strategy: - matrix: - python-version: ["3.10", "3.11", "3.12"] - steps: - - name: Checkout - uses: actions/checkout@v4 - - - name: Set up Python ${{ matrix.python-version }} - uses: actions/setup-python@v5 - with: - python-version: ${{ matrix.python-version }} - - - name: Install Dependencies - run: ./test_scripts/install_deps_github_macos.sh - - - name: Install Package - run: pip3 install . - - - name: Run Serial Test - run: MPI_DISABLE=1 python3 -c 'import pshmem.test; pshmem.test.run()' - - - name: Run MPI Test on 1 Process - run: mpirun -np 1 python3 -c 'import pshmem.test; pshmem.test.run()' - - - name: Run MPI Test on 2 Processes - run: mpirun -np 2 python3 -c 'import pshmem.test; pshmem.test.run()' + - name: Checkout + uses: actions/checkout@v4 + + - name: Setup Conda Base + run: | + sudo rm -rf /usr/share/miniconda \ + && sudo rm -rf /usr/local/miniconda \ + && curl -SL -o miniforge.sh https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-${{ matrix.arch }}.sh \ + && bash miniforge.sh -b -f -p ~/conda \ + && source ~/conda/etc/profile.d/conda.sh \ + && conda activate base \ + && conda update -n base --yes conda + + - name: Check Conda Config + run: | + source ~/conda/etc/profile.d/conda.sh \ + && conda activate base \ + && conda info \ + && conda list \ + && conda config --show-sources \ + && conda config --show + + - name: Install Dependencies + run: | + source ~/conda/etc/profile.d/conda.sh \ + && conda create --yes -n test python==${{ matrix.python }} \ + && conda activate test \ + && conda install --yes numpy mpich mpi4py + + - name: Install + run: | + source ~/conda/etc/profile.d/conda.sh \ + && conda activate test \ + && pip install . + + - name: Run Serial Tests + run: | + source ~/conda/etc/profile.d/conda.sh \ + && conda activate test \ + && mkdir -p test \ + && pushd test >/dev/null 2>&1 \ + && MPI_DISABLE=1 python3 -c 'import pshmem.test; pshmem.test.run()' \ + && popd >/dev/null 2>&1 + + - name: Run MPI Tests + run: | + source ~/conda/etc/profile.d/conda.sh \ + && conda activate test \ + && mkdir -p test \ + && pushd test >/dev/null 2>&1 \ + && mpirun -np 1 python3 -c 'import pshmem.test; pshmem.test.run()' \ + && mpirun -np 2 python3 -c 'import pshmem.test; pshmem.test.run()' \ + && popd >/dev/null 2>&1 diff --git a/README.md b/README.md index 8da4a9d..fd67dba 100644 --- a/README.md +++ b/README.md @@ -8,16 +8,24 @@ shared memory constructs. This package needs a recent version of the `mpi4py` package in order to be useful. However, the classes also accept a value of `None` for the communicator, in which case a trivial local implementation is used. The code uses other widely available packages -(like numpy) and requires a recent Python3 installation. You can install the code from -a git checkout with: +(like numpy) and requires a recent Python3 installation. - pip install . +### Binary Packages + +Wheels are available on PyPI: + + pip install pshmem + +Or you can install packages from conda-forge: -Or: + conda install -c conda-forge pshmem - python3 setup.py install +### Installing from Source + +You can install the code from a git checkout with: + + pip install . -Or directly from github. ## MPIShared Class diff --git a/pshmem/__init__.py b/pshmem/__init__.py index 37420e2..78b28bf 100644 --- a/pshmem/__init__.py +++ b/pshmem/__init__.py @@ -1,5 +1,5 @@ ## -# Copyright (c) 2017-2020, all rights reserved. Use of this source code +# Copyright (c) 2017-2025, all rights reserved. Use of this source code # is governed by a BSD license that can be found in the top-level # LICENSE file. ## diff --git a/pshmem/locking.py b/pshmem/locking.py index 148d793..2d0a698 100644 --- a/pshmem/locking.py +++ b/pshmem/locking.py @@ -1,5 +1,5 @@ ## -# Copyright (c) 2017-2024, all rights reserved. Use of this source code +# Copyright (c) 2017-2025, all rights reserved. Use of this source code # is governed by a BSD license that can be found in the top-level # LICENSE file. ## diff --git a/pshmem/registry.py b/pshmem/registry.py new file mode 100644 index 0000000..f979197 --- /dev/null +++ b/pshmem/registry.py @@ -0,0 +1,60 @@ +## +# Copyright (c) 2017-2025, all rights reserved. Use of this source code +# is governed by a BSD license that can be found in the top-level +# LICENSE file. +## + +import sys +import atexit +import signal + + +class MPISharedRegistry: + """Registry of shared memory buffers. + + This tracks all MPIShared memory buffers on the current process and + ensures they are cleaned up when the process is terminated. + + """ + def __init__(self): + self.reg = dict() + + def register(self, name, buffer, noderank): + self.reg[name] = (buffer, noderank) + + def unregister(self, name): + del self.reg[name] + + def cleanup(self): + for name, (buf, noderank) in self.reg.items(): + buf.close() + if noderank == 0: + buf.unlink() + self.reg.clear() + + +"""Single instance of the registry""" +registry = MPISharedRegistry() + + +def _signal_handler(sig, frame): + global registry + registry.cleanup() + sys.exit(0) + + +@atexit.register +def _atexit_handler(): + global registry + registry.cleanup() + + +def _register_signals(): + signal.signal(signal.SIGINT, _signal_handler) + signal.signal(signal.SIGTERM, _signal_handler) + signal.signal(signal.SIGQUIT, _signal_handler) + signal.signal(signal.SIGHUP, _signal_handler) + + +# Register signal handlers on import +_register_signals() diff --git a/pshmem/shmem.py b/pshmem/shmem.py index aab4b2b..64ff318 100644 --- a/pshmem/shmem.py +++ b/pshmem/shmem.py @@ -1,24 +1,20 @@ ## -# Copyright (c) 2017-2024, all rights reserved. Use of this source code +# Copyright (c) 2017-2025, all rights reserved. Use of this source code # is governed by a BSD license that can be found in the top-level # LICENSE file. ## import sys -from multiprocessing import shared_memory import numpy as np +from .registry import registry from .utils import ( + SharedMemory, mpi_data_type, random_shm_key, - remove_shm_from_resource_tracker, ) -# Monkey patch resource_tracker. Remove once upstream CPython -# changes are merged. -remove_shm_from_resource_tracker() - class MPIShared(object): """Create a shared memory buffer that is replicated across nodes. @@ -183,10 +179,14 @@ def __init__(self, shape, dtype, comm, comm_node=None, comm_node_rank=None): self.data = self._flat.reshape(self._shape) else: # First rank on each node creates the buffer + mem_err = 0 if self._noderank == 0: try: - self._shmem = shared_memory.SharedMemory( - name=self._name, create=True, size=int(nbytes), + self._shmem = SharedMemory( + name=self._name, + create=True, + size=int(nbytes), + track=False, ) except Exception as e: msg = "Process {}: {}".format(self._rank, self._name) @@ -196,17 +196,22 @@ def __init__(self, shape, dtype, comm, comm_node=None, comm_node_rank=None): ) msg += ": {}".format(e) print(msg, flush=True) - raise - - # Wait for that to be created + mem_err = 1 + # All ranks check for error if self._nodecomm is not None: - self._nodecomm.barrier() + mem_err = self._nodecomm.bcast(mem_err, root=0) + if mem_err != 0: + raise RuntimeError("Failed to allocate shared memory") # Other ranks on the node attach + mem_err = 0 if self._noderank != 0: try: - self._shmem = shared_memory.SharedMemory( - name=self._name, create=False, size=int(nbytes) + self._shmem = SharedMemory( + name=self._name, + create=False, + size=int(nbytes), + track=False, ) except Exception as e: msg = "Process {}: {}".format(self._rank, self._name) @@ -216,7 +221,15 @@ def __init__(self, shape, dtype, comm, comm_node=None, comm_node_rank=None): ) msg += ": {}".format(e) print(msg, flush=True) - raise + mem_err = 1 + if self._nodecomm is not None: + mem_err = self._nodecomm.allreduce(mem_err, op=MPI.SUM) + if mem_err != 0: + raise RuntimeError("Failed to attach to shared memory") + + # Register the shared memory buffer for cleanup if program is + # terminated. + registry.register(self._name, self._shmem, self._noderank) # Create a numpy array which acts as a view of the buffer. self._flat = np.ndarray( @@ -363,6 +376,8 @@ def close(self): del self._flat if hasattr(self, "_shmem"): if self._shmem is not None: + # Unregister the shared memory buffer, since we are about to close it. + registry.unregister(self._name) self._shmem.close() if self._noderank == 0: self._shmem.unlink() diff --git a/pshmem/test.py b/pshmem/test.py index 22de215..48479b4 100644 --- a/pshmem/test.py +++ b/pshmem/test.py @@ -1,14 +1,11 @@ ## -# Copyright (c) 2017-2024, all rights reserved. Use of this source code +# Copyright (c) 2017-2025, all rights reserved. Use of this source code # is governed by a BSD license that can be found in the top-level # LICENSE file. ## import os import sys -import time -import traceback - import unittest import numpy as np @@ -16,6 +13,7 @@ from .shmem import MPIShared from .locking import MPILock +from .utils import exception_guard MPI = None use_mpi = True @@ -27,7 +25,7 @@ try: import mpi4py.MPI as MPI except ImportError: - print("Cannot import mpi4py, will only test serial functionality.") + print("Cannot import mpi4py, will only test serial functionality.", flush=True) class ShmemTest(unittest.TestCase): @@ -44,20 +42,16 @@ def setUp(self): def tearDown(self): pass - def read_write(self, comm, comm_node=None, comm_node_rank=None): - """Run a sequence of various access tests.""" + def _write_read_buffer(self, local, shm): rank = 0 procs = 1 - if comm is not None: - rank = comm.rank - procs = comm.size - - # Dimensions of our shared memory array - datadims = (2, 5, 10) - + if shm.comm is not None: + rank = shm.comm.rank + procs = shm.comm.size + datadims = local.shape # Dimensions of the incremental slab that we will # copy during each set() call. - updatedims = (1, 1, 5) + updatedims = (1, 1, datadims[2] // 2) # How many updates are there to cover the whole # data array? @@ -65,8 +59,113 @@ def read_write(self, comm, comm_node=None, comm_node_rank=None): for d in range(len(datadims)): nupdate *= datadims[d] // updatedims[d] - for datatype in [np.int32, np.int64, np.float32, np.float64]: + for p in range(procs): + # Every process takes turns writing to the buffer. + setdata = None + setoffset = (0, 0, 0) + + # Write to the whole data volume, but in small blocks + for upd in range(nupdate): + if p == rank: + # My turn! Write my process rank to the buffer slab. + setdata = local[ + setoffset[0] : setoffset[0] + updatedims[0], + setoffset[1] : setoffset[1] + updatedims[1], + setoffset[2] : setoffset[2] + updatedims[2], + ] + try: + # All processes call set(), but only data on rank p matters. + shm.set(setdata, setoffset, fromrank=p) + except (RuntimeError, ValueError): + print( + "proc {} threw exception during set()".format(rank), + flush=True, + ) + raise + try: + # Same as set(), but using __setitem__ with an + # allreduce to find which process is setting. + + # key as a tuple slices + if setdata is None: + shm[None] = setdata + else: + shm[ + setoffset[0] : setoffset[0] + setdata.shape[0], + setoffset[1] : setoffset[1] + setdata.shape[1], + setoffset[2] : setoffset[2] + setdata.shape[2], + ] = setdata + except (RuntimeError, ValueError): + print( + "proc {} threw exception during __setitem__".format(rank), + flush=True, + ) + raise + + # Increment the write offset within the array + + x = setoffset[0] + y = setoffset[1] + z = setoffset[2] + + z += updatedims[2] + if z >= datadims[2]: + z = 0 + y += updatedims[1] + if y >= datadims[1]: + y = 0 + x += updatedims[0] + + setoffset = (x, y, z) + + # Every process is now going to read a copy from the shared memory + # and make sure that they see the data written by the current process. + check = np.zeros_like(local) + check[:, :, :] = shm[:, :, :] + + truth = np.ones_like(local) + truth *= p + + # This should be bitwise identical, even for floats + nt.assert_equal(check[:, :, :], truth[:, :, :]) + + # Try full array assignment with slices containing None start + # values + if p != rank: + shm[None] = None + else: + shm[:, :, :] = local + + check[:, :, :] = shm[:, :, :] + nt.assert_equal(check[:, :, :], truth[:, :, :]) + + # Ensure that we can reference the memory buffer from numpy without + # a memory copy. The intention is that a slice of the shared memory + # buffer should appear as a C-contiguous ndarray whenever we slice + # along the last dimension. + + for p in range(procs): + if p == rank: + slc = shm[1, 2] + print( + "proc {} slice has dims {}, dtype {}, C = {}".format( + p, slc.shape, slc.dtype.str, slc.flags["C_CONTIGUOUS"] + ), + flush=True, + ) + if shm.comm is not None: + shm.comm.barrier() + + def context_write_read(self, comm, comm_node=None, comm_node_rank=None): + """Run a sequence of various access tests.""" + rank = 0 + if comm is not None: + rank = comm.rank + + # Dimensions of our shared memory array + datadims = (2, 5, 10) + for datatype in [np.int32, np.int64, np.float32, np.float64]: # For testing the "set()" method, every process is going to # create a full-sized data buffer and fill it with its process rank. local = np.ones(datadims, dtype=datatype) @@ -75,7 +174,6 @@ def read_write(self, comm, comm_node=None, comm_node_rank=None): # A context manager is the pythonic way to make sure that the # object has no dangling reference counts after leaving the context, # and will ensure that the shared memory is freed properly. - with MPIShared( local.shape, local.dtype, @@ -83,124 +181,50 @@ def read_write(self, comm, comm_node=None, comm_node_rank=None): comm_node=comm_node, comm_node_rank=comm_node_rank, ) as shm: - for p in range(procs): - # Every process takes turns writing to the buffer. - setdata = None - setoffset = (0, 0, 0) - - # Write to the whole data volume, but in small blocks - for upd in range(nupdate): - if p == rank: - # My turn! Write my process rank to the buffer slab. - setdata = local[ - setoffset[0] : setoffset[0] + updatedims[0], - setoffset[1] : setoffset[1] + updatedims[1], - setoffset[2] : setoffset[2] + updatedims[2], - ] - try: - # All processes call set(), but only data on rank p matters. - shm.set(setdata, setoffset, fromrank=p) - except (RuntimeError, ValueError): - print( - "proc {} threw exception during set()".format(rank), - flush=True, - ) - if comm is not None: - comm.Abort() - else: - sys.exit(1) - - try: - # Same as set(), but using __setitem__ with an - # allreduce to find which process is setting. - - # key as a tuple slices - if setdata is None: - shm[None] = setdata - else: - shm[ - setoffset[0] : setoffset[0] + setdata.shape[0], - setoffset[1] : setoffset[1] + setdata.shape[1], - setoffset[2] : setoffset[2] + setdata.shape[2], - ] = setdata - except (RuntimeError, ValueError): - print( - "proc {} threw exception during __setitem__".format( - rank - ), - flush=True, - ) - if comm is not None: - exc_type, exc_value, exc_traceback = sys.exc_info() - lines = traceback.format_exception( - exc_type, exc_value, exc_traceback - ) - lines = ["Proc {}: {}".format(rank, x) for x in lines] - print("".join(lines), flush=True) - comm.Abort() - else: - raise - - # Increment the write offset within the array - - x = setoffset[0] - y = setoffset[1] - z = setoffset[2] - - z += updatedims[2] - if z >= datadims[2]: - z = 0 - y += updatedims[1] - if y >= datadims[1]: - y = 0 - x += updatedims[0] - - setoffset = (x, y, z) - - # Every process is now going to read a copy from the shared memory - # and make sure that they see the data written by the current process. - check = np.zeros_like(local) - check[:, :, :] = shm[:, :, :] - - truth = np.ones_like(local) - truth *= p - - # This should be bitwise identical, even for floats - nt.assert_equal(check[:, :, :], truth[:, :, :]) - - # Try full array assignment with slices containing None start - # values - if p != rank: - shm[None] = None - else: - shm[:, :, :] = local - - check[:, :, :] = shm[:, :, :] - nt.assert_equal(check[:, :, :], truth[:, :, :]) - - # Ensure that we can reference the memory buffer from numpy without - # a memory copy. The intention is that a slice of the shared memory - # buffer should appear as a C-contiguous ndarray whenever we slice - # along the last dimension. - - for p in range(procs): - if p == rank: - slc = shm[1, 2] - print( - "proc {} slice has dims {}, dtype {}, C = {}".format( - p, slc.shape, slc.dtype.str, slc.flags["C_CONTIGUOUS"] - ), - flush=True, - ) - if comm is not None: - comm.barrier() + self._write_read_buffer(local, shm) + + def create_separated(self, comm): + # Test creation of shared memory objects outside of a context manager + data = dict() + rank = 0 + if comm is not None: + rank = comm.rank + + # Dimensions of our shared memory array + datadims = (2, 5, 10) + + for datatype in [np.int32, np.int64, np.float32, np.float64]: + local = np.ones(datadims, dtype=datatype) + local *= rank + data[datatype] = MPIShared( + local.shape, + local.dtype, + comm, + ) + self._write_read_buffer(local, data[datatype]) + return data + + def close_separated(self, data): + for dt, shm in data.items(): + shm.close() + del data def test_world(self): if self.comm is None: print("Testing MPIShared without MPI...", flush=True) elif self.comm.rank == 0: print("Testing MPIShared with world communicator...", flush=True) - self.read_write(self.comm) + self.context_write_read(self.comm) + + def test_separated(self): + if self.comm is None: + print("Testing separated create/close without MPI...", flush=True) + elif self.comm.rank == 0: + print( + "Testing separated create/close with world communicator...", flush=True + ) + data = self.create_separated(self.comm) + self.close_separated(data) def test_split(self): if self.comm is not None: @@ -282,16 +306,24 @@ def test_split(self): # Test the access and creation of shared memory objects across all # these different communicators. - self.read_write(wcomm, comm_node=nodecomm, comm_node_rank=noderankcomm) + self.context_write_read( + wcomm, comm_node=nodecomm, comm_node_rank=noderankcomm + ) wcomm.barrier() - self.read_write(gcomm, comm_node=gnodecomm, comm_node_rank=gnoderankcomm) + self.context_write_read( + gcomm, comm_node=gnodecomm, comm_node_rank=gnoderankcomm + ) wcomm.barrier() - self.read_write(comm_row, comm_node=comm_row_node, comm_node_rank=comm_row_rank_node) + self.context_write_read( + comm_row, comm_node=comm_row_node, comm_node_rank=comm_row_rank_node + ) wcomm.barrier() - self.read_write(comm_col, comm_node=comm_col_node, comm_node_rank=comm_col_rank_node) + self.context_write_read( + comm_col, comm_node=comm_col_node, comm_node_rank=comm_col_rank_node + ) wcomm.barrier() # Clean up row / column communicators @@ -316,7 +348,7 @@ def test_comm_self(self): if self.comm.rank == 0: print("Testing MPIShared with COMM_SELF...", flush=True) # Every process does the operations on COMM_SELF - self.read_write(MPI.COMM_SELF) + self.context_write_read(MPI.COMM_SELF) def test_comm_reuse(self): if self.comm is not None: @@ -329,7 +361,9 @@ def test_comm_reuse(self): mynode = self.comm.rank // nodeprocs rankcomm = self.comm.Split(noderank, mynode) - self.read_write(self.comm, comm_node=nodecomm, comm_node_rank=rankcomm) + self.context_write_read( + self.comm, comm_node=nodecomm, comm_node_rank=rankcomm + ) if nodes > 1 and nodeprocs > 2: # We have at least one node, test passing in an incorrect @@ -343,10 +377,13 @@ def test_comm_reuse(self): comm_node=evenoddcomm, comm_node_rank=evenoddcomm, ) - print("Failed to catch construction with bad node comm") + print("Failed to catch construction with bad node comm", flush=True) self.assertTrue(False) except ValueError: - print("Successfully caught construction with bad node comm") + print( + "Successfully caught construction with bad node comm", + flush=True, + ) def test_shape(self): good_dims = [ @@ -397,8 +434,8 @@ def test_array(self): view = np.array(shm, copy=False) vptr, vflag = view.__array_interface__["data"] sptr, sflag = shm._flat.__array_interface__["data"] - print(f"numpy view address = {vptr}") - print(f"original address = {sptr}") + print(f"numpy view address = {vptr}", flush=True) + print(f"original address = {sptr}", flush=True) self.assertTrue(vptr == sptr) def test_zero(self): @@ -425,10 +462,26 @@ def test_zero(self): except RuntimeError: print("successful raise with no data during set()", flush=True) + def test_max_shmem_segments(self): + handles = list() + n_seg = 0 + failed = False + while not failed and n_seg < 10000: + try: + shm = MPIShared((5, 5), np.float64, self.comm) + handles.append(shm) + n_seg += 1 + except Exception: + failed = True + print(f"Allocated {n_seg} shared memory segments without OS error", flush=True) + for h in handles: + h.close() + handles.clear() + # def test_hang(self): # # Run this while monitoring memory usage (e.g. with htop) and then - # # do kill -9 on one of the processes to verify that the kernel - # # releases shared memory. + # # do kill -15 (SIGTERM) on one of the processes to verify that the signal + # # handler is executed to cleanup memory. # dims = (200, 1000000) # dt = np.float64 # shm = MPIShared(dims, dt, self.comm) @@ -440,6 +493,7 @@ def test_zero(self): # del temp # import time # time.sleep(60) + # shm.close() # del shm # return @@ -477,14 +531,15 @@ def run(): comm = MPI.COMM_WORLD suite = unittest.TestSuite() - suite.addTest(unittest.makeSuite(LockTest)) - suite.addTest(unittest.makeSuite(ShmemTest)) + suite.addTest(unittest.defaultTestLoader.loadTestsFromTestCase(LockTest)) + suite.addTest(unittest.defaultTestLoader.loadTestsFromTestCase(ShmemTest)) runner = unittest.TextTestRunner() ret = 0 - _ret = runner.run(suite) - if not _ret.wasSuccessful(): - ret += 1 + with exception_guard(comm=comm): + _ret = runner.run(suite) + if not _ret.wasSuccessful(): + ret += 1 if comm is not None: ret = comm.allreduce(ret, op=MPI.SUM) diff --git a/pshmem/utils.py b/pshmem/utils.py index abc1522..72d3a69 100644 --- a/pshmem/utils.py +++ b/pshmem/utils.py @@ -1,17 +1,64 @@ ## -# Copyright (c) 2017-2024, all rights reserved. Use of this source code +# Copyright (c) 2017-2025, all rights reserved. Use of this source code # is governed by a BSD license that can be found in the top-level # LICENSE file. ## import random import sys -# Import for monkey patching resource tracker -from multiprocessing import resource_tracker +import threading +import time +import traceback +from contextlib import contextmanager +from multiprocessing import resource_tracker as _mprt +from multiprocessing import shared_memory as _mpshm import numpy as np +"""Backport the new `track` option from python 3.13 to older versions +More details at: https://github.com/python/cpython/issues/82300 +""" +if sys.version_info >= (3, 13): + SharedMemory = _mpshm.SharedMemory +else: + class SharedMemory(_mpshm.SharedMemory): + __lock = threading.Lock() + + def __init__( + self, name=None, create=False, size=0, *, track=True + ) -> None: + self._track = track + + # if tracking, normal init will suffice + if track: + return super().__init__(name=name, create=create, size=size) + + # lock so that other threads don't attempt to use the + # register function during this time + with self.__lock: + # temporarily disable registration during initialization + orig_register = _mprt.register + _mprt.register = self.__tmp_register + + # initialize; ensure original register function is + # re-instated + try: + super().__init__(name=name, create=create, size=size) + finally: + _mprt.register = orig_register + + @staticmethod + def __tmp_register(*args, **kwargs) -> None: + return + + def unlink(self) -> None: + if _mpshm._USE_POSIX and self._name: + _mpshm._posixshmem.shm_unlink(self._name) + if self._track: + _mprt.unregister(self._name, "shared_memory") + + def mpi_data_type(comm, dt): """Helper function to return the byte size and MPI datatype. @@ -66,23 +113,34 @@ def random_shm_key(): return random.randint(min_val, max_val) -def remove_shm_from_resource_tracker(): - """Monkey-patch multiprocessing.resource_tracker so SharedMemory won't be tracked - - More details at: https://bugs.python.org/issue38119 - """ - - def fix_register(name, rtype): - if rtype == "shared_memory": - return - return resource_tracker._resource_tracker.register(self, name, rtype) - resource_tracker.register = fix_register +@contextmanager +def exception_guard(comm=None, timeout=5): + """Ensure if one MPI process raises an un-caught exception, the program shuts down. - def fix_unregister(name, rtype): - if rtype == "shared_memory": - return - return resource_tracker._resource_tracker.unregister(self, name, rtype) - resource_tracker.unregister = fix_unregister + Args: + comm (mpi4py.MPI.Comm): The MPI communicator or None. + timeout (int): The number of seconds to wait before aborting all processes - if "shared_memory" in resource_tracker._CLEANUP_FUNCS: - del resource_tracker._CLEANUP_FUNCS["shared_memory"] + """ + rank = 0 if comm is None else comm.rank + try: + yield + except Exception: + # Note that the intention of this function is to handle *any* exception. + # The typical use case is to wrap main() and ensure that the job exits + # cleanly. + exc_type, exc_value, exc_traceback = sys.exc_info() + lines = traceback.format_exception(exc_type, exc_value, exc_traceback) + lines = [f"Proc {rank}: {x}" for x in lines] + msg = "".join(lines) + print(msg, flush=True) + # kills the job + if comm is None or comm.size == 1: + # Raising the exception allows for debugging + raise + else: + if comm.size > 1: + # gives other processes a bit of time to see whether + # they encounter the same error + time.sleep(timeout) + comm.Abort(1) diff --git a/pyproject.toml b/pyproject.toml new file mode 100644 index 0000000..2d8f8ea --- /dev/null +++ b/pyproject.toml @@ -0,0 +1,45 @@ +[build-system] +requires = ["setuptools", "versioneer[toml]==0.29"] +build-backend = "setuptools.build_meta" + +[project] +name = "pshmem" +dynamic = ["version"] +description = "Parallel shared memory and locking with MPI" +readme = "README.md" +maintainers = [ + { name = "Theodore Kisner", email = "tskisner.public@gmail.com" }, +] +requires-python = ">=3.9" +dependencies = [ + "numpy", +] +classifiers = [ + "Development Status :: 4 - Beta", + "License :: OSI Approved :: BSD License", + "Programming Language :: Python", + "Programming Language :: Python :: 3", + "Programming Language :: Python :: 3.9", + "Programming Language :: Python :: 3.10", + "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", + "Programming Language :: Python :: 3.13", + "Topic :: Software Development :: Libraries", +] + +[project.optional-dependencies] +mpi = [ + "mpi4py", +] + +[project.urls] +"Source" = "https://github.com/tskisner/pshmem" +"Issue Tracker" = "https://github.com/tskisner/pshmem/issues" + +[tool.versioneer] +VCS = "git" +style = "pep440" +versionfile_source = "pshmem/_version.py" +versionfile_build = "pshmem/_version.py" +tag_prefix = "" +parentdir_prefix = "pshmem-" diff --git a/setup.cfg b/setup.cfg deleted file mode 100644 index f5253f5..0000000 --- a/setup.cfg +++ /dev/null @@ -1,12 +0,0 @@ - -# See the docstring in versioneer.py for instructions. Note that you must -# re-run 'versioneer.py setup' after changing this section, and commit the -# resulting files. - -[versioneer] -VCS = git -style = pep440 -versionfile_source = pshmem/_version.py -versionfile_build = pshmem/_version.py -tag_prefix = -parentdir_prefix = pshmem- diff --git a/setup.py b/setup.py index c943ae8..0d3cede 100644 --- a/setup.py +++ b/setup.py @@ -1,44 +1,7 @@ -import os -import sys - -from setuptools import find_packages, setup - +import setuptools import versioneer - -def readme(): - with open("README.md") as f: - return f.read() - - -setup( - name="pshmem", - provides="pshmem", - version=versioneer.get_version(), - description="Parallel shared memory and locking with MPI", - long_description=readme(), - long_description_content_type="text/markdown", - author="Theodore Kisner", - author_email="work@theodorekisner.com", - url="https://github.com/tskisner/pshmem", +setuptools.setup( packages=["pshmem"], - scripts=None, - license="BSD", - python_requires=">=3.8.0", - install_requires=["numpy"], - extras_require={"mpi": ["mpi4py>=3.0"]}, - cmdclass=versioneer.get_cmdclass(), - classifiers=[ - "Development Status :: 4 - Beta", - "Environment :: Console", - "Intended Audience :: Science/Research", - "License :: OSI Approved :: BSD License", - "Operating System :: POSIX", - "Programming Language :: Python :: 3.8", - "Programming Language :: Python :: 3.9", - "Programming Language :: Python :: 3.10", - "Programming Language :: Python :: 3.11", - "Programming Language :: Python :: 3.12", - "Topic :: Utilities", - ], + version=versioneer.get_version(), ) diff --git a/test_scripts/install_deps_cibuildwheel_linux.sh b/test_scripts/install_deps_cibuildwheel_linux.sh deleted file mode 100755 index 1a9fb29..0000000 --- a/test_scripts/install_deps_cibuildwheel_linux.sh +++ /dev/null @@ -1,13 +0,0 @@ -#!/bin/bash -# -# This installs mpich using yum and then installs mpi4py with pip. -# - -# Location of this script -pushd $(dirname $0) >/dev/null 2>&1 -topdir=$(pwd) -popd >/dev/null 2>&1 - -# Install mpich -yum -y update -yum -y install mpich-3.2-devel.x86_64 mpich-3.2-autoload.x86_64 diff --git a/test_scripts/install_deps_cibuildwheel_macos.sh b/test_scripts/install_deps_cibuildwheel_macos.sh deleted file mode 100755 index 204f8c7..0000000 --- a/test_scripts/install_deps_cibuildwheel_macos.sh +++ /dev/null @@ -1,12 +0,0 @@ -#!/bin/bash -# -# This installs mpich using homebrew and then installs mpi4py with pip. -# - -# Location of this script -pushd $(dirname $0) >/dev/null 2>&1 -topdir=$(pwd) -popd >/dev/null 2>&1 - -# Install mpich -brew install mpich diff --git a/test_scripts/install_deps_github_macos.sh b/test_scripts/install_deps_github_macos.sh deleted file mode 100755 index d4a1eea..0000000 --- a/test_scripts/install_deps_github_macos.sh +++ /dev/null @@ -1,19 +0,0 @@ -#!/bin/bash -# -# This installs mpich using homebrew and then installs mpi4py with pip. -# - -# Location of this script -pushd $(dirname $0) >/dev/null 2>&1 -topdir=$(pwd) -popd >/dev/null 2>&1 - -# Install mpich -brew install mpich - -# Install mpi4py -pip3 install setuptools -pip3 install wheel -pip3 install numpy -pip3 install posix_ipc -pip3 install mpi4py diff --git a/test_scripts/install_deps_github_ubuntu.sh b/test_scripts/install_deps_github_ubuntu.sh deleted file mode 100755 index 099cf0e..0000000 --- a/test_scripts/install_deps_github_ubuntu.sh +++ /dev/null @@ -1,20 +0,0 @@ -#!/bin/bash -# -# This installs mpich using apt and then installs mpi4py with pip. -# - -# Location of this script -pushd $(dirname $0) >/dev/null 2>&1 -topdir=$(pwd) -popd >/dev/null 2>&1 - -# Install mpich -sudo apt-get -y update -sudo apt-get install -y build-essential libmpich-dev - -# Install mpi4py -pip3 install setuptools -pip3 install wheel -pip3 install numpy -pip3 install posix_ipc -pip3 install mpi4py diff --git a/test_scripts/run_cori_haswell.slurm b/test_scripts/run_cori_haswell.slurm deleted file mode 100644 index ea37319..0000000 --- a/test_scripts/run_cori_haswell.slurm +++ /dev/null @@ -1,32 +0,0 @@ -#!/bin/bash -l - -#SBATCH --partition=debug -#SBATCH --constraint=haswell -#SBATCH --nodes=2 -#SBATCH --time=00:10:00 -#SBATCH --job-name=pshmem - -# Nodes used in the job -NODES=${SLURM_JOB_NUM_NODES} - -# set procs and threads -NODE_SLOTS=64 -NODE_PROC=32 -PROC_THREADS=1 -PROC_DEPTH=$(( NODE_SLOTS / NODE_PROC )) - -# total number of processes on all nodes -NPROC=$(( NODES * NODE_PROC )) - -echo "Using ${NODES} node(s), which have ${NODE_SLOTS} thread slots each." -echo "Starting ${NODE_PROC} process(es) per node (${NPROC} total), each with ${PROC_THREADS} OpenMP threads." - -export OMP_NUM_THREADS=${PROC_THREADS} -export OMP_PROC_BIND=spread -export OMP_PLACES=threads - -com="srun -n ${NPROC} -N ${NODES} -c ${PROC_DEPTH} --cpu_bind=cores \ -python3 -c 'import pshmem; pshmem.test()'" - -echo ${com} -eval ${com} diff --git a/test_scripts/run_cori_knl.slurm b/test_scripts/run_perlmutter.slurm similarity index 66% rename from test_scripts/run_cori_knl.slurm rename to test_scripts/run_perlmutter.slurm index 16d6989..bb5b508 100644 --- a/test_scripts/run_cori_knl.slurm +++ b/test_scripts/run_perlmutter.slurm @@ -1,18 +1,17 @@ #!/bin/bash -l -#SBATCH --partition=debug -#SBATCH --constraint=knl,quad,cache +#SBATCH --qos=debug +#SBATCH --constraint=cpu #SBATCH --nodes=2 -#SBATCH --core-spec=4 #SBATCH --time=00:10:00 #SBATCH --job-name=pshmem -# Nodes used in the job +# nodes used by this job NODES=${SLURM_JOB_NUM_NODES} # set procs and threads NODE_SLOTS=256 -NODE_PROC=64 +NODE_PROC=128 PROC_THREADS=1 PROC_DEPTH=$(( NODE_SLOTS / NODE_PROC )) @@ -26,8 +25,11 @@ export OMP_NUM_THREADS=${PROC_THREADS} export OMP_PROC_BIND=spread export OMP_PLACES=threads -com="srun -n ${NPROC} -N ${NODES} -c ${PROC_DEPTH} --cpu_bind=cores \ -python3 -c 'import pshmem; pshmem.test()'" +# The launching command and options +launch_str="srun -n ${NPROC} -N ${NODES} -c ${PROC_DEPTH} --cpu_bind=cores" +# Run the unit test suite +com="${launch_str} python3 -c 'import pshmem.test; pshmem.test.run()'" echo ${com} eval ${com} + diff --git a/test_scripts/test_cibuild.sh b/test_scripts/test_cibuild.sh deleted file mode 100755 index 725aaee..0000000 --- a/test_scripts/test_cibuild.sh +++ /dev/null @@ -1,15 +0,0 @@ -#!/bin/bash - - -docker run \ --it \ --v $(pwd):/home/pshmem \ -quay.io/pypa/manylinux2014_x86_64:latest \ -/bin/bash - -# export PATH=/opt/python/cp38-cp38/bin:${PATH} -# python3 -m pip install --upgrade pip -# yum -y update -# yum -y install mpich-3.2-devel.x86_64 mpich-3.2-autoload.x86_64 -# source /etc/profile.d/modules.sh -# source /etc/profile.d/mpich-3.2-x86_64.sh diff --git a/versioneer.py b/versioneer.py deleted file mode 100644 index 1e3753e..0000000 --- a/versioneer.py +++ /dev/null @@ -1,2277 +0,0 @@ - -# Version: 0.29 - -"""The Versioneer - like a rocketeer, but for versions. - -The Versioneer -============== - -* like a rocketeer, but for versions! -* https://github.com/python-versioneer/python-versioneer -* Brian Warner -* License: Public Domain (Unlicense) -* Compatible with: Python 3.7, 3.8, 3.9, 3.10, 3.11 and pypy3 -* [![Latest Version][pypi-image]][pypi-url] -* [![Build Status][travis-image]][travis-url] - -This is a tool for managing a recorded version number in setuptools-based -python projects. The goal is to remove the tedious and error-prone "update -the embedded version string" step from your release process. Making a new -release should be as easy as recording a new tag in your version-control -system, and maybe making new tarballs. - - -## Quick Install - -Versioneer provides two installation modes. The "classic" vendored mode installs -a copy of versioneer into your repository. The experimental build-time dependency mode -is intended to allow you to skip this step and simplify the process of upgrading. - -### Vendored mode - -* `pip install versioneer` to somewhere in your $PATH - * A [conda-forge recipe](https://github.com/conda-forge/versioneer-feedstock) is - available, so you can also use `conda install -c conda-forge versioneer` -* add a `[tool.versioneer]` section to your `pyproject.toml` or a - `[versioneer]` section to your `setup.cfg` (see [Install](INSTALL.md)) - * Note that you will need to add `tomli; python_version < "3.11"` to your - build-time dependencies if you use `pyproject.toml` -* run `versioneer install --vendor` in your source tree, commit the results -* verify version information with `python setup.py version` - -### Build-time dependency mode - -* `pip install versioneer` to somewhere in your $PATH - * A [conda-forge recipe](https://github.com/conda-forge/versioneer-feedstock) is - available, so you can also use `conda install -c conda-forge versioneer` -* add a `[tool.versioneer]` section to your `pyproject.toml` or a - `[versioneer]` section to your `setup.cfg` (see [Install](INSTALL.md)) -* add `versioneer` (with `[toml]` extra, if configuring in `pyproject.toml`) - to the `requires` key of the `build-system` table in `pyproject.toml`: - ```toml - [build-system] - requires = ["setuptools", "versioneer[toml]"] - build-backend = "setuptools.build_meta" - ``` -* run `versioneer install --no-vendor` in your source tree, commit the results -* verify version information with `python setup.py version` - -## Version Identifiers - -Source trees come from a variety of places: - -* a version-control system checkout (mostly used by developers) -* a nightly tarball, produced by build automation -* a snapshot tarball, produced by a web-based VCS browser, like github's - "tarball from tag" feature -* a release tarball, produced by "setup.py sdist", distributed through PyPI - -Within each source tree, the version identifier (either a string or a number, -this tool is format-agnostic) can come from a variety of places: - -* ask the VCS tool itself, e.g. "git describe" (for checkouts), which knows - about recent "tags" and an absolute revision-id -* the name of the directory into which the tarball was unpacked -* an expanded VCS keyword ($Id$, etc) -* a `_version.py` created by some earlier build step - -For released software, the version identifier is closely related to a VCS -tag. Some projects use tag names that include more than just the version -string (e.g. "myproject-1.2" instead of just "1.2"), in which case the tool -needs to strip the tag prefix to extract the version identifier. For -unreleased software (between tags), the version identifier should provide -enough information to help developers recreate the same tree, while also -giving them an idea of roughly how old the tree is (after version 1.2, before -version 1.3). Many VCS systems can report a description that captures this, -for example `git describe --tags --dirty --always` reports things like -"0.7-1-g574ab98-dirty" to indicate that the checkout is one revision past the -0.7 tag, has a unique revision id of "574ab98", and is "dirty" (it has -uncommitted changes). - -The version identifier is used for multiple purposes: - -* to allow the module to self-identify its version: `myproject.__version__` -* to choose a name and prefix for a 'setup.py sdist' tarball - -## Theory of Operation - -Versioneer works by adding a special `_version.py` file into your source -tree, where your `__init__.py` can import it. This `_version.py` knows how to -dynamically ask the VCS tool for version information at import time. - -`_version.py` also contains `$Revision$` markers, and the installation -process marks `_version.py` to have this marker rewritten with a tag name -during the `git archive` command. As a result, generated tarballs will -contain enough information to get the proper version. - -To allow `setup.py` to compute a version too, a `versioneer.py` is added to -the top level of your source tree, next to `setup.py` and the `setup.cfg` -that configures it. This overrides several distutils/setuptools commands to -compute the version when invoked, and changes `setup.py build` and `setup.py -sdist` to replace `_version.py` with a small static file that contains just -the generated version data. - -## Installation - -See [INSTALL.md](./INSTALL.md) for detailed installation instructions. - -## Version-String Flavors - -Code which uses Versioneer can learn about its version string at runtime by -importing `_version` from your main `__init__.py` file and running the -`get_versions()` function. From the "outside" (e.g. in `setup.py`), you can -import the top-level `versioneer.py` and run `get_versions()`. - -Both functions return a dictionary with different flavors of version -information: - -* `['version']`: A condensed version string, rendered using the selected - style. This is the most commonly used value for the project's version - string. The default "pep440" style yields strings like `0.11`, - `0.11+2.g1076c97`, or `0.11+2.g1076c97.dirty`. See the "Styles" section - below for alternative styles. - -* `['full-revisionid']`: detailed revision identifier. For Git, this is the - full SHA1 commit id, e.g. "1076c978a8d3cfc70f408fe5974aa6c092c949ac". - -* `['date']`: Date and time of the latest `HEAD` commit. For Git, it is the - commit date in ISO 8601 format. This will be None if the date is not - available. - -* `['dirty']`: a boolean, True if the tree has uncommitted changes. Note that - this is only accurate if run in a VCS checkout, otherwise it is likely to - be False or None - -* `['error']`: if the version string could not be computed, this will be set - to a string describing the problem, otherwise it will be None. It may be - useful to throw an exception in setup.py if this is set, to avoid e.g. - creating tarballs with a version string of "unknown". - -Some variants are more useful than others. Including `full-revisionid` in a -bug report should allow developers to reconstruct the exact code being tested -(or indicate the presence of local changes that should be shared with the -developers). `version` is suitable for display in an "about" box or a CLI -`--version` output: it can be easily compared against release notes and lists -of bugs fixed in various releases. - -The installer adds the following text to your `__init__.py` to place a basic -version in `YOURPROJECT.__version__`: - - from ._version import get_versions - __version__ = get_versions()['version'] - del get_versions - -## Styles - -The setup.cfg `style=` configuration controls how the VCS information is -rendered into a version string. - -The default style, "pep440", produces a PEP440-compliant string, equal to the -un-prefixed tag name for actual releases, and containing an additional "local -version" section with more detail for in-between builds. For Git, this is -TAG[+DISTANCE.gHEX[.dirty]] , using information from `git describe --tags ---dirty --always`. For example "0.11+2.g1076c97.dirty" indicates that the -tree is like the "1076c97" commit but has uncommitted changes (".dirty"), and -that this commit is two revisions ("+2") beyond the "0.11" tag. For released -software (exactly equal to a known tag), the identifier will only contain the -stripped tag, e.g. "0.11". - -Other styles are available. See [details.md](details.md) in the Versioneer -source tree for descriptions. - -## Debugging - -Versioneer tries to avoid fatal errors: if something goes wrong, it will tend -to return a version of "0+unknown". To investigate the problem, run `setup.py -version`, which will run the version-lookup code in a verbose mode, and will -display the full contents of `get_versions()` (including the `error` string, -which may help identify what went wrong). - -## Known Limitations - -Some situations are known to cause problems for Versioneer. This details the -most significant ones. More can be found on Github -[issues page](https://github.com/python-versioneer/python-versioneer/issues). - -### Subprojects - -Versioneer has limited support for source trees in which `setup.py` is not in -the root directory (e.g. `setup.py` and `.git/` are *not* siblings). The are -two common reasons why `setup.py` might not be in the root: - -* Source trees which contain multiple subprojects, such as - [Buildbot](https://github.com/buildbot/buildbot), which contains both - "master" and "slave" subprojects, each with their own `setup.py`, - `setup.cfg`, and `tox.ini`. Projects like these produce multiple PyPI - distributions (and upload multiple independently-installable tarballs). -* Source trees whose main purpose is to contain a C library, but which also - provide bindings to Python (and perhaps other languages) in subdirectories. - -Versioneer will look for `.git` in parent directories, and most operations -should get the right version string. However `pip` and `setuptools` have bugs -and implementation details which frequently cause `pip install .` from a -subproject directory to fail to find a correct version string (so it usually -defaults to `0+unknown`). - -`pip install --editable .` should work correctly. `setup.py install` might -work too. - -Pip-8.1.1 is known to have this problem, but hopefully it will get fixed in -some later version. - -[Bug #38](https://github.com/python-versioneer/python-versioneer/issues/38) is tracking -this issue. The discussion in -[PR #61](https://github.com/python-versioneer/python-versioneer/pull/61) describes the -issue from the Versioneer side in more detail. -[pip PR#3176](https://github.com/pypa/pip/pull/3176) and -[pip PR#3615](https://github.com/pypa/pip/pull/3615) contain work to improve -pip to let Versioneer work correctly. - -Versioneer-0.16 and earlier only looked for a `.git` directory next to the -`setup.cfg`, so subprojects were completely unsupported with those releases. - -### Editable installs with setuptools <= 18.5 - -`setup.py develop` and `pip install --editable .` allow you to install a -project into a virtualenv once, then continue editing the source code (and -test) without re-installing after every change. - -"Entry-point scripts" (`setup(entry_points={"console_scripts": ..})`) are a -convenient way to specify executable scripts that should be installed along -with the python package. - -These both work as expected when using modern setuptools. When using -setuptools-18.5 or earlier, however, certain operations will cause -`pkg_resources.DistributionNotFound` errors when running the entrypoint -script, which must be resolved by re-installing the package. This happens -when the install happens with one version, then the egg_info data is -regenerated while a different version is checked out. Many setup.py commands -cause egg_info to be rebuilt (including `sdist`, `wheel`, and installing into -a different virtualenv), so this can be surprising. - -[Bug #83](https://github.com/python-versioneer/python-versioneer/issues/83) describes -this one, but upgrading to a newer version of setuptools should probably -resolve it. - - -## Updating Versioneer - -To upgrade your project to a new release of Versioneer, do the following: - -* install the new Versioneer (`pip install -U versioneer` or equivalent) -* edit `setup.cfg` and `pyproject.toml`, if necessary, - to include any new configuration settings indicated by the release notes. - See [UPGRADING](./UPGRADING.md) for details. -* re-run `versioneer install --[no-]vendor` in your source tree, to replace - `SRC/_version.py` -* commit any changed files - -## Future Directions - -This tool is designed to make it easily extended to other version-control -systems: all VCS-specific components are in separate directories like -src/git/ . The top-level `versioneer.py` script is assembled from these -components by running make-versioneer.py . In the future, make-versioneer.py -will take a VCS name as an argument, and will construct a version of -`versioneer.py` that is specific to the given VCS. It might also take the -configuration arguments that are currently provided manually during -installation by editing setup.py . Alternatively, it might go the other -direction and include code from all supported VCS systems, reducing the -number of intermediate scripts. - -## Similar projects - -* [setuptools_scm](https://github.com/pypa/setuptools_scm/) - a non-vendored build-time - dependency -* [minver](https://github.com/jbweston/miniver) - a lightweight reimplementation of - versioneer -* [versioningit](https://github.com/jwodder/versioningit) - a PEP 518-based setuptools - plugin - -## License - -To make Versioneer easier to embed, all its code is dedicated to the public -domain. The `_version.py` that it creates is also in the public domain. -Specifically, both are released under the "Unlicense", as described in -https://unlicense.org/. - -[pypi-image]: https://img.shields.io/pypi/v/versioneer.svg -[pypi-url]: https://pypi.python.org/pypi/versioneer/ -[travis-image]: -https://img.shields.io/travis/com/python-versioneer/python-versioneer.svg -[travis-url]: https://travis-ci.com/github/python-versioneer/python-versioneer - -""" -# pylint:disable=invalid-name,import-outside-toplevel,missing-function-docstring -# pylint:disable=missing-class-docstring,too-many-branches,too-many-statements -# pylint:disable=raise-missing-from,too-many-lines,too-many-locals,import-error -# pylint:disable=too-few-public-methods,redefined-outer-name,consider-using-with -# pylint:disable=attribute-defined-outside-init,too-many-arguments - -import configparser -import errno -import json -import os -import re -import subprocess -import sys -from pathlib import Path -from typing import Any, Callable, cast, Dict, List, Optional, Tuple, Union -from typing import NoReturn -import functools - -have_tomllib = True -if sys.version_info >= (3, 11): - import tomllib -else: - try: - import tomli as tomllib - except ImportError: - have_tomllib = False - - -class VersioneerConfig: - """Container for Versioneer configuration parameters.""" - - VCS: str - style: str - tag_prefix: str - versionfile_source: str - versionfile_build: Optional[str] - parentdir_prefix: Optional[str] - verbose: Optional[bool] - - -def get_root() -> str: - """Get the project root directory. - - We require that all commands are run from the project root, i.e. the - directory that contains setup.py, setup.cfg, and versioneer.py . - """ - root = os.path.realpath(os.path.abspath(os.getcwd())) - setup_py = os.path.join(root, "setup.py") - pyproject_toml = os.path.join(root, "pyproject.toml") - versioneer_py = os.path.join(root, "versioneer.py") - if not ( - os.path.exists(setup_py) - or os.path.exists(pyproject_toml) - or os.path.exists(versioneer_py) - ): - # allow 'python path/to/setup.py COMMAND' - root = os.path.dirname(os.path.realpath(os.path.abspath(sys.argv[0]))) - setup_py = os.path.join(root, "setup.py") - pyproject_toml = os.path.join(root, "pyproject.toml") - versioneer_py = os.path.join(root, "versioneer.py") - if not ( - os.path.exists(setup_py) - or os.path.exists(pyproject_toml) - or os.path.exists(versioneer_py) - ): - err = ("Versioneer was unable to run the project root directory. " - "Versioneer requires setup.py to be executed from " - "its immediate directory (like 'python setup.py COMMAND'), " - "or in a way that lets it use sys.argv[0] to find the root " - "(like 'python path/to/setup.py COMMAND').") - raise VersioneerBadRootError(err) - try: - # Certain runtime workflows (setup.py install/develop in a setuptools - # tree) execute all dependencies in a single python process, so - # "versioneer" may be imported multiple times, and python's shared - # module-import table will cache the first one. So we can't use - # os.path.dirname(__file__), as that will find whichever - # versioneer.py was first imported, even in later projects. - my_path = os.path.realpath(os.path.abspath(__file__)) - me_dir = os.path.normcase(os.path.splitext(my_path)[0]) - vsr_dir = os.path.normcase(os.path.splitext(versioneer_py)[0]) - if me_dir != vsr_dir and "VERSIONEER_PEP518" not in globals(): - print("Warning: build in %s is using versioneer.py from %s" - % (os.path.dirname(my_path), versioneer_py)) - except NameError: - pass - return root - - -def get_config_from_root(root: str) -> VersioneerConfig: - """Read the project setup.cfg file to determine Versioneer config.""" - # This might raise OSError (if setup.cfg is missing), or - # configparser.NoSectionError (if it lacks a [versioneer] section), or - # configparser.NoOptionError (if it lacks "VCS="). See the docstring at - # the top of versioneer.py for instructions on writing your setup.cfg . - root_pth = Path(root) - pyproject_toml = root_pth / "pyproject.toml" - setup_cfg = root_pth / "setup.cfg" - section: Union[Dict[str, Any], configparser.SectionProxy, None] = None - if pyproject_toml.exists() and have_tomllib: - try: - with open(pyproject_toml, 'rb') as fobj: - pp = tomllib.load(fobj) - section = pp['tool']['versioneer'] - except (tomllib.TOMLDecodeError, KeyError) as e: - print(f"Failed to load config from {pyproject_toml}: {e}") - print("Try to load it from setup.cfg") - if not section: - parser = configparser.ConfigParser() - with open(setup_cfg) as cfg_file: - parser.read_file(cfg_file) - parser.get("versioneer", "VCS") # raise error if missing - - section = parser["versioneer"] - - # `cast`` really shouldn't be used, but its simplest for the - # common VersioneerConfig users at the moment. We verify against - # `None` values elsewhere where it matters - - cfg = VersioneerConfig() - cfg.VCS = section['VCS'] - cfg.style = section.get("style", "") - cfg.versionfile_source = cast(str, section.get("versionfile_source")) - cfg.versionfile_build = section.get("versionfile_build") - cfg.tag_prefix = cast(str, section.get("tag_prefix")) - if cfg.tag_prefix in ("''", '""', None): - cfg.tag_prefix = "" - cfg.parentdir_prefix = section.get("parentdir_prefix") - if isinstance(section, configparser.SectionProxy): - # Make sure configparser translates to bool - cfg.verbose = section.getboolean("verbose") - else: - cfg.verbose = section.get("verbose") - - return cfg - - -class NotThisMethod(Exception): - """Exception raised if a method is not valid for the current scenario.""" - - -# these dictionaries contain VCS-specific tools -LONG_VERSION_PY: Dict[str, str] = {} -HANDLERS: Dict[str, Dict[str, Callable]] = {} - - -def register_vcs_handler(vcs: str, method: str) -> Callable: # decorator - """Create decorator to mark a method as the handler of a VCS.""" - def decorate(f: Callable) -> Callable: - """Store f in HANDLERS[vcs][method].""" - HANDLERS.setdefault(vcs, {})[method] = f - return f - return decorate - - -def run_command( - commands: List[str], - args: List[str], - cwd: Optional[str] = None, - verbose: bool = False, - hide_stderr: bool = False, - env: Optional[Dict[str, str]] = None, -) -> Tuple[Optional[str], Optional[int]]: - """Call the given command(s).""" - assert isinstance(commands, list) - process = None - - popen_kwargs: Dict[str, Any] = {} - if sys.platform == "win32": - # This hides the console window if pythonw.exe is used - startupinfo = subprocess.STARTUPINFO() - startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW - popen_kwargs["startupinfo"] = startupinfo - - for command in commands: - try: - dispcmd = str([command] + args) - # remember shell=False, so use git.cmd on windows, not just git - process = subprocess.Popen([command] + args, cwd=cwd, env=env, - stdout=subprocess.PIPE, - stderr=(subprocess.PIPE if hide_stderr - else None), **popen_kwargs) - break - except OSError as e: - if e.errno == errno.ENOENT: - continue - if verbose: - print("unable to run %s" % dispcmd) - print(e) - return None, None - else: - if verbose: - print("unable to find command, tried %s" % (commands,)) - return None, None - stdout = process.communicate()[0].strip().decode() - if process.returncode != 0: - if verbose: - print("unable to run %s (error)" % dispcmd) - print("stdout was %s" % stdout) - return None, process.returncode - return stdout, process.returncode - - -LONG_VERSION_PY['git'] = r''' -# This file helps to compute a version number in source trees obtained from -# git-archive tarball (such as those provided by githubs download-from-tag -# feature). Distribution tarballs (built by setup.py sdist) and build -# directories (produced by setup.py build) will contain a much shorter file -# that just contains the computed version number. - -# This file is released into the public domain. -# Generated by versioneer-0.29 -# https://github.com/python-versioneer/python-versioneer - -"""Git implementation of _version.py.""" - -import errno -import os -import re -import subprocess -import sys -from typing import Any, Callable, Dict, List, Optional, Tuple -import functools - - -def get_keywords() -> Dict[str, str]: - """Get the keywords needed to look up the version information.""" - # these strings will be replaced by git during git-archive. - # setup.py/versioneer.py will grep for the variable names, so they must - # each be defined on a line of their own. _version.py will just call - # get_keywords(). - git_refnames = "%(DOLLAR)sFormat:%%d%(DOLLAR)s" - git_full = "%(DOLLAR)sFormat:%%H%(DOLLAR)s" - git_date = "%(DOLLAR)sFormat:%%ci%(DOLLAR)s" - keywords = {"refnames": git_refnames, "full": git_full, "date": git_date} - return keywords - - -class VersioneerConfig: - """Container for Versioneer configuration parameters.""" - - VCS: str - style: str - tag_prefix: str - parentdir_prefix: str - versionfile_source: str - verbose: bool - - -def get_config() -> VersioneerConfig: - """Create, populate and return the VersioneerConfig() object.""" - # these strings are filled in when 'setup.py versioneer' creates - # _version.py - cfg = VersioneerConfig() - cfg.VCS = "git" - cfg.style = "%(STYLE)s" - cfg.tag_prefix = "%(TAG_PREFIX)s" - cfg.parentdir_prefix = "%(PARENTDIR_PREFIX)s" - cfg.versionfile_source = "%(VERSIONFILE_SOURCE)s" - cfg.verbose = False - return cfg - - -class NotThisMethod(Exception): - """Exception raised if a method is not valid for the current scenario.""" - - -LONG_VERSION_PY: Dict[str, str] = {} -HANDLERS: Dict[str, Dict[str, Callable]] = {} - - -def register_vcs_handler(vcs: str, method: str) -> Callable: # decorator - """Create decorator to mark a method as the handler of a VCS.""" - def decorate(f: Callable) -> Callable: - """Store f in HANDLERS[vcs][method].""" - if vcs not in HANDLERS: - HANDLERS[vcs] = {} - HANDLERS[vcs][method] = f - return f - return decorate - - -def run_command( - commands: List[str], - args: List[str], - cwd: Optional[str] = None, - verbose: bool = False, - hide_stderr: bool = False, - env: Optional[Dict[str, str]] = None, -) -> Tuple[Optional[str], Optional[int]]: - """Call the given command(s).""" - assert isinstance(commands, list) - process = None - - popen_kwargs: Dict[str, Any] = {} - if sys.platform == "win32": - # This hides the console window if pythonw.exe is used - startupinfo = subprocess.STARTUPINFO() - startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW - popen_kwargs["startupinfo"] = startupinfo - - for command in commands: - try: - dispcmd = str([command] + args) - # remember shell=False, so use git.cmd on windows, not just git - process = subprocess.Popen([command] + args, cwd=cwd, env=env, - stdout=subprocess.PIPE, - stderr=(subprocess.PIPE if hide_stderr - else None), **popen_kwargs) - break - except OSError as e: - if e.errno == errno.ENOENT: - continue - if verbose: - print("unable to run %%s" %% dispcmd) - print(e) - return None, None - else: - if verbose: - print("unable to find command, tried %%s" %% (commands,)) - return None, None - stdout = process.communicate()[0].strip().decode() - if process.returncode != 0: - if verbose: - print("unable to run %%s (error)" %% dispcmd) - print("stdout was %%s" %% stdout) - return None, process.returncode - return stdout, process.returncode - - -def versions_from_parentdir( - parentdir_prefix: str, - root: str, - verbose: bool, -) -> Dict[str, Any]: - """Try to determine the version from the parent directory name. - - Source tarballs conventionally unpack into a directory that includes both - the project name and a version string. We will also support searching up - two directory levels for an appropriately named parent directory - """ - rootdirs = [] - - for _ in range(3): - dirname = os.path.basename(root) - if dirname.startswith(parentdir_prefix): - return {"version": dirname[len(parentdir_prefix):], - "full-revisionid": None, - "dirty": False, "error": None, "date": None} - rootdirs.append(root) - root = os.path.dirname(root) # up a level - - if verbose: - print("Tried directories %%s but none started with prefix %%s" %% - (str(rootdirs), parentdir_prefix)) - raise NotThisMethod("rootdir doesn't start with parentdir_prefix") - - -@register_vcs_handler("git", "get_keywords") -def git_get_keywords(versionfile_abs: str) -> Dict[str, str]: - """Extract version information from the given file.""" - # the code embedded in _version.py can just fetch the value of these - # keywords. When used from setup.py, we don't want to import _version.py, - # so we do it with a regexp instead. This function is not used from - # _version.py. - keywords: Dict[str, str] = {} - try: - with open(versionfile_abs, "r") as fobj: - for line in fobj: - if line.strip().startswith("git_refnames ="): - mo = re.search(r'=\s*"(.*)"', line) - if mo: - keywords["refnames"] = mo.group(1) - if line.strip().startswith("git_full ="): - mo = re.search(r'=\s*"(.*)"', line) - if mo: - keywords["full"] = mo.group(1) - if line.strip().startswith("git_date ="): - mo = re.search(r'=\s*"(.*)"', line) - if mo: - keywords["date"] = mo.group(1) - except OSError: - pass - return keywords - - -@register_vcs_handler("git", "keywords") -def git_versions_from_keywords( - keywords: Dict[str, str], - tag_prefix: str, - verbose: bool, -) -> Dict[str, Any]: - """Get version information from git keywords.""" - if "refnames" not in keywords: - raise NotThisMethod("Short version file found") - date = keywords.get("date") - if date is not None: - # Use only the last line. Previous lines may contain GPG signature - # information. - date = date.splitlines()[-1] - - # git-2.2.0 added "%%cI", which expands to an ISO-8601 -compliant - # datestamp. However we prefer "%%ci" (which expands to an "ISO-8601 - # -like" string, which we must then edit to make compliant), because - # it's been around since git-1.5.3, and it's too difficult to - # discover which version we're using, or to work around using an - # older one. - date = date.strip().replace(" ", "T", 1).replace(" ", "", 1) - refnames = keywords["refnames"].strip() - if refnames.startswith("$Format"): - if verbose: - print("keywords are unexpanded, not using") - raise NotThisMethod("unexpanded keywords, not a git-archive tarball") - refs = {r.strip() for r in refnames.strip("()").split(",")} - # starting in git-1.8.3, tags are listed as "tag: foo-1.0" instead of - # just "foo-1.0". If we see a "tag: " prefix, prefer those. - TAG = "tag: " - tags = {r[len(TAG):] for r in refs if r.startswith(TAG)} - if not tags: - # Either we're using git < 1.8.3, or there really are no tags. We use - # a heuristic: assume all version tags have a digit. The old git %%d - # expansion behaves like git log --decorate=short and strips out the - # refs/heads/ and refs/tags/ prefixes that would let us distinguish - # between branches and tags. By ignoring refnames without digits, we - # filter out many common branch names like "release" and - # "stabilization", as well as "HEAD" and "master". - tags = {r for r in refs if re.search(r'\d', r)} - if verbose: - print("discarding '%%s', no digits" %% ",".join(refs - tags)) - if verbose: - print("likely tags: %%s" %% ",".join(sorted(tags))) - for ref in sorted(tags): - # sorting will prefer e.g. "2.0" over "2.0rc1" - if ref.startswith(tag_prefix): - r = ref[len(tag_prefix):] - # Filter out refs that exactly match prefix or that don't start - # with a number once the prefix is stripped (mostly a concern - # when prefix is '') - if not re.match(r'\d', r): - continue - if verbose: - print("picking %%s" %% r) - return {"version": r, - "full-revisionid": keywords["full"].strip(), - "dirty": False, "error": None, - "date": date} - # no suitable tags, so version is "0+unknown", but full hex is still there - if verbose: - print("no suitable tags, using unknown + full revision id") - return {"version": "0+unknown", - "full-revisionid": keywords["full"].strip(), - "dirty": False, "error": "no suitable tags", "date": None} - - -@register_vcs_handler("git", "pieces_from_vcs") -def git_pieces_from_vcs( - tag_prefix: str, - root: str, - verbose: bool, - runner: Callable = run_command -) -> Dict[str, Any]: - """Get version from 'git describe' in the root of the source tree. - - This only gets called if the git-archive 'subst' keywords were *not* - expanded, and _version.py hasn't already been rewritten with a short - version string, meaning we're inside a checked out source tree. - """ - GITS = ["git"] - if sys.platform == "win32": - GITS = ["git.cmd", "git.exe"] - - # GIT_DIR can interfere with correct operation of Versioneer. - # It may be intended to be passed to the Versioneer-versioned project, - # but that should not change where we get our version from. - env = os.environ.copy() - env.pop("GIT_DIR", None) - runner = functools.partial(runner, env=env) - - _, rc = runner(GITS, ["rev-parse", "--git-dir"], cwd=root, - hide_stderr=not verbose) - if rc != 0: - if verbose: - print("Directory %%s not under git control" %% root) - raise NotThisMethod("'git rev-parse --git-dir' returned error") - - # if there is a tag matching tag_prefix, this yields TAG-NUM-gHEX[-dirty] - # if there isn't one, this yields HEX[-dirty] (no NUM) - describe_out, rc = runner(GITS, [ - "describe", "--tags", "--dirty", "--always", "--long", - "--match", f"{tag_prefix}[[:digit:]]*" - ], cwd=root) - # --long was added in git-1.5.5 - if describe_out is None: - raise NotThisMethod("'git describe' failed") - describe_out = describe_out.strip() - full_out, rc = runner(GITS, ["rev-parse", "HEAD"], cwd=root) - if full_out is None: - raise NotThisMethod("'git rev-parse' failed") - full_out = full_out.strip() - - pieces: Dict[str, Any] = {} - pieces["long"] = full_out - pieces["short"] = full_out[:7] # maybe improved later - pieces["error"] = None - - branch_name, rc = runner(GITS, ["rev-parse", "--abbrev-ref", "HEAD"], - cwd=root) - # --abbrev-ref was added in git-1.6.3 - if rc != 0 or branch_name is None: - raise NotThisMethod("'git rev-parse --abbrev-ref' returned error") - branch_name = branch_name.strip() - - if branch_name == "HEAD": - # If we aren't exactly on a branch, pick a branch which represents - # the current commit. If all else fails, we are on a branchless - # commit. - branches, rc = runner(GITS, ["branch", "--contains"], cwd=root) - # --contains was added in git-1.5.4 - if rc != 0 or branches is None: - raise NotThisMethod("'git branch --contains' returned error") - branches = branches.split("\n") - - # Remove the first line if we're running detached - if "(" in branches[0]: - branches.pop(0) - - # Strip off the leading "* " from the list of branches. - branches = [branch[2:] for branch in branches] - if "master" in branches: - branch_name = "master" - elif not branches: - branch_name = None - else: - # Pick the first branch that is returned. Good or bad. - branch_name = branches[0] - - pieces["branch"] = branch_name - - # parse describe_out. It will be like TAG-NUM-gHEX[-dirty] or HEX[-dirty] - # TAG might have hyphens. - git_describe = describe_out - - # look for -dirty suffix - dirty = git_describe.endswith("-dirty") - pieces["dirty"] = dirty - if dirty: - git_describe = git_describe[:git_describe.rindex("-dirty")] - - # now we have TAG-NUM-gHEX or HEX - - if "-" in git_describe: - # TAG-NUM-gHEX - mo = re.search(r'^(.+)-(\d+)-g([0-9a-f]+)$', git_describe) - if not mo: - # unparsable. Maybe git-describe is misbehaving? - pieces["error"] = ("unable to parse git-describe output: '%%s'" - %% describe_out) - return pieces - - # tag - full_tag = mo.group(1) - if not full_tag.startswith(tag_prefix): - if verbose: - fmt = "tag '%%s' doesn't start with prefix '%%s'" - print(fmt %% (full_tag, tag_prefix)) - pieces["error"] = ("tag '%%s' doesn't start with prefix '%%s'" - %% (full_tag, tag_prefix)) - return pieces - pieces["closest-tag"] = full_tag[len(tag_prefix):] - - # distance: number of commits since tag - pieces["distance"] = int(mo.group(2)) - - # commit: short hex revision ID - pieces["short"] = mo.group(3) - - else: - # HEX: no tags - pieces["closest-tag"] = None - out, rc = runner(GITS, ["rev-list", "HEAD", "--left-right"], cwd=root) - pieces["distance"] = len(out.split()) # total number of commits - - # commit date: see ISO-8601 comment in git_versions_from_keywords() - date = runner(GITS, ["show", "-s", "--format=%%ci", "HEAD"], cwd=root)[0].strip() - # Use only the last line. Previous lines may contain GPG signature - # information. - date = date.splitlines()[-1] - pieces["date"] = date.strip().replace(" ", "T", 1).replace(" ", "", 1) - - return pieces - - -def plus_or_dot(pieces: Dict[str, Any]) -> str: - """Return a + if we don't already have one, else return a .""" - if "+" in pieces.get("closest-tag", ""): - return "." - return "+" - - -def render_pep440(pieces: Dict[str, Any]) -> str: - """Build up version string, with post-release "local version identifier". - - Our goal: TAG[+DISTANCE.gHEX[.dirty]] . Note that if you - get a tagged build and then dirty it, you'll get TAG+0.gHEX.dirty - - Exceptions: - 1: no tags. git_describe was just HEX. 0+untagged.DISTANCE.gHEX[.dirty] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - rendered += plus_or_dot(pieces) - rendered += "%%d.g%%s" %% (pieces["distance"], pieces["short"]) - if pieces["dirty"]: - rendered += ".dirty" - else: - # exception #1 - rendered = "0+untagged.%%d.g%%s" %% (pieces["distance"], - pieces["short"]) - if pieces["dirty"]: - rendered += ".dirty" - return rendered - - -def render_pep440_branch(pieces: Dict[str, Any]) -> str: - """TAG[[.dev0]+DISTANCE.gHEX[.dirty]] . - - The ".dev0" means not master branch. Note that .dev0 sorts backwards - (a feature branch will appear "older" than the master branch). - - Exceptions: - 1: no tags. 0[.dev0]+untagged.DISTANCE.gHEX[.dirty] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - if pieces["branch"] != "master": - rendered += ".dev0" - rendered += plus_or_dot(pieces) - rendered += "%%d.g%%s" %% (pieces["distance"], pieces["short"]) - if pieces["dirty"]: - rendered += ".dirty" - else: - # exception #1 - rendered = "0" - if pieces["branch"] != "master": - rendered += ".dev0" - rendered += "+untagged.%%d.g%%s" %% (pieces["distance"], - pieces["short"]) - if pieces["dirty"]: - rendered += ".dirty" - return rendered - - -def pep440_split_post(ver: str) -> Tuple[str, Optional[int]]: - """Split pep440 version string at the post-release segment. - - Returns the release segments before the post-release and the - post-release version number (or -1 if no post-release segment is present). - """ - vc = str.split(ver, ".post") - return vc[0], int(vc[1] or 0) if len(vc) == 2 else None - - -def render_pep440_pre(pieces: Dict[str, Any]) -> str: - """TAG[.postN.devDISTANCE] -- No -dirty. - - Exceptions: - 1: no tags. 0.post0.devDISTANCE - """ - if pieces["closest-tag"]: - if pieces["distance"]: - # update the post release segment - tag_version, post_version = pep440_split_post(pieces["closest-tag"]) - rendered = tag_version - if post_version is not None: - rendered += ".post%%d.dev%%d" %% (post_version + 1, pieces["distance"]) - else: - rendered += ".post0.dev%%d" %% (pieces["distance"]) - else: - # no commits, use the tag as the version - rendered = pieces["closest-tag"] - else: - # exception #1 - rendered = "0.post0.dev%%d" %% pieces["distance"] - return rendered - - -def render_pep440_post(pieces: Dict[str, Any]) -> str: - """TAG[.postDISTANCE[.dev0]+gHEX] . - - The ".dev0" means dirty. Note that .dev0 sorts backwards - (a dirty tree will appear "older" than the corresponding clean one), - but you shouldn't be releasing software with -dirty anyways. - - Exceptions: - 1: no tags. 0.postDISTANCE[.dev0] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - rendered += ".post%%d" %% pieces["distance"] - if pieces["dirty"]: - rendered += ".dev0" - rendered += plus_or_dot(pieces) - rendered += "g%%s" %% pieces["short"] - else: - # exception #1 - rendered = "0.post%%d" %% pieces["distance"] - if pieces["dirty"]: - rendered += ".dev0" - rendered += "+g%%s" %% pieces["short"] - return rendered - - -def render_pep440_post_branch(pieces: Dict[str, Any]) -> str: - """TAG[.postDISTANCE[.dev0]+gHEX[.dirty]] . - - The ".dev0" means not master branch. - - Exceptions: - 1: no tags. 0.postDISTANCE[.dev0]+gHEX[.dirty] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - rendered += ".post%%d" %% pieces["distance"] - if pieces["branch"] != "master": - rendered += ".dev0" - rendered += plus_or_dot(pieces) - rendered += "g%%s" %% pieces["short"] - if pieces["dirty"]: - rendered += ".dirty" - else: - # exception #1 - rendered = "0.post%%d" %% pieces["distance"] - if pieces["branch"] != "master": - rendered += ".dev0" - rendered += "+g%%s" %% pieces["short"] - if pieces["dirty"]: - rendered += ".dirty" - return rendered - - -def render_pep440_old(pieces: Dict[str, Any]) -> str: - """TAG[.postDISTANCE[.dev0]] . - - The ".dev0" means dirty. - - Exceptions: - 1: no tags. 0.postDISTANCE[.dev0] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - rendered += ".post%%d" %% pieces["distance"] - if pieces["dirty"]: - rendered += ".dev0" - else: - # exception #1 - rendered = "0.post%%d" %% pieces["distance"] - if pieces["dirty"]: - rendered += ".dev0" - return rendered - - -def render_git_describe(pieces: Dict[str, Any]) -> str: - """TAG[-DISTANCE-gHEX][-dirty]. - - Like 'git describe --tags --dirty --always'. - - Exceptions: - 1: no tags. HEX[-dirty] (note: no 'g' prefix) - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"]: - rendered += "-%%d-g%%s" %% (pieces["distance"], pieces["short"]) - else: - # exception #1 - rendered = pieces["short"] - if pieces["dirty"]: - rendered += "-dirty" - return rendered - - -def render_git_describe_long(pieces: Dict[str, Any]) -> str: - """TAG-DISTANCE-gHEX[-dirty]. - - Like 'git describe --tags --dirty --always -long'. - The distance/hash is unconditional. - - Exceptions: - 1: no tags. HEX[-dirty] (note: no 'g' prefix) - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - rendered += "-%%d-g%%s" %% (pieces["distance"], pieces["short"]) - else: - # exception #1 - rendered = pieces["short"] - if pieces["dirty"]: - rendered += "-dirty" - return rendered - - -def render(pieces: Dict[str, Any], style: str) -> Dict[str, Any]: - """Render the given version pieces into the requested style.""" - if pieces["error"]: - return {"version": "unknown", - "full-revisionid": pieces.get("long"), - "dirty": None, - "error": pieces["error"], - "date": None} - - if not style or style == "default": - style = "pep440" # the default - - if style == "pep440": - rendered = render_pep440(pieces) - elif style == "pep440-branch": - rendered = render_pep440_branch(pieces) - elif style == "pep440-pre": - rendered = render_pep440_pre(pieces) - elif style == "pep440-post": - rendered = render_pep440_post(pieces) - elif style == "pep440-post-branch": - rendered = render_pep440_post_branch(pieces) - elif style == "pep440-old": - rendered = render_pep440_old(pieces) - elif style == "git-describe": - rendered = render_git_describe(pieces) - elif style == "git-describe-long": - rendered = render_git_describe_long(pieces) - else: - raise ValueError("unknown style '%%s'" %% style) - - return {"version": rendered, "full-revisionid": pieces["long"], - "dirty": pieces["dirty"], "error": None, - "date": pieces.get("date")} - - -def get_versions() -> Dict[str, Any]: - """Get version information or return default if unable to do so.""" - # I am in _version.py, which lives at ROOT/VERSIONFILE_SOURCE. If we have - # __file__, we can work backwards from there to the root. Some - # py2exe/bbfreeze/non-CPython implementations don't do __file__, in which - # case we can only use expanded keywords. - - cfg = get_config() - verbose = cfg.verbose - - try: - return git_versions_from_keywords(get_keywords(), cfg.tag_prefix, - verbose) - except NotThisMethod: - pass - - try: - root = os.path.realpath(__file__) - # versionfile_source is the relative path from the top of the source - # tree (where the .git directory might live) to this file. Invert - # this to find the root from __file__. - for _ in cfg.versionfile_source.split('/'): - root = os.path.dirname(root) - except NameError: - return {"version": "0+unknown", "full-revisionid": None, - "dirty": None, - "error": "unable to find root of source tree", - "date": None} - - try: - pieces = git_pieces_from_vcs(cfg.tag_prefix, root, verbose) - return render(pieces, cfg.style) - except NotThisMethod: - pass - - try: - if cfg.parentdir_prefix: - return versions_from_parentdir(cfg.parentdir_prefix, root, verbose) - except NotThisMethod: - pass - - return {"version": "0+unknown", "full-revisionid": None, - "dirty": None, - "error": "unable to compute version", "date": None} -''' - - -@register_vcs_handler("git", "get_keywords") -def git_get_keywords(versionfile_abs: str) -> Dict[str, str]: - """Extract version information from the given file.""" - # the code embedded in _version.py can just fetch the value of these - # keywords. When used from setup.py, we don't want to import _version.py, - # so we do it with a regexp instead. This function is not used from - # _version.py. - keywords: Dict[str, str] = {} - try: - with open(versionfile_abs, "r") as fobj: - for line in fobj: - if line.strip().startswith("git_refnames ="): - mo = re.search(r'=\s*"(.*)"', line) - if mo: - keywords["refnames"] = mo.group(1) - if line.strip().startswith("git_full ="): - mo = re.search(r'=\s*"(.*)"', line) - if mo: - keywords["full"] = mo.group(1) - if line.strip().startswith("git_date ="): - mo = re.search(r'=\s*"(.*)"', line) - if mo: - keywords["date"] = mo.group(1) - except OSError: - pass - return keywords - - -@register_vcs_handler("git", "keywords") -def git_versions_from_keywords( - keywords: Dict[str, str], - tag_prefix: str, - verbose: bool, -) -> Dict[str, Any]: - """Get version information from git keywords.""" - if "refnames" not in keywords: - raise NotThisMethod("Short version file found") - date = keywords.get("date") - if date is not None: - # Use only the last line. Previous lines may contain GPG signature - # information. - date = date.splitlines()[-1] - - # git-2.2.0 added "%cI", which expands to an ISO-8601 -compliant - # datestamp. However we prefer "%ci" (which expands to an "ISO-8601 - # -like" string, which we must then edit to make compliant), because - # it's been around since git-1.5.3, and it's too difficult to - # discover which version we're using, or to work around using an - # older one. - date = date.strip().replace(" ", "T", 1).replace(" ", "", 1) - refnames = keywords["refnames"].strip() - if refnames.startswith("$Format"): - if verbose: - print("keywords are unexpanded, not using") - raise NotThisMethod("unexpanded keywords, not a git-archive tarball") - refs = {r.strip() for r in refnames.strip("()").split(",")} - # starting in git-1.8.3, tags are listed as "tag: foo-1.0" instead of - # just "foo-1.0". If we see a "tag: " prefix, prefer those. - TAG = "tag: " - tags = {r[len(TAG):] for r in refs if r.startswith(TAG)} - if not tags: - # Either we're using git < 1.8.3, or there really are no tags. We use - # a heuristic: assume all version tags have a digit. The old git %d - # expansion behaves like git log --decorate=short and strips out the - # refs/heads/ and refs/tags/ prefixes that would let us distinguish - # between branches and tags. By ignoring refnames without digits, we - # filter out many common branch names like "release" and - # "stabilization", as well as "HEAD" and "master". - tags = {r for r in refs if re.search(r'\d', r)} - if verbose: - print("discarding '%s', no digits" % ",".join(refs - tags)) - if verbose: - print("likely tags: %s" % ",".join(sorted(tags))) - for ref in sorted(tags): - # sorting will prefer e.g. "2.0" over "2.0rc1" - if ref.startswith(tag_prefix): - r = ref[len(tag_prefix):] - # Filter out refs that exactly match prefix or that don't start - # with a number once the prefix is stripped (mostly a concern - # when prefix is '') - if not re.match(r'\d', r): - continue - if verbose: - print("picking %s" % r) - return {"version": r, - "full-revisionid": keywords["full"].strip(), - "dirty": False, "error": None, - "date": date} - # no suitable tags, so version is "0+unknown", but full hex is still there - if verbose: - print("no suitable tags, using unknown + full revision id") - return {"version": "0+unknown", - "full-revisionid": keywords["full"].strip(), - "dirty": False, "error": "no suitable tags", "date": None} - - -@register_vcs_handler("git", "pieces_from_vcs") -def git_pieces_from_vcs( - tag_prefix: str, - root: str, - verbose: bool, - runner: Callable = run_command -) -> Dict[str, Any]: - """Get version from 'git describe' in the root of the source tree. - - This only gets called if the git-archive 'subst' keywords were *not* - expanded, and _version.py hasn't already been rewritten with a short - version string, meaning we're inside a checked out source tree. - """ - GITS = ["git"] - if sys.platform == "win32": - GITS = ["git.cmd", "git.exe"] - - # GIT_DIR can interfere with correct operation of Versioneer. - # It may be intended to be passed to the Versioneer-versioned project, - # but that should not change where we get our version from. - env = os.environ.copy() - env.pop("GIT_DIR", None) - runner = functools.partial(runner, env=env) - - _, rc = runner(GITS, ["rev-parse", "--git-dir"], cwd=root, - hide_stderr=not verbose) - if rc != 0: - if verbose: - print("Directory %s not under git control" % root) - raise NotThisMethod("'git rev-parse --git-dir' returned error") - - # if there is a tag matching tag_prefix, this yields TAG-NUM-gHEX[-dirty] - # if there isn't one, this yields HEX[-dirty] (no NUM) - describe_out, rc = runner(GITS, [ - "describe", "--tags", "--dirty", "--always", "--long", - "--match", f"{tag_prefix}[[:digit:]]*" - ], cwd=root) - # --long was added in git-1.5.5 - if describe_out is None: - raise NotThisMethod("'git describe' failed") - describe_out = describe_out.strip() - full_out, rc = runner(GITS, ["rev-parse", "HEAD"], cwd=root) - if full_out is None: - raise NotThisMethod("'git rev-parse' failed") - full_out = full_out.strip() - - pieces: Dict[str, Any] = {} - pieces["long"] = full_out - pieces["short"] = full_out[:7] # maybe improved later - pieces["error"] = None - - branch_name, rc = runner(GITS, ["rev-parse", "--abbrev-ref", "HEAD"], - cwd=root) - # --abbrev-ref was added in git-1.6.3 - if rc != 0 or branch_name is None: - raise NotThisMethod("'git rev-parse --abbrev-ref' returned error") - branch_name = branch_name.strip() - - if branch_name == "HEAD": - # If we aren't exactly on a branch, pick a branch which represents - # the current commit. If all else fails, we are on a branchless - # commit. - branches, rc = runner(GITS, ["branch", "--contains"], cwd=root) - # --contains was added in git-1.5.4 - if rc != 0 or branches is None: - raise NotThisMethod("'git branch --contains' returned error") - branches = branches.split("\n") - - # Remove the first line if we're running detached - if "(" in branches[0]: - branches.pop(0) - - # Strip off the leading "* " from the list of branches. - branches = [branch[2:] for branch in branches] - if "master" in branches: - branch_name = "master" - elif not branches: - branch_name = None - else: - # Pick the first branch that is returned. Good or bad. - branch_name = branches[0] - - pieces["branch"] = branch_name - - # parse describe_out. It will be like TAG-NUM-gHEX[-dirty] or HEX[-dirty] - # TAG might have hyphens. - git_describe = describe_out - - # look for -dirty suffix - dirty = git_describe.endswith("-dirty") - pieces["dirty"] = dirty - if dirty: - git_describe = git_describe[:git_describe.rindex("-dirty")] - - # now we have TAG-NUM-gHEX or HEX - - if "-" in git_describe: - # TAG-NUM-gHEX - mo = re.search(r'^(.+)-(\d+)-g([0-9a-f]+)$', git_describe) - if not mo: - # unparsable. Maybe git-describe is misbehaving? - pieces["error"] = ("unable to parse git-describe output: '%s'" - % describe_out) - return pieces - - # tag - full_tag = mo.group(1) - if not full_tag.startswith(tag_prefix): - if verbose: - fmt = "tag '%s' doesn't start with prefix '%s'" - print(fmt % (full_tag, tag_prefix)) - pieces["error"] = ("tag '%s' doesn't start with prefix '%s'" - % (full_tag, tag_prefix)) - return pieces - pieces["closest-tag"] = full_tag[len(tag_prefix):] - - # distance: number of commits since tag - pieces["distance"] = int(mo.group(2)) - - # commit: short hex revision ID - pieces["short"] = mo.group(3) - - else: - # HEX: no tags - pieces["closest-tag"] = None - out, rc = runner(GITS, ["rev-list", "HEAD", "--left-right"], cwd=root) - pieces["distance"] = len(out.split()) # total number of commits - - # commit date: see ISO-8601 comment in git_versions_from_keywords() - date = runner(GITS, ["show", "-s", "--format=%ci", "HEAD"], cwd=root)[0].strip() - # Use only the last line. Previous lines may contain GPG signature - # information. - date = date.splitlines()[-1] - pieces["date"] = date.strip().replace(" ", "T", 1).replace(" ", "", 1) - - return pieces - - -def do_vcs_install(versionfile_source: str, ipy: Optional[str]) -> None: - """Git-specific installation logic for Versioneer. - - For Git, this means creating/changing .gitattributes to mark _version.py - for export-subst keyword substitution. - """ - GITS = ["git"] - if sys.platform == "win32": - GITS = ["git.cmd", "git.exe"] - files = [versionfile_source] - if ipy: - files.append(ipy) - if "VERSIONEER_PEP518" not in globals(): - try: - my_path = __file__ - if my_path.endswith((".pyc", ".pyo")): - my_path = os.path.splitext(my_path)[0] + ".py" - versioneer_file = os.path.relpath(my_path) - except NameError: - versioneer_file = "versioneer.py" - files.append(versioneer_file) - present = False - try: - with open(".gitattributes", "r") as fobj: - for line in fobj: - if line.strip().startswith(versionfile_source): - if "export-subst" in line.strip().split()[1:]: - present = True - break - except OSError: - pass - if not present: - with open(".gitattributes", "a+") as fobj: - fobj.write(f"{versionfile_source} export-subst\n") - files.append(".gitattributes") - run_command(GITS, ["add", "--"] + files) - - -def versions_from_parentdir( - parentdir_prefix: str, - root: str, - verbose: bool, -) -> Dict[str, Any]: - """Try to determine the version from the parent directory name. - - Source tarballs conventionally unpack into a directory that includes both - the project name and a version string. We will also support searching up - two directory levels for an appropriately named parent directory - """ - rootdirs = [] - - for _ in range(3): - dirname = os.path.basename(root) - if dirname.startswith(parentdir_prefix): - return {"version": dirname[len(parentdir_prefix):], - "full-revisionid": None, - "dirty": False, "error": None, "date": None} - rootdirs.append(root) - root = os.path.dirname(root) # up a level - - if verbose: - print("Tried directories %s but none started with prefix %s" % - (str(rootdirs), parentdir_prefix)) - raise NotThisMethod("rootdir doesn't start with parentdir_prefix") - - -SHORT_VERSION_PY = """ -# This file was generated by 'versioneer.py' (0.29) from -# revision-control system data, or from the parent directory name of an -# unpacked source archive. Distribution tarballs contain a pre-generated copy -# of this file. - -import json - -version_json = ''' -%s -''' # END VERSION_JSON - - -def get_versions(): - return json.loads(version_json) -""" - - -def versions_from_file(filename: str) -> Dict[str, Any]: - """Try to determine the version from _version.py if present.""" - try: - with open(filename) as f: - contents = f.read() - except OSError: - raise NotThisMethod("unable to read _version.py") - mo = re.search(r"version_json = '''\n(.*)''' # END VERSION_JSON", - contents, re.M | re.S) - if not mo: - mo = re.search(r"version_json = '''\r\n(.*)''' # END VERSION_JSON", - contents, re.M | re.S) - if not mo: - raise NotThisMethod("no version_json in _version.py") - return json.loads(mo.group(1)) - - -def write_to_version_file(filename: str, versions: Dict[str, Any]) -> None: - """Write the given version number to the given _version.py file.""" - contents = json.dumps(versions, sort_keys=True, - indent=1, separators=(",", ": ")) - with open(filename, "w") as f: - f.write(SHORT_VERSION_PY % contents) - - print("set %s to '%s'" % (filename, versions["version"])) - - -def plus_or_dot(pieces: Dict[str, Any]) -> str: - """Return a + if we don't already have one, else return a .""" - if "+" in pieces.get("closest-tag", ""): - return "." - return "+" - - -def render_pep440(pieces: Dict[str, Any]) -> str: - """Build up version string, with post-release "local version identifier". - - Our goal: TAG[+DISTANCE.gHEX[.dirty]] . Note that if you - get a tagged build and then dirty it, you'll get TAG+0.gHEX.dirty - - Exceptions: - 1: no tags. git_describe was just HEX. 0+untagged.DISTANCE.gHEX[.dirty] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - rendered += plus_or_dot(pieces) - rendered += "%d.g%s" % (pieces["distance"], pieces["short"]) - if pieces["dirty"]: - rendered += ".dirty" - else: - # exception #1 - rendered = "0+untagged.%d.g%s" % (pieces["distance"], - pieces["short"]) - if pieces["dirty"]: - rendered += ".dirty" - return rendered - - -def render_pep440_branch(pieces: Dict[str, Any]) -> str: - """TAG[[.dev0]+DISTANCE.gHEX[.dirty]] . - - The ".dev0" means not master branch. Note that .dev0 sorts backwards - (a feature branch will appear "older" than the master branch). - - Exceptions: - 1: no tags. 0[.dev0]+untagged.DISTANCE.gHEX[.dirty] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - if pieces["branch"] != "master": - rendered += ".dev0" - rendered += plus_or_dot(pieces) - rendered += "%d.g%s" % (pieces["distance"], pieces["short"]) - if pieces["dirty"]: - rendered += ".dirty" - else: - # exception #1 - rendered = "0" - if pieces["branch"] != "master": - rendered += ".dev0" - rendered += "+untagged.%d.g%s" % (pieces["distance"], - pieces["short"]) - if pieces["dirty"]: - rendered += ".dirty" - return rendered - - -def pep440_split_post(ver: str) -> Tuple[str, Optional[int]]: - """Split pep440 version string at the post-release segment. - - Returns the release segments before the post-release and the - post-release version number (or -1 if no post-release segment is present). - """ - vc = str.split(ver, ".post") - return vc[0], int(vc[1] or 0) if len(vc) == 2 else None - - -def render_pep440_pre(pieces: Dict[str, Any]) -> str: - """TAG[.postN.devDISTANCE] -- No -dirty. - - Exceptions: - 1: no tags. 0.post0.devDISTANCE - """ - if pieces["closest-tag"]: - if pieces["distance"]: - # update the post release segment - tag_version, post_version = pep440_split_post(pieces["closest-tag"]) - rendered = tag_version - if post_version is not None: - rendered += ".post%d.dev%d" % (post_version + 1, pieces["distance"]) - else: - rendered += ".post0.dev%d" % (pieces["distance"]) - else: - # no commits, use the tag as the version - rendered = pieces["closest-tag"] - else: - # exception #1 - rendered = "0.post0.dev%d" % pieces["distance"] - return rendered - - -def render_pep440_post(pieces: Dict[str, Any]) -> str: - """TAG[.postDISTANCE[.dev0]+gHEX] . - - The ".dev0" means dirty. Note that .dev0 sorts backwards - (a dirty tree will appear "older" than the corresponding clean one), - but you shouldn't be releasing software with -dirty anyways. - - Exceptions: - 1: no tags. 0.postDISTANCE[.dev0] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - rendered += ".post%d" % pieces["distance"] - if pieces["dirty"]: - rendered += ".dev0" - rendered += plus_or_dot(pieces) - rendered += "g%s" % pieces["short"] - else: - # exception #1 - rendered = "0.post%d" % pieces["distance"] - if pieces["dirty"]: - rendered += ".dev0" - rendered += "+g%s" % pieces["short"] - return rendered - - -def render_pep440_post_branch(pieces: Dict[str, Any]) -> str: - """TAG[.postDISTANCE[.dev0]+gHEX[.dirty]] . - - The ".dev0" means not master branch. - - Exceptions: - 1: no tags. 0.postDISTANCE[.dev0]+gHEX[.dirty] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - rendered += ".post%d" % pieces["distance"] - if pieces["branch"] != "master": - rendered += ".dev0" - rendered += plus_or_dot(pieces) - rendered += "g%s" % pieces["short"] - if pieces["dirty"]: - rendered += ".dirty" - else: - # exception #1 - rendered = "0.post%d" % pieces["distance"] - if pieces["branch"] != "master": - rendered += ".dev0" - rendered += "+g%s" % pieces["short"] - if pieces["dirty"]: - rendered += ".dirty" - return rendered - - -def render_pep440_old(pieces: Dict[str, Any]) -> str: - """TAG[.postDISTANCE[.dev0]] . - - The ".dev0" means dirty. - - Exceptions: - 1: no tags. 0.postDISTANCE[.dev0] - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"] or pieces["dirty"]: - rendered += ".post%d" % pieces["distance"] - if pieces["dirty"]: - rendered += ".dev0" - else: - # exception #1 - rendered = "0.post%d" % pieces["distance"] - if pieces["dirty"]: - rendered += ".dev0" - return rendered - - -def render_git_describe(pieces: Dict[str, Any]) -> str: - """TAG[-DISTANCE-gHEX][-dirty]. - - Like 'git describe --tags --dirty --always'. - - Exceptions: - 1: no tags. HEX[-dirty] (note: no 'g' prefix) - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - if pieces["distance"]: - rendered += "-%d-g%s" % (pieces["distance"], pieces["short"]) - else: - # exception #1 - rendered = pieces["short"] - if pieces["dirty"]: - rendered += "-dirty" - return rendered - - -def render_git_describe_long(pieces: Dict[str, Any]) -> str: - """TAG-DISTANCE-gHEX[-dirty]. - - Like 'git describe --tags --dirty --always -long'. - The distance/hash is unconditional. - - Exceptions: - 1: no tags. HEX[-dirty] (note: no 'g' prefix) - """ - if pieces["closest-tag"]: - rendered = pieces["closest-tag"] - rendered += "-%d-g%s" % (pieces["distance"], pieces["short"]) - else: - # exception #1 - rendered = pieces["short"] - if pieces["dirty"]: - rendered += "-dirty" - return rendered - - -def render(pieces: Dict[str, Any], style: str) -> Dict[str, Any]: - """Render the given version pieces into the requested style.""" - if pieces["error"]: - return {"version": "unknown", - "full-revisionid": pieces.get("long"), - "dirty": None, - "error": pieces["error"], - "date": None} - - if not style or style == "default": - style = "pep440" # the default - - if style == "pep440": - rendered = render_pep440(pieces) - elif style == "pep440-branch": - rendered = render_pep440_branch(pieces) - elif style == "pep440-pre": - rendered = render_pep440_pre(pieces) - elif style == "pep440-post": - rendered = render_pep440_post(pieces) - elif style == "pep440-post-branch": - rendered = render_pep440_post_branch(pieces) - elif style == "pep440-old": - rendered = render_pep440_old(pieces) - elif style == "git-describe": - rendered = render_git_describe(pieces) - elif style == "git-describe-long": - rendered = render_git_describe_long(pieces) - else: - raise ValueError("unknown style '%s'" % style) - - return {"version": rendered, "full-revisionid": pieces["long"], - "dirty": pieces["dirty"], "error": None, - "date": pieces.get("date")} - - -class VersioneerBadRootError(Exception): - """The project root directory is unknown or missing key files.""" - - -def get_versions(verbose: bool = False) -> Dict[str, Any]: - """Get the project version from whatever source is available. - - Returns dict with two keys: 'version' and 'full'. - """ - if "versioneer" in sys.modules: - # see the discussion in cmdclass.py:get_cmdclass() - del sys.modules["versioneer"] - - root = get_root() - cfg = get_config_from_root(root) - - assert cfg.VCS is not None, "please set [versioneer]VCS= in setup.cfg" - handlers = HANDLERS.get(cfg.VCS) - assert handlers, "unrecognized VCS '%s'" % cfg.VCS - verbose = verbose or bool(cfg.verbose) # `bool()` used to avoid `None` - assert cfg.versionfile_source is not None, \ - "please set versioneer.versionfile_source" - assert cfg.tag_prefix is not None, "please set versioneer.tag_prefix" - - versionfile_abs = os.path.join(root, cfg.versionfile_source) - - # extract version from first of: _version.py, VCS command (e.g. 'git - # describe'), parentdir. This is meant to work for developers using a - # source checkout, for users of a tarball created by 'setup.py sdist', - # and for users of a tarball/zipball created by 'git archive' or github's - # download-from-tag feature or the equivalent in other VCSes. - - get_keywords_f = handlers.get("get_keywords") - from_keywords_f = handlers.get("keywords") - if get_keywords_f and from_keywords_f: - try: - keywords = get_keywords_f(versionfile_abs) - ver = from_keywords_f(keywords, cfg.tag_prefix, verbose) - if verbose: - print("got version from expanded keyword %s" % ver) - return ver - except NotThisMethod: - pass - - try: - ver = versions_from_file(versionfile_abs) - if verbose: - print("got version from file %s %s" % (versionfile_abs, ver)) - return ver - except NotThisMethod: - pass - - from_vcs_f = handlers.get("pieces_from_vcs") - if from_vcs_f: - try: - pieces = from_vcs_f(cfg.tag_prefix, root, verbose) - ver = render(pieces, cfg.style) - if verbose: - print("got version from VCS %s" % ver) - return ver - except NotThisMethod: - pass - - try: - if cfg.parentdir_prefix: - ver = versions_from_parentdir(cfg.parentdir_prefix, root, verbose) - if verbose: - print("got version from parentdir %s" % ver) - return ver - except NotThisMethod: - pass - - if verbose: - print("unable to compute version") - - return {"version": "0+unknown", "full-revisionid": None, - "dirty": None, "error": "unable to compute version", - "date": None} - - -def get_version() -> str: - """Get the short version string for this project.""" - return get_versions()["version"] - - -def get_cmdclass(cmdclass: Optional[Dict[str, Any]] = None): - """Get the custom setuptools subclasses used by Versioneer. - - If the package uses a different cmdclass (e.g. one from numpy), it - should be provide as an argument. - """ - if "versioneer" in sys.modules: - del sys.modules["versioneer"] - # this fixes the "python setup.py develop" case (also 'install' and - # 'easy_install .'), in which subdependencies of the main project are - # built (using setup.py bdist_egg) in the same python process. Assume - # a main project A and a dependency B, which use different versions - # of Versioneer. A's setup.py imports A's Versioneer, leaving it in - # sys.modules by the time B's setup.py is executed, causing B to run - # with the wrong versioneer. Setuptools wraps the sub-dep builds in a - # sandbox that restores sys.modules to it's pre-build state, so the - # parent is protected against the child's "import versioneer". By - # removing ourselves from sys.modules here, before the child build - # happens, we protect the child from the parent's versioneer too. - # Also see https://github.com/python-versioneer/python-versioneer/issues/52 - - cmds = {} if cmdclass is None else cmdclass.copy() - - # we add "version" to setuptools - from setuptools import Command - - class cmd_version(Command): - description = "report generated version string" - user_options: List[Tuple[str, str, str]] = [] - boolean_options: List[str] = [] - - def initialize_options(self) -> None: - pass - - def finalize_options(self) -> None: - pass - - def run(self) -> None: - vers = get_versions(verbose=True) - print("Version: %s" % vers["version"]) - print(" full-revisionid: %s" % vers.get("full-revisionid")) - print(" dirty: %s" % vers.get("dirty")) - print(" date: %s" % vers.get("date")) - if vers["error"]: - print(" error: %s" % vers["error"]) - cmds["version"] = cmd_version - - # we override "build_py" in setuptools - # - # most invocation pathways end up running build_py: - # distutils/build -> build_py - # distutils/install -> distutils/build ->.. - # setuptools/bdist_wheel -> distutils/install ->.. - # setuptools/bdist_egg -> distutils/install_lib -> build_py - # setuptools/install -> bdist_egg ->.. - # setuptools/develop -> ? - # pip install: - # copies source tree to a tempdir before running egg_info/etc - # if .git isn't copied too, 'git describe' will fail - # then does setup.py bdist_wheel, or sometimes setup.py install - # setup.py egg_info -> ? - - # pip install -e . and setuptool/editable_wheel will invoke build_py - # but the build_py command is not expected to copy any files. - - # we override different "build_py" commands for both environments - if 'build_py' in cmds: - _build_py: Any = cmds['build_py'] - else: - from setuptools.command.build_py import build_py as _build_py - - class cmd_build_py(_build_py): - def run(self) -> None: - root = get_root() - cfg = get_config_from_root(root) - versions = get_versions() - _build_py.run(self) - if getattr(self, "editable_mode", False): - # During editable installs `.py` and data files are - # not copied to build_lib - return - # now locate _version.py in the new build/ directory and replace - # it with an updated value - if cfg.versionfile_build: - target_versionfile = os.path.join(self.build_lib, - cfg.versionfile_build) - print("UPDATING %s" % target_versionfile) - write_to_version_file(target_versionfile, versions) - cmds["build_py"] = cmd_build_py - - if 'build_ext' in cmds: - _build_ext: Any = cmds['build_ext'] - else: - from setuptools.command.build_ext import build_ext as _build_ext - - class cmd_build_ext(_build_ext): - def run(self) -> None: - root = get_root() - cfg = get_config_from_root(root) - versions = get_versions() - _build_ext.run(self) - if self.inplace: - # build_ext --inplace will only build extensions in - # build/lib<..> dir with no _version.py to write to. - # As in place builds will already have a _version.py - # in the module dir, we do not need to write one. - return - # now locate _version.py in the new build/ directory and replace - # it with an updated value - if not cfg.versionfile_build: - return - target_versionfile = os.path.join(self.build_lib, - cfg.versionfile_build) - if not os.path.exists(target_versionfile): - print(f"Warning: {target_versionfile} does not exist, skipping " - "version update. This can happen if you are running build_ext " - "without first running build_py.") - return - print("UPDATING %s" % target_versionfile) - write_to_version_file(target_versionfile, versions) - cmds["build_ext"] = cmd_build_ext - - if "cx_Freeze" in sys.modules: # cx_freeze enabled? - from cx_Freeze.dist import build_exe as _build_exe # type: ignore - # nczeczulin reports that py2exe won't like the pep440-style string - # as FILEVERSION, but it can be used for PRODUCTVERSION, e.g. - # setup(console=[{ - # "version": versioneer.get_version().split("+", 1)[0], # FILEVERSION - # "product_version": versioneer.get_version(), - # ... - - class cmd_build_exe(_build_exe): - def run(self) -> None: - root = get_root() - cfg = get_config_from_root(root) - versions = get_versions() - target_versionfile = cfg.versionfile_source - print("UPDATING %s" % target_versionfile) - write_to_version_file(target_versionfile, versions) - - _build_exe.run(self) - os.unlink(target_versionfile) - with open(cfg.versionfile_source, "w") as f: - LONG = LONG_VERSION_PY[cfg.VCS] - f.write(LONG % - {"DOLLAR": "$", - "STYLE": cfg.style, - "TAG_PREFIX": cfg.tag_prefix, - "PARENTDIR_PREFIX": cfg.parentdir_prefix, - "VERSIONFILE_SOURCE": cfg.versionfile_source, - }) - cmds["build_exe"] = cmd_build_exe - del cmds["build_py"] - - if 'py2exe' in sys.modules: # py2exe enabled? - try: - from py2exe.setuptools_buildexe import py2exe as _py2exe # type: ignore - except ImportError: - from py2exe.distutils_buildexe import py2exe as _py2exe # type: ignore - - class cmd_py2exe(_py2exe): - def run(self) -> None: - root = get_root() - cfg = get_config_from_root(root) - versions = get_versions() - target_versionfile = cfg.versionfile_source - print("UPDATING %s" % target_versionfile) - write_to_version_file(target_versionfile, versions) - - _py2exe.run(self) - os.unlink(target_versionfile) - with open(cfg.versionfile_source, "w") as f: - LONG = LONG_VERSION_PY[cfg.VCS] - f.write(LONG % - {"DOLLAR": "$", - "STYLE": cfg.style, - "TAG_PREFIX": cfg.tag_prefix, - "PARENTDIR_PREFIX": cfg.parentdir_prefix, - "VERSIONFILE_SOURCE": cfg.versionfile_source, - }) - cmds["py2exe"] = cmd_py2exe - - # sdist farms its file list building out to egg_info - if 'egg_info' in cmds: - _egg_info: Any = cmds['egg_info'] - else: - from setuptools.command.egg_info import egg_info as _egg_info - - class cmd_egg_info(_egg_info): - def find_sources(self) -> None: - # egg_info.find_sources builds the manifest list and writes it - # in one shot - super().find_sources() - - # Modify the filelist and normalize it - root = get_root() - cfg = get_config_from_root(root) - self.filelist.append('versioneer.py') - if cfg.versionfile_source: - # There are rare cases where versionfile_source might not be - # included by default, so we must be explicit - self.filelist.append(cfg.versionfile_source) - self.filelist.sort() - self.filelist.remove_duplicates() - - # The write method is hidden in the manifest_maker instance that - # generated the filelist and was thrown away - # We will instead replicate their final normalization (to unicode, - # and POSIX-style paths) - from setuptools import unicode_utils - normalized = [unicode_utils.filesys_decode(f).replace(os.sep, '/') - for f in self.filelist.files] - - manifest_filename = os.path.join(self.egg_info, 'SOURCES.txt') - with open(manifest_filename, 'w') as fobj: - fobj.write('\n'.join(normalized)) - - cmds['egg_info'] = cmd_egg_info - - # we override different "sdist" commands for both environments - if 'sdist' in cmds: - _sdist: Any = cmds['sdist'] - else: - from setuptools.command.sdist import sdist as _sdist - - class cmd_sdist(_sdist): - def run(self) -> None: - versions = get_versions() - self._versioneer_generated_versions = versions - # unless we update this, the command will keep using the old - # version - self.distribution.metadata.version = versions["version"] - return _sdist.run(self) - - def make_release_tree(self, base_dir: str, files: List[str]) -> None: - root = get_root() - cfg = get_config_from_root(root) - _sdist.make_release_tree(self, base_dir, files) - # now locate _version.py in the new base_dir directory - # (remembering that it may be a hardlink) and replace it with an - # updated value - target_versionfile = os.path.join(base_dir, cfg.versionfile_source) - print("UPDATING %s" % target_versionfile) - write_to_version_file(target_versionfile, - self._versioneer_generated_versions) - cmds["sdist"] = cmd_sdist - - return cmds - - -CONFIG_ERROR = """ -setup.cfg is missing the necessary Versioneer configuration. You need -a section like: - - [versioneer] - VCS = git - style = pep440 - versionfile_source = src/myproject/_version.py - versionfile_build = myproject/_version.py - tag_prefix = - parentdir_prefix = myproject- - -You will also need to edit your setup.py to use the results: - - import versioneer - setup(version=versioneer.get_version(), - cmdclass=versioneer.get_cmdclass(), ...) - -Please read the docstring in ./versioneer.py for configuration instructions, -edit setup.cfg, and re-run the installer or 'python versioneer.py setup'. -""" - -SAMPLE_CONFIG = """ -# See the docstring in versioneer.py for instructions. Note that you must -# re-run 'versioneer.py setup' after changing this section, and commit the -# resulting files. - -[versioneer] -#VCS = git -#style = pep440 -#versionfile_source = -#versionfile_build = -#tag_prefix = -#parentdir_prefix = - -""" - -OLD_SNIPPET = """ -from ._version import get_versions -__version__ = get_versions()['version'] -del get_versions -""" - -INIT_PY_SNIPPET = """ -from . import {0} -__version__ = {0}.get_versions()['version'] -""" - - -def do_setup() -> int: - """Do main VCS-independent setup function for installing Versioneer.""" - root = get_root() - try: - cfg = get_config_from_root(root) - except (OSError, configparser.NoSectionError, - configparser.NoOptionError) as e: - if isinstance(e, (OSError, configparser.NoSectionError)): - print("Adding sample versioneer config to setup.cfg", - file=sys.stderr) - with open(os.path.join(root, "setup.cfg"), "a") as f: - f.write(SAMPLE_CONFIG) - print(CONFIG_ERROR, file=sys.stderr) - return 1 - - print(" creating %s" % cfg.versionfile_source) - with open(cfg.versionfile_source, "w") as f: - LONG = LONG_VERSION_PY[cfg.VCS] - f.write(LONG % {"DOLLAR": "$", - "STYLE": cfg.style, - "TAG_PREFIX": cfg.tag_prefix, - "PARENTDIR_PREFIX": cfg.parentdir_prefix, - "VERSIONFILE_SOURCE": cfg.versionfile_source, - }) - - ipy = os.path.join(os.path.dirname(cfg.versionfile_source), - "__init__.py") - maybe_ipy: Optional[str] = ipy - if os.path.exists(ipy): - try: - with open(ipy, "r") as f: - old = f.read() - except OSError: - old = "" - module = os.path.splitext(os.path.basename(cfg.versionfile_source))[0] - snippet = INIT_PY_SNIPPET.format(module) - if OLD_SNIPPET in old: - print(" replacing boilerplate in %s" % ipy) - with open(ipy, "w") as f: - f.write(old.replace(OLD_SNIPPET, snippet)) - elif snippet not in old: - print(" appending to %s" % ipy) - with open(ipy, "a") as f: - f.write(snippet) - else: - print(" %s unmodified" % ipy) - else: - print(" %s doesn't exist, ok" % ipy) - maybe_ipy = None - - # Make VCS-specific changes. For git, this means creating/changing - # .gitattributes to mark _version.py for export-subst keyword - # substitution. - do_vcs_install(cfg.versionfile_source, maybe_ipy) - return 0 - - -def scan_setup_py() -> int: - """Validate the contents of setup.py against Versioneer's expectations.""" - found = set() - setters = False - errors = 0 - with open("setup.py", "r") as f: - for line in f.readlines(): - if "import versioneer" in line: - found.add("import") - if "versioneer.get_cmdclass()" in line: - found.add("cmdclass") - if "versioneer.get_version()" in line: - found.add("get_version") - if "versioneer.VCS" in line: - setters = True - if "versioneer.versionfile_source" in line: - setters = True - if len(found) != 3: - print("") - print("Your setup.py appears to be missing some important items") - print("(but I might be wrong). Please make sure it has something") - print("roughly like the following:") - print("") - print(" import versioneer") - print(" setup( version=versioneer.get_version(),") - print(" cmdclass=versioneer.get_cmdclass(), ...)") - print("") - errors += 1 - if setters: - print("You should remove lines like 'versioneer.VCS = ' and") - print("'versioneer.versionfile_source = ' . This configuration") - print("now lives in setup.cfg, and should be removed from setup.py") - print("") - errors += 1 - return errors - - -def setup_command() -> NoReturn: - """Set up Versioneer and exit with appropriate error code.""" - errors = do_setup() - errors += scan_setup_py() - sys.exit(1 if errors else 0) - - -if __name__ == "__main__": - cmd = sys.argv[1] - if cmd == "setup": - setup_command()