Skip to content

v0.6.0: Checkpointing and bfloat16 support

Compare
Choose a tag to compare
@sgugger sgugger released this 18 Mar 13:47
· 1363 commits to main since this release

This release adds support for bloat16 mixed precision training (requires PyTorch >= 1.10) and a brand-new checkpoint utility to help with resuming interrupted trainings. We also get a completely revamped documentation frontend.

Checkpoints

Save the current state of all your objects (models, optimizers, RNG states) with accelerator.save_state(path_to_checkpoint) and reload everything by calling accelerator.load_state(path_to_checkpoint)

BFloat16 support

Accelerate now supports bfloat16 mixed precision training. As a result the old --fp16 argument has been deprecated to be replaced by the more generic --mixed-precision.

New env subcommand

You can now type accelerate env to have a copy-pastable summary of your environment and default configuration. Very convenient when opening a new issue!

New doc frontend

The documentation has been switched to the new Hugging Face frontend, like Transformers and Datasets.

  • Convert documentation to the new front by @sgugger in #271

What's Changed

New Contributors

Full Changelog: v0.5.1...v0.6.0