Skip to content

v0.9.0 (Robot Stop)

Compare
Choose a tag to compare
@danieldk danieldk released this 13 Jul 13:14
· 105 commits to main since this release

We are very happy to announce this major new release of Curated Transformers! 🎉

Curated Transformers started as a small transformer library for spaCy pipelines. Over the last two months we made it a pure PyTorch library that is completely independent of spaCy and Thinc. We also added support for popular LLM models, generation, 8-bit/4-bit quantization, and many other features:

  • Curated Transformers is now a pure PyTorch library.
  • Support for popular LLMs such as Falcon, LLaMA, and Dolly v2.
  • Greedy generation and generation with sampling.
  • 8-bit and 4-bit quantization of models through bitsandbytes.
  • Flash attention and other optimizations through PyTorch Scaled Dot Product Attention.
  • Efficient model loading without unneeded allocations and initialization through the Torch meta devices.
  • Support for modern tokenizer.json tokenizers.
  • Load models from Hugging Face Hub without requiring the transformers package.
  • Extensive API documentation and examples.

Curated Transformers can be used in spaCy using the spacy-curated-transformers package.

👥 Contributors

@danieldk, @honnibal, @ines, @shadeMe