Skip to content

jasonmaoverlord/LLM-Distillery

 
 

Repository files navigation

LLM-Distillery

LLM-Distillery is a pipeline for distillation of one or multiple teacher models into a student.

Main features:

  • Single and Multi-Teacher distillation
  • Distillation on instruct and completion text
  • Offline distillation: collects the dataset, and only then trains (Yes, you can share the collected datasets)
  • Windows and Linux support
  • Automatic hdf5 dataset synchronization, with continued collection after force-exit
  • Lots of knobs to tweak! From temperature to the device mapping strategy
  • And a lot more!

Installation

See Wiki for installation instructions

Contributions

Big thanks to kalomaze for help and keeping me sane while I was building this project!
Also, thanks to AlpinDale for giving access to compute during the development!

If you want to contribute to this project, feel free!
Open issues when you encounter them, and make PRs when you feel like it.

About

A pipeline for LLM knowledge distillation

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 99.7%
  • Other 0.3%