Example of distilling LLM knowledge using LoRa
We use this dataset juancavallotti/multilingual-gec
from the huggingface Hub. It is a synthetic grammar correction dataset.
Torch
conda install pytorch torchvision torchaudio pytorch-cuda=12.1 -c pytorch -c nvidia
Other libs:
pip install -r requirements.txt
All steps to run the experiments are listed in order in the file scripts/run.sh
that you can run as bash scripts/run.sh
% of exact match between ground truth and prediction:
LLama 2–70B: 42%
Base Tiny-LLama: 11%
Distilled Tiny-LLama: 31%