You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
LMOps: General technology for enabling AI capabilities w/ LLMs and MLLMs. [link]
lit-gpt: Hackable implementation of state-of-the-art open-source LLMs based on nanoGPT. Supports flash attention, 4-bit and 8-bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.. [link]
llmtools: 4-Bit Finetuning of Large Language Models on One Consumer GPU. [link]
torchdistill: A coding-free framework built on PyTorch for reproducible deep learning studies. 🏆20 knowledge distillation methods presented at CVPR, ICLR, ECCV, NeurIPS, ICCV, etc are implemented so far. 🎁 Trained models, training logs and configurations are available for ensuring the reproducibiliy and benchmark.. [link][paper]
gpt4all: open-source LLM chatbots that you can run anywhere. [link][paper]
low_bit_llama: Advanced Ultra-Low Bitrate Compression Techniques for the LLaMA Family of LLMs. [link]
exllama: A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.. [link]
Open-source Lightweight LLM
TinyLlama: The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.. [link]