Skip to content

Tensorflow implementation of LoRA layers in Google's FLAN-T5 LLM

Notifications You must be signed in to change notification settings

ponder-lab/LoRA-FlanT5-Finetune

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 

Repository files navigation

LoRA Flan T5 Finetune

This repo hosts an example of adding LoRA (Low Rank Adaptation) layers in Google's Flan T5 model and evaluates its performance for news summarisation. The data set used is the CNN/Dailymail dataset

What are LoRA (Low Rank Adaptation)

Low Rank Adaptions allow us to finetune the weights of the model in a manner that the dimensions of the weights matrix is reduced. The weights matrix is transformed into a product of two matrices, each having a lower rank than the original weights matrix. This essentially number of parameters to fine-tune, allowing for faster training of large models.

About

Tensorflow implementation of LoRA layers in Google's FLAN-T5 LLM

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 100.0%