Skip to content

Implemented Transformer Architecture from scratch using PyTorch. Also trained the model for Machine Translation task.

Notifications You must be signed in to change notification settings

SathvikNayak123/Transformers-from-Scratch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Transformers-from-Scratch

Implementation of Transformer Architecture introduced in the paper "Attention is All You Need" from scratch using PyTorch.

  • Encompasses all key components of the Transformer model, including multi-head self-attention, positional encodings, feedforward layers, and layer normalization.
  • Scalable and modular design, allowing for extensions or experimentation with Transformer variants.
  • Also trained the model for Machine translation task (English-Italian Language Translation)

pic

About

Implemented Transformer Architecture from scratch using PyTorch. Also trained the model for Machine Translation task.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages