Skip to content

My implementation of the "Attention Is All You Need" paper, introducing the Transformer Architecture.

Notifications You must be signed in to change notification settings

sagnikc395/transformer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

transformer

My implementation of the transformer architecture, after reading Attention is All You Need paper.

Structure:

./
├── README.md
├── requirements.txt
├── ruff.toml
└── src
    ├── config.py
    ├── dataset.py
    ├── model.py
    └── train.py

2 directories, 7 files

Contents:

  • README.md

    this file (duh!)

  • src/model.py

    has the basic structure of our transformer model.

  • src/train.py

    has utiltiies for getting the data from huggingface, building out the tokenizer , getting all sentences and building the vocabulary and then training the model on top our architecture.

  • src/config.py

    has our config for the dataset like source lang, target lang, epoch size,learning rate, where to store the files etc.

  • src/dataset.py

    builds on top of the dataset and adds out SOS and EOS tokens and the padding tokens ,also adds causal mask

  • ./ruff.toml

    python formatting rules for this project.

About

My implementation of the "Attention Is All You Need" paper, introducing the Transformer Architecture.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages