Skip to content

A TensorFlow 2.0 Implementation of the Transformer: Attention Is All You Need

License

Notifications You must be signed in to change notification settings

Aveek-Saha/Transformer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

26 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Transformer

A TensorFlow 2.x implementation of the Transformer from Attention Is All You Need (Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, arxiv, 2017).

This is my attempt at trying to understand and recreate the transformer from the research paper. This is just for my own understanding of the subject and is by no means perfect.

In order to understand and implement the transformer I've taken the help of various tutorials and code guides, which I'll be linking in the resources section.

Requirements

  • tensorflow==2.1.0
  • numpy==1.16.5
  • tensorflow_datasets==3.2.1

How to run

python train.py

Resources