Aveek-Saha / Transformer

A TensorFlow 2.0 Implementation of the Transformer: Attention Is All You Need

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Transformer

A TensorFlow 2.x implementation of the Transformer from Attention Is All You Need (Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, arxiv, 2017).

This is my attempt at trying to understand and recreate the transformer from the research paper. This is just for my own understanding of the subject and is by no means perfect.

In order to understand and implement the transformer I've taken the help of various tutorials and code guides, which I'll be linking in the resources section.

Requirements

  • tensorflow==2.1.0
  • numpy==1.16.5
  • tensorflow_datasets==3.2.1

How to run

python train.py

Resources

About

A TensorFlow 2.0 Implementation of the Transformer: Attention Is All You Need

License:Apache License 2.0


Languages

Language:Python 100.0%