yingying123321 / LSTM_with_attention

Seq2seq using LSTM with attention from Luong et al

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

LSTM_with_attention

My interest in languages and deep learning found a natural intersection in Neural Machine Translation (NMT). This notebook represents my first attempt at coding a seq2seq model to build a fully functioning English-French translator.

I implemented the attention model as outlined by Luong et al. Using just the small dataset provided, it will build a translator capable of quite basic tasks within two hours of training (on a single GPU).

However it seems RNN models are being taken over by novel approaches such as The Transformer and Temporal Convolutional Networks. I abandoned this approach eventually to build the Transformer as it trained far faster. See the program I created here, which allows you to train the Transformer on your own datasets.

About

Seq2seq using LSTM with attention from Luong et al


Languages

Language:Jupyter Notebook 100.0%