Dadajon / cpd-chatbot

Learn Transformer model (Attention is all you need, Google Brain, 2017) from implementation code written by @hyunwoongko in 2021.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Transformer

Learn Transformer model (Attention is all you need, Google Brain, 2017) from implementation code written by @hyunwoongko in 2021.

Python PyTorch VSCode

Results

graph

  • Min train loss = 2.7348
  • Min validation loss = 3.2860
  • Max blue score = 24.3375
Model Dataset BLEU Score
Hyunwoong Ko's Multi30K EN-DE 26.4
My Implementation Multi30K EN-DE 24.3

Reference

GPT

OpenAI GPT model using 🤗HuggingFace

BERT

BERT example from the Hugging Face transformer library

About

Learn Transformer model (Attention is all you need, Google Brain, 2017) from implementation code written by @hyunwoongko in 2021.


Languages

Language:Jupyter Notebook 91.1%Language:Python 8.9%