mahakal001 / nanogpt-jax

An implementation of nanogpt in jax from scratch based on Andrej Karpathy's tutorial series on Neural Networks.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Nano GPT-jax

An implementation of nanogpt in jax from scratch ( Other than Optax for optimization and Equinox for handling PyTrees ) based on Andrej Karpathy's Let's build GPT Lecture.

Usage

  • The Shakespeare dataset is in data folder. You only need to configure hyper-parameters in nanogpt-jax/train.py as per your test settings and then run :
$ python train.py

TODOS

  • Write DropOut Layers.
  • LayerNorm.
  • Apply weight initializers.
  • Implement Adam.

References

About

An implementation of nanogpt in jax from scratch based on Andrej Karpathy's tutorial series on Neural Networks.

License:MIT License


Languages

Language:Python 100.0%