Kenta Izumi's repositories
Megatron-LM
Ongoing research training transformer models at scale
NOASSERTION000
Language:PythonMIT000
Language:PythonApache-2.0000
zoology
Understand and test language model architectures on synthetic tasks.
000
gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
BSD-3-Clause000
Apache-2.0000
Language:Python000
Language:TypeScriptMIT000
llama-2-jax
JAX implementation of the Llama 2 model
CC0-1.0000
flax
Flax is a neural network library for JAX that is designed for flexibility.
Language:PythonApache-2.0000
mesh-transformer-jax
Model parallel transformers in JAX and Haiku
Apache-2.0000