Kenta Izumi (kenkenpa2126)

kenkenpa2126

Geek Repo

Company:SB Intuitions

Location:Tokyo, Japan

Github PK Tool:Github PK Tool

Kenta Izumi's repositories

Megatron-LM

Ongoing research training transformer models at scale

License:NOASSERTIONStargazers:0Issues:0Issues:0
Language:PythonLicense:MITStargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:Jupyter NotebookLicense:MITStargazers:10Issues:0Issues:0

zoology

Understand and test language model architectures on synthetic tasks.

Stargazers:0Issues:0Issues:0

gpt-fast

Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.

License:BSD-3-ClauseStargazers:0Issues:0Issues:0
License:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:0Issues:0
Language:TypeScriptLicense:MITStargazers:0Issues:0Issues:0

llama-2-jax

JAX implementation of the Llama 2 model

License:CC0-1.0Stargazers:0Issues:0Issues:0

EasyLM

Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.

License:Apache-2.0Stargazers:1Issues:0Issues:0

flax

Flax is a neural network library for JAX that is designed for flexibility.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

mesh-transformer-jax

Model parallel transformers in JAX and Haiku

License:Apache-2.0Stargazers:0Issues:0Issues:0