powergiant's starred repositories

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:34965Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:10164Issues:0Issues:0

gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries

Language:PythonLicense:Apache-2.0Stargazers:6856Issues:0Issues:0

nanoGPT

The simplest, fastest repository for training/finetuning medium-sized GPTs.

Language:PythonLicense:MITStargazers:36499Issues:0Issues:0

open_llama

OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset

License:Apache-2.0Stargazers:7355Issues:0Issues:0

EasyLM

Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.

Language:PythonLicense:Apache-2.0Stargazers:2377Issues:0Issues:0
License:Apache-2.0Stargazers:238Issues:0Issues:0

information-bottleneck

demonstration of the information bottleneck theory for deep learning

Language:Jupyter NotebookStargazers:58Issues:0Issues:0

mean-field-theory-deep-learning

paper lists and information on mean-field theory of deep learning

License:MITStargazers:75Issues:0Issues:0

Compositional_Deep_Learning

Deep learning via category theory and functional programming

Language:HaskellLicense:MITStargazers:138Issues:0Issues:0

RL-Theory-book

Reinforcement learning theory book about foundations of deep RL algorithms with proofs.

Language:TeXStargazers:269Issues:0Issues:0

AutoGPT

AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.

Language:PythonLicense:MITStargazers:167210Issues:0Issues:0