EleutherAI

EleutherAI

Geek Repo

Location:The Internet

Home Page:www.eleuther.ai

Twitter:@AIEleuther

Github PK Tool:Github PK Tool

EleutherAI's repositories

gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries

Language:PythonLicense:Apache-2.0Stargazers:6624Issues:119Issues:423

lm-evaluation-harness

A framework for few-shot evaluation of language models.

Language:PythonLicense:MITStargazers:5233Issues:35Issues:830

pythia

The hub for EleutherAI's work on interpretability and learning dynamics

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:2069Issues:32Issues:96
Language:PythonLicense:MITStargazers:984Issues:16Issues:43

concept-erasure

Erasing concepts from neural representations with provable guarantees

Language:PythonLicense:MITStargazers:193Issues:9Issues:5

cookbook

Deep learning for dummies. All the practical details and useful utilities that go into working with real models.

Language:PythonLicense:Apache-2.0Stargazers:180Issues:6Issues:10

elk

Keeping language models honest by directly eliciting knowledge encoded in their activations.

Language:PythonLicense:MITStargazers:171Issues:6Issues:88

DeeperSpeed

DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:160Issues:4Issues:0

improved-t5

Experiments for efforts to train a new and improved t5

minetest

Minetest is an open source voxel game engine with easy modding and game creation

Language:C++License:NOASSERTIONStargazers:57Issues:3Issues:31
Language:PythonLicense:Apache-2.0Stargazers:36Issues:3Issues:0
Language:Jupyter NotebookStargazers:34Issues:4Issues:0

rnngineering

Engineering the state of RNN language models (Mamba, RWKV, etc.)

Language:Jupyter NotebookLicense:MITStargazers:31Issues:4Issues:1

features-across-time

Understanding how features learned by neural networks evolve throughout training

Language:PythonLicense:MITStargazers:25Issues:4Issues:0

elk-generalization

Investigating the generalization behavior of LM probes trained to predict truth labels: (1) from one annotator to another, and (2) from easy questions to hard

Language:PythonLicense:MITStargazers:20Issues:2Issues:1

tokengrams

Efficiently computing & storing token n-grams from large corpora

Language:RustLicense:MITStargazers:7Issues:2Issues:0

aria-amt

Efficient and robust implementation of seq-to-seq automatic piano transcription.

Language:PythonLicense:Apache-2.0Stargazers:6Issues:2Issues:0
Language:PythonLicense:MITStargazers:3Issues:3Issues:1

variance-across-time

Studying the variance in neural net predictions across training time

Language:PythonLicense:MITStargazers:3Issues:2Issues:0

website

New website for EleutherAI based on Hugo static site generator

Language:HTMLStargazers:3Issues:2Issues:0

tuned-lens

Tools for understanding how transformer predictions are built layer-by-layer

Language:PythonLicense:MITStargazers:2Issues:1Issues:0

bayesian-adam

Exactly what it says on the tin

Language:PythonLicense:MITStargazers:1Issues:2Issues:0

conceptual-constraints

Applying LEACE to models during training

Language:Jupyter NotebookLicense:MITStargazers:1Issues:3Issues:0
Language:Jupyter NotebookLicense:Apache-2.0Stargazers:1Issues:2Issues:0

RWKV-LM

RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.

Language:PythonLicense:Apache-2.0Stargazers:1Issues:1Issues:0

CAA

Steering Llama 2 with Contrastive Activation Addition

Language:Jupyter NotebookLicense:MITStargazers:0Issues:1Issues:0

cupbearer

A library for mechanistic anomaly detection

License:MITStargazers:0Issues:0Issues:0

TransformerEngine

A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0
Language:PythonLicense:MITStargazers:0Issues:0Issues:0