EleutherAI's repositories
lm-evaluation-harness
A framework for few-shot evaluation of language models.
concept-erasure
Erasing concepts from neural representations with provable guarantees
DeeperSpeed
DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.
nanoGPT-mup
The simplest, fastest repository for training/finetuning medium-sized GPTs.
polyapprox
Closed-form polynomial approximations to neural networks
transformer-reasoning
Experiments in transformer knowledge and reasoning
sae_overlap
Acompanying code for our research on SAE feature overlap when trained on different seeds.
aria-utils
MIDI tokenizers and pre-processing utils.
open-r1
Fully open reproduction of DeepSeek-R1
POSER
Poser: Unmasking Alignment Faking LLMs by Manipulating Their Internals
TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.