Eugene Yan (eugeneyan)

eugeneyan

Geek Repo

Company:Amazon

Location:Seattle

Home Page:eugeneyan.com

Twitter:@eugeneyan

Github PK Tool:Github PK Tool

Eugene Yan's starred repositories

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:32636Issues:328Issues:2504

dspy

DSPy: The framework for programming—not prompting—foundation models

Language:PythonLicense:MITStargazers:10415Issues:101Issues:396

semantic

Parsing, analyzing, and comparing source code across many languages

trl

Train transformer language models with reinforcement learning.

Language:PythonLicense:Apache-2.0Stargazers:8088Issues:74Issues:871

minbpe

Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.

Language:PythonLicense:MITStargazers:7923Issues:78Issues:27
Language:PythonLicense:NOASSERTIONStargazers:7770Issues:148Issues:0

accelerate

🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support

Language:PythonLicense:Apache-2.0Stargazers:6956Issues:100Issues:1352

axolotl

Go ahead and axolotl questions

Language:PythonLicense:Apache-2.0Stargazers:5719Issues:47Issues:520

bitsandbytes

Accessible large language models via k-bit quantization for PyTorch.

Language:PythonLicense:MITStargazers:5382Issues:46Issues:920

instructor

structured outputs for llms

Language:PythonLicense:MITStargazers:4936Issues:36Issues:181

captum

Model interpretability and understanding for PyTorch

Language:PythonLicense:BSD-3-ClauseStargazers:4567Issues:239Issues:513

trlx

A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)

Language:PythonLicense:MITStargazers:4317Issues:49Issues:282

alignment-handbook

Robust recipes to align language models with human and AI preferences

Language:PythonLicense:Apache-2.0Stargazers:3763Issues:110Issues:109

higgsfield

Fault-tolerant, highly scalable GPU orchestration, and a machine learning framework designed for training models with billions to trillions of parameters

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:3252Issues:79Issues:1

picoGPT

An unnecessarily tiny implementation of GPT-2 in NumPy.

Language:PythonLicense:MITStargazers:3082Issues:29Issues:10

CTranslate2

Fast inference engine for Transformer models

Language:C++License:MITStargazers:2790Issues:56Issues:611

AutoAWQ

AutoAWQ implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference. Documentation:

Language:PythonLicense:MITStargazers:1185Issues:11Issues:283

lm-human-preferences

Code for the paper Fine-Tuning Language Models from Human Preferences

Language:PythonLicense:MITStargazers:1107Issues:23Issues:15

hallucination-leaderboard

Leaderboard Comparing LLM Performance at Producing Hallucinations when Summarizing Short Documents

summarize-from-feedback

Code for "Learning to summarize from human feedback"

Language:PythonLicense:NOASSERTIONStargazers:948Issues:145Issues:21

obsidian-kindle-plugin

Sync your Kindle notes and highlights directly into your Obsidian vault

Language:TypeScriptLicense:MITStargazers:846Issues:7Issues:157

bigcode-evaluation-harness

A framework for the evaluation of autoregressive code generation language models.

Language:PythonLicense:Apache-2.0Stargazers:627Issues:13Issues:111

Platypus

Code for fine-tuning Platypus fam LLMs using LoRA

llm_distillation_playbook

Best practices for distilling large language models.

Language:Jupyter NotebookStargazers:279Issues:10Issues:0

datagen

a pipeline for using api calls to agnostically convert unstructured data into structured training data

Language:PythonStargazers:26Issues:0Issues:0
Language:Jupyter NotebookLicense:Apache-2.0Stargazers:20Issues:2Issues:0

get-lambda

Use Actions to acquire those precious lambda GPUs

Language:PythonStargazers:19Issues:2Issues:0

swe-study-group

Code for the SWE study group

Language:PythonLicense:Apache-2.0Stargazers:6Issues:2Issues:0