Eugene Yan (eugeneyan)

eugeneyan

Geek Repo

Company:Amazon

Location:Seattle

Home Page:eugeneyan.com

Twitter:@eugeneyan

Github PK Tool:Github PK Tool

Eugene Yan's starred repositories

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:33089Issues:335Issues:2557

dspy

DSPy: The framework for programming—not prompting—foundation models

Language:PythonLicense:MITStargazers:11798Issues:114Issues:474

minbpe

Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.

Language:PythonLicense:MITStargazers:8322Issues:79Issues:31

trl

Train transformer language models with reinforcement learning.

Language:PythonLicense:Apache-2.0Stargazers:8321Issues:75Issues:909
Language:PythonLicense:NOASSERTIONStargazers:7929Issues:149Issues:0

bitsandbytes

Accessible large language models via k-bit quantization for PyTorch.

Language:PythonLicense:MITStargazers:5553Issues:47Issues:938

instructor

structured outputs for llms

Language:PythonLicense:MITStargazers:5535Issues:47Issues:202

captum

Model interpretability and understanding for PyTorch

Language:PythonLicense:BSD-3-ClauseStargazers:4623Issues:245Issues:519

trlx

A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)

Language:PythonLicense:MITStargazers:4348Issues:49Issues:283

alignment-handbook

Robust recipes to align language models with human and AI preferences

Language:PythonLicense:Apache-2.0Stargazers:3960Issues:110Issues:115

higgsfield

Fault-tolerant, highly scalable GPU orchestration, and a machine learning framework designed for training models with billions to trillions of parameters

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:3264Issues:79Issues:1

picoGPT

An unnecessarily tiny implementation of GPT-2 in NumPy.

Language:PythonLicense:MITStargazers:3103Issues:29Issues:10

CTranslate2

Fast inference engine for Transformer models

Language:C++License:MITStargazers:2884Issues:56Issues:636

vocode-python

🤖 Build voice-based LLM agents. Modular + open source.

Language:PythonLicense:MITStargazers:2385Issues:42Issues:139

ChainForge

An open-source visual programming environment for battle-testing prompts to LLMs.

Language:TypeScriptLicense:MITStargazers:2037Issues:24Issues:164

AutoAWQ

AutoAWQ implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference. Documentation:

Language:PythonLicense:MITStargazers:1286Issues:11Issues:303

mup

maximal update parametrization (µP)

Language:Jupyter NotebookLicense:MITStargazers:1201Issues:29Issues:57

lm-human-preferences

Code for the paper Fine-Tuning Language Models from Human Preferences

Language:PythonLicense:MITStargazers:1127Issues:23Issues:15

hallucination-leaderboard

Leaderboard Comparing LLM Performance at Producing Hallucinations when Summarizing Short Documents

summarize-from-feedback

Code for "Learning to summarize from human feedback"

Language:PythonLicense:NOASSERTIONStargazers:952Issues:147Issues:21

obsidian-kindle-plugin

Sync your Kindle notes and highlights directly into your Obsidian vault

Language:TypeScriptLicense:MITStargazers:864Issues:7Issues:160

bleurt

BLEURT is a metric for Natural Language Generation based on transfer learning.

Language:PythonLicense:Apache-2.0Stargazers:653Issues:13Issues:49

COMET

A Neural Framework for MT Evaluation

Language:PythonLicense:Apache-2.0Stargazers:415Issues:17Issues:157

llm_distillation_playbook

Best practices for distilling large language models.

Language:Jupyter NotebookStargazers:286Issues:10Issues:0

Whisper-transcription_and_diarization-speaker-identification-

How to use OpenAIs Whisper to transcribe and diarize audio files

Language:Jupyter NotebookStargazers:244Issues:5Issues:8
Language:Jupyter NotebookLicense:Apache-2.0Stargazers:21Issues:2Issues:1

ft-drift

Check for data drift between two OpenAI multi-turn chat jsonl files.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:11Issues:2Issues:0

swe-study-group

Code for the SWE study group

Language:PythonLicense:Apache-2.0Stargazers:6Issues:2Issues:0
Language:PythonLicense:MITStargazers:6Issues:0Issues:0