Yu Zhang's starred repositories
open_llama
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
lm-evaluation-harness
A framework for few-shot evaluation of language models.
IP-Adapter
The image prompt adapter is designed to enable a pretrained text-to-image diffusion model to generate images with image prompt.
distil-whisper
Distilled variant of Whisper for speech recognition. 6x faster, 50% smaller, within 1% word error rate.
consistencydecoder
Consistency Distilled Diff VAE
cuda_programming
Code from the "CUDA Crash Course" YouTube series by CoffeeBeforeArch
gpt_paper_assistant
GPT4 based personalized ArXiv paper assistant bot
AutoCompressors
[EMNLP 2023] Adapting Language Models to Compress Long Contexts
flash-fft-conv
FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores
aft-pytorch
Unofficial PyTorch implementation of Attention Free Transformer (AFT) layers by Apple Inc.
triton-transformer
Implementation of a Transformer, but completely in Triton
CoLT5-attention
Implementation of the conditionally routed attention in the CoLT5 architecture, in Pytorch
ModuleFormer
ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward experts. We released a collection of ModuleFormer-based Language Models (MoLM) ranging in scale from 4 billion to 8 billion parameters.
heinsen_sequence
Code implementing "Efficient Parallelization of a Ubiquitious Sequential Computation" (Heinsen, 2023)
torchscale
Transformers at any scale
fairseq-evo
Fairseq with transformer evolution