Benjamin Warner's starred repositories
tuning_playbook
A playbook for systematically maximizing the performance of deep learning models.
flash-attention
Fast and memory-efficient exact attention
ml-engineering
Machine Learning Engineering Open Book
LLMsPracticalGuide
A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)
llama-recipes
Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger
latexify_py
A library to generate LaTeX expression from Python code.
bitsandbytes
Accessible large language models via k-bit quantization for PyTorch.
RAGatouille
Easily use and train state of the art late-interaction retrieval methods (ColBERT) in any RAG pipeline. Designed for modularity and ease-of-use, backed by research.
mamba-minimal
Simple, minimal implementation of the Mamba SSM in one file of PyTorch.
pillow-simd
The friendly PIL fork
sd-akashic
A compendium of informations regarding Stable Diffusion (SD)
awesome-stable-diffusion
Curated list of awesome resources for the Stable Diffusion AI Model.
stable-fast
Best inference performance optimization framework for HuggingFace Diffusers on NVIDIA GPUs.
the-art-of-debugging
The Art of Debugging
CushyStudio
🛋 The AI and Generative Art platform for everyone
CLIP-Guided-Diffusion
Just playing with getting CLIP Guided Diffusion running locally, rather than having to use colab.
multipack_sampler
Multipack distributed sampler for fast padding-free training of LLMs
triton-autodiff
Experiment of using Tangent to autodiff triton