Liangyu Chen's repositories
random_hacks
Random hacks that I need to keep happy
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Emu
Emu: An Open Multimodal Generalist
fast-stable-diffusion
fast-stable-diffusion, +25-50% speed increase + memory efficient + DreamBooth
gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
litellm
Call all LLM APIs using the OpenAI format. Use Azure, OpenAI, Cohere, Anthropic, Ollama, VLLM, Sagemaker, HuggingFace, Replicate (100+ LLMs)
llama-recipes
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama3 for WhatsApp & Messenger.
llama3
the main Llama 3 GitHub site - will be moved under Meta-Llama
LLaVA
Visual Instruction Tuning: Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities.
LLMSpeculativeSampling
Fast inference from large lauguage models via speculative decoding
Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
OFA
Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
open_flamingo
An open-source framework for training large multimodal models.
optimum
🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools
visitor-badge
A badge generator service to count visitors of your markdown file.
visual-chatgpt
VisualChatGPT
yang-song.github.io
Personal website