Zhou Yu's starred repositories
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
llama-recipes
Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger
DeepSpeedExamples
Example models using DeepSpeed
Baichuan-7B
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
Informer2020
The GitHub repository for the paper "Informer" accepted by AAAI 2021.
pytorch-GAT
My implementation of the original GAT paper (Veličković et al.). I've additionally included the playground.py file for visualizing the Cora dataset, GAT embeddings, an attention mechanism, and entropy histograms. I've supported both Cora (transductive) and PPI (inductive) examples!
FlagAttention
A collection of memory efficient attention operators implemented in the Triton language.
hyperparameter
Hyperparameter, Make configurable AI applications.Build for Python hackers.