Zewen Chi's starred repositories
LLaMA-Adapter
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
GPTQ-for-LLaMa
4 bits quantization of LLaMA using GPTQ
flops-counter.pytorch
Flops counter for convolutional networks in pytorch framework
LLMAgentPapers
Must-read Papers on LLM Agents.
prompt-in-context-learning
Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date and cutting-edge updates.
DeepSeek-LLM
DeepSeek LLM: Let there be answers
bigscience
Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.
proteinnet
Standardized data set for machine learning of protein structure
ModuleFormer
ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward experts. We released a collection of ModuleFormer-based Language Models (MoLM) ranging in scale from 4 billion to 8 billion parameters.
awesome-jekyll-websites
Awesome list of jekyll website and awesome resources. Create a pull request and add your jekyll website!
OpenAlpaca
OpenAlpaca: A Fully Open-Source Instruction-Following Model Based On OpenLLaMA