Baohao Liao's starred repositories
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
generative-models
Generative Models by Stability AI
Chinese-LLaMA-Alpaca
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
lm-evaluation-harness
A framework for few-shot evaluation of language models.
RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
llm_interview_note
主要记录大语言大模型(LLMs) 算法(应用)工程师相关的知识及面试题
LLM-Adapters
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
Instruction-Tuning-Papers
Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).
unify-parameter-efficient-tuning
Implementation of paper "Towards a Unified View of Parameter-Efficient Transfer Learning" (ICLR 2022)
minimal-text-diffusion
A minimal implementation of diffusion models for text generation
academic-budget-bert
Repository containing code for "How to Train BERT with an Academic Budget" paper
DinkyTrain
Princeton NLP's pre-training library based on fairseq with DeepSpeed kernel integration 🚃
SeqDiffuSeq
Text Diffusion Model with Encoder-Decoder Transformers for Sequence-to-Sequence Generation [NAACL 2024]
PrefixQuant
An algorithm for static activation quantization of LLMs
StableMask
PyTorch implementation of StableMask (ICML'24)
chat-task-2024-data
Data for WMT 2024 Chat Shared Task
NLP-reproduction
Offer straightforward guidance to reproduce the results of NLP papers.