putizi-super's starred repositories
Awesome-ChatTTS
官方推荐的 ChatTTS 资源汇总项目,整理了全网相关资源和常见问题 || Officially recommended ChatTTS resource collection project
Mixture-of-depths
Unofficial implementation for the paper "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
mixture-of-depths
An unofficial implementation of "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
Awesome-Chinese-LLM
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
LLaMA3-Quantization
A repository dedicated to evaluating the performance of quantizied LLaMA3 using various quantization methods..
Mixture-of-Depths
Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
alpaca-lora
Instruct-tune LLaMA on consumer hardware
Efficient-LLMs-Survey
[TMLR 2024] Efficient Large Language Models: A Survey
tuning_playbook
A playbook for systematically maximizing the performance of deep learning models.
Awesome-Knowledge-Distillation-of-LLMs
This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & Vertical Distillation of LLMs.
Awesome-LLM-Prune
Awesome list for LLM pruning.
lm-evaluation-harness
A framework for few-shot evaluation of language models.
llama-recipes
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama3 for WhatsApp & Messenger.
PowerInfer
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
mmdetection
OpenMMLab Detection Toolbox and Benchmark
DeepSpeedExamples
Example models using DeepSpeed