There are 0 repository under efficient-llm topic.
A curated list for Efficient Large Language Models
[ICLR-2025-SLLM Spotlight 🔥]MobiLlama : Small Language Model tailored for edge devices
[ICML 2024] CLLMs: Consistency Large Language Models
[NAACL' 25 main] Lillama: Large Language Model Compression via Low-Rank Feature Distillation
There is a summary repo for Efficient AI direction. If you want to contribute to this repo, feel free to pr(pull request)!
Colab-friendly BitNet distillation engine: collect KD traces from a teacher, train a ternary Mini-BitNet, and dry-run 7B memory. Multi-provider + Drive/S3
A Curated Paper List for Efficient Large Models