There are 0 repository under small-models topic.
[ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Overview of self-supervised learning of tiny models, including distillation-based methods (aks. self-supervised distillation) and non-distillation methods.
Help us define the Pareto front of small models for MNIST classification. Frugal AI.