Minsoo Kim's repositories
kd-qat-large-enc
[EMNLP 2022 main] Code for "Understanding and Improving Knowledge Distillation for Quantization-Aware-Training of Large Transformer Encoders"
mbv1_brevitas
This is MobileNetv1 Brevitas based Quantization-aware-Training Framework
efficientdet-pytorch
A PyTorch impl of EfficientDet faithful to the original Google impl w/ ported weights
finetune-transformer-lm
Code and model for the paper "Improving Language Understanding by Generative Pre-Training"
llm-awq
AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
lsq-net
Unofficial implementation of LSQ-Net, a neural network quantization framework
model-quantization
Collections of model quantization algorithms
Pretrained-Language-Model
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
TernGEMM
TernGEMM: General Matrix Multiply Library with Ternary Weights for Fast DNN Inference
TSLD
[NeurIPS 2023] Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
Yet-Another-EfficientDet-Pytorch
The pytorch re-implement of the official efficientdet with SOTA performance in real time and pretrained weights.