The quantization of the compressed models
lihuang258 opened this issue · comments
Liguangyan @UCAS commented
If I want to further quantize the pruned model, how should I proceed? I saw this mentioned in the paper
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support LLaMA, Llama-2, BLOOM, Vicuna, Baichuan, etc.
lihuang258 opened this issue · comments
If I want to further quantize the pruned model, how should I proceed? I saw this mentioned in the paper