Why does the model quantization prompt KILLED at the end?
g558800 opened this issue · comments
It's a RAM issue, try to increase the RAM size, or add a memory swap
Resolved. thanks
4 bits quantization of LLaMA using GPTQ
g558800 opened this issue · comments
It's a RAM issue, try to increase the RAM size, or add a memory swap
Resolved. thanks