4 bits quantization of LLaMA using GPTQ
Geek Repo:Geek Repo
Github PK Tool:Github PK Tool
Ted8000 opened this issue a year ago · comments
i use it for openllama, but th result generate seems not right.