wanghz18's repositories
exllama
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
Language:PythonMIT000
GPTQ-triton
GPTQ inference Triton kernel
Language:Jupyter NotebookApache-2.0000
Language:Python000
000