simeks / doris

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

doris

Conversational AI sandbox

Notes

Vicuna

Vicuna-13B is a fine-tuned version of LLaMA [5].

LLaMA/Alpaca

  • LLaMA [1]: 65B LLM by Meta
  • Alpaca [2]: fine-tuned version of LLaMA
  • GPTQ repo [3]
  • GPTQ-for-LLaMA LLaMA-specific implementation of GOTQ [3]

Weights

Install GPTQ-for-LLaMa

Assumes requirements.txt has been installed.

  1. Clone GPTQ-for-LLaMA: git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa.git into third_party folder.
  2. In GPTQ folder: CUDA_PATH=/usr/local/cuda-11.7 python setup_cuda.py install (assuming project environment is active)
  3. Test installation: CUDA_VISIBLE_DEVIES=0 python test_kernel.py

References

  1. Hugo Touvron, et al., LLaMA: Open and Efficient Foundation Language Models, https://arxiv.org/abs/2302.13971

  2. Rohan Taori, et al., Stanford Alpaca: An Instruction-following LLaMA model, https://github.com/tatsu-lab/stanford_alpaca

  3. Elias Frantar, et al., GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers, https://arxiv.org/abs/2210.17323

  4. Edward J. Hu, et al., LoRA: Low-Rank Adaptation of Large Language Models, https://arxiv.org/abs/2106.09685

  5. https://vicuna.lmsys.org/

About


Languages

Language:Python 100.0%