zetavg / LLaMA-LoRA-Tuner

UI tool for fine-tuning and testing your own LoRA models base on LLaMA, GPT-J and more. One-click run on Google Colab. + A Gradio ChatGPT-like Chat UI to demonstrate your language models.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Support for Larger Models

scblaze opened this issue · comments

It would be great if there was a way to use this with the 13B, 30B or 60B LLaMa model sizes.

In theory, it will work by specifying a larger LLaMA base model via the --base_model flag, e.g. --base_model=decapoda-research/llama-13b-hf, then select a LoRA model that's trained on top of that base model (such as chansung/alpaca-lora-13b). However, I still need to test it. If you had the chance to try it first, your sharing of how it works would be appreciated! 🚀

BTW I think I'll be adding the ability to switch between base models without restarting the app and support non-llama models in the near future.

Update 2023/4/20: The ability of switching between base models has been added.

I can confirm that it working with llama-13b-h, it use 93% from 80GB VRAM A100, Lora trained successfully and working