DARK-art108 / QLora-101

Quantized LLMs with Low-Rank Adapters, Fine Tuning GPT-NeoX-12B

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

QLora-101 : Quantized LLMs with Low-Rank Adapters, Fine Tuning GPT-NeoX-12B

LoRa adds a tiny amount of trainable parameters, i.e., adapters, for each layer of the LLM and freezes all the original parameters. For fine-tuning, we only have to update the adapter weights which significantly reduces the memory footprint.

alt text

About

Quantized LLMs with Low-Rank Adapters, Fine Tuning GPT-NeoX-12B


Languages

Language:Jupyter Notebook 100.0%