codingchild2424 / alpaca-lora

Code for reproducing the Stanford Alpaca InstructLLaMA result on consumer hardware

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

🦙🌲🤏 Alpaca-LoRA: Low-Rank LLaMA Instruct-Tuning

This repository contains code for reproducing the Stanford Alpaca results using low-rank adaptation (LoRA). We aim to provide an Instruct model of similar quality to text-davinci-003 that can run on a Raspberry Pi (for research), but extensions to the 13b, 30b, and 65b models should be feasible with simple changes to the code.

In addition to the training code, which runs within five hours on a single RTX 4090, we publish the resulting LoRA weights on the Huggingface model hub, along with a script for downloading and inference on the foundation model and LoRA. For fine-tuning LoRAs we use Huggingface's PEFT as well as Tim Dettmers' bitsandbytes.

Without hyperparameter tuning or validation-based checkpointing, the LoRA model produces outputs comparable to the Stanford Alpaca model, though possibly with more minor mistakes. (Please see the outputs included below.) Further tuning might be able to achieve better performance; I invite interested users to give it a try and report their results.

As usual, I can be reached at https://twitter.com/ecjwg.

Setup

Until Jason Phang's LLaMA implementation is merged, users will need to replace their local transformers package.

  1. Install dependencies (install zphang's transformers fork)
pip install -q datasets loralib sentencepiece

pip uninstall transformers
pip install -q git+https://github.com/zphang/transformers@c3dc391

pip install -q git+https://github.com/huggingface/peft.git
  1. Install bitsandbytes from source.

Inference (generate.py)

This file reads the foundation model from the Huggingface model hub and the LoRA weights from tloen/alpaca-lora-7b, and runs inference on a specified input. Users should treat this as example code for the use of the model, and modify it as needed.

Training (finetune.py)

This file contains a straightforward application of PEFT to the LLaMA model, as well as some code related to prompt construction and tokenization. Near the top of this file is a set of hardcoded hyperparameters that you should feel free to modify. PRs adapting this code to multi-GPU setups and larger models are always welcome.

To do

  • Merge LoRA weights into LLaMA weights to remove inference dependency on PEFT
  • Train/val/test split
  • Hyperparameter tuning code
  • Documentation for notebook
  • Support for 13b, 30b, 65b
  • Train a version that doesn't waste tokens on the prompt header
  • Inference CLI and evaluation
  • Better disclaimers about why using LLaMA without permission is very bad!

Example outputs

Note errors and hallucinations here: Emmanuel Macron is not the king of France, territories are not provinces / listed in the wrong order, the FizzBuzz program is incorrect, and I'm unaware of a dialect in which 'awkward' rhymes with 'shock'.


Instruction: Tell me about alpacas.

Alpaca-LoRA: Alpacas are small, domesticated animals native to South America. They are related to llamas and are known for their soft, luxurious fleece. Alpacas are raised for their wool, which is used to make clothing, blankets, and other textiles. They are also raised for their meat and are often used as a source of milk. Alpacas are social animals and live in herds. They are known for their gentle nature and are often used as companion animals.

Stanford Alpaca: Alpacas are small, fluffy animals related to camels and llamas. They are native to Peru and Bolivia, and were first domesticated around 5,000 years ago. They are kept mainly for their fine, soft fleece, which is used to make knitwear and other garments. Alpacas are herd animals and live in small family groups, led by an older male. They are highly social animals and can form strong bonds with their owners.


Instruction: Tell me about the president of Mexico in 2019.

Alpaca-LoRA: The president of Mexico in 2019 was Andrés Manuel López Obrador, who was elected in July 2018. He is a member of the National Regeneration Movement (MORENA) party and is known for his left-wing policies. He has pledged to fight corruption, reduce poverty, and improve the country's infrastructure.

Stanford Alpaca: In 2019, Andres Manuel Lopez Obrador was sworn in as President of Mexico. He is the country’s 58th president and the first from the left-wing Morena party. He was elected in 2018 on a promise to combat corruption and reduce poverty and has faced criticism for his handling of the economy, crime and the country’s relationship with the United States. He is also known for his ambitious social welfare programs and has been praised for raising the minimum wage and providing aid to low-income families. He is seen as a champion of democracy and human rights in Mexico.


Instruction: Tell me about the king of France in 2019.

Alpaca-LoRA: The king of France in 2019 is Emmanuel Macron. He was elected in 2017 and is the youngest president in French history. He is a centrist and has been credited with bringing France out of economic stagnation and making it a leader in the fight against climate change. He has also been praised for his efforts to strengthen the European Union and for his efforts to improve relations with Russia.

Stanford Alpaca: In 2019, the King of France was Emmanuel Macron. He was sworn in as President of France on May 14, 2017, and has been leading the country ever since. He is a member of the centrist La République En Marche! Party. He is known for his reformist agenda and has faced a number of protests from both the right and left of the political spectrum. He is the youngest French president in history.


Instruction: List all Canadian provinces in alphabetical order.

Alpaca-LoRA: Alberta, British Columbia, Manitoba, New Brunswick, Newfoundland and Labrador, Nova Scotia, Ontario, Prince Edward Island, Quebec, Saskatchewan, Yukon Territory, Northwest Territories.

Stanford Alpaca: Alberta, British Columbia, Manitoba, New Brunswick, Newfoundland and Labrador, Nova Scotia, Ontario, Prince Edward Island, Quebec, Saskatchewan, Northwest Territories, Nunavut, Yukon.


Instruction: Write a Python program that prints the first 10 Fibonacci numbers.

Alpaca-LoRA:

# Python program that prints the first 10 Fibonacci numbers

def fibonacci(n):
    if n == 0:
        return 0
    elif n == 1:
        return 1
    else:
        return fibonacci(n - 1) + fibonacci(n - 2)

for i in range(10):
    print(fibonacci(i))

Stanford Alpaca:

numbers = [0, 1]

for i in range(2, 11):
    numbers.append(numbers[i-2] + numbers[i-1])

print(numbers)
# Output: [0, 1, 1, 2, 3, 5, 8, 13, 21, 34]

Instruction: Write a program that prints the numbers from 1 to 100. But for multiples of three print 'Fizz' instead of the number and for the multiples of five print 'Buzz'. For numbers which are multiples of both three and five print 'FizzBuzz'.

Alpaca-LoRA:

def print_numbers(start, end):
    for i in range(start, end):
        if i % 3 == 0 or i % 5 == 0:
            print("FizzBuzz")
        else:
            print(i)

Stanford Alpaca:

for i in range(1, 101):
    if i % 3 == 0 and i % 5 == 0:
        print('FizzBuzz')
    elif i % 3 == 0:
        print('Fizz')
    elif i % 5 == 0:
        print('Buzz')
    else:
        print(i)

Instruction: Tell me five words that rhyme with 'shock'.

Alpaca-LoRA: Awkward, bark, chock, lock, and rock.

Stanford Alpaca: Five words that rhyme with shock are: rock, pop, shock, cook, and snock.


Instruction: Translate the sentence 'I have no mouth but I must scream' into Spanish.

Alpaca-LoRA: No tengo boca pero tengo que gritar.

Stanford Alpaca: No tengo boca, pero debo gritar.

About

Code for reproducing the Stanford Alpaca InstructLLaMA result on consumer hardware

License:Apache License 2.0


Languages

Language:Jupyter Notebook 91.6%Language:Python 8.4%