sunggo / EBFT

Code for paper "EBFT: Effective and Block-Wise Fine-Tuning for Sparse LLMs"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

EBFT: Effective and Block-Wise Fine-Tuning for Sparse LLMs

fine-tuning

1. LlamaV1-7B
python main.py --model decapoda-research/llama-7b-hf \
    --prune_method wanda \
    --density 0.5 \
    --sparsity_type unstructured \
    --learning_rate 0.0002 \
    --eval_zero_shot \

Acknowledgments

Our implementation partially reuses Wanda's code.

About

Code for paper "EBFT: Effective and Block-Wise Fine-Tuning for Sparse LLMs"


Languages

Language:Python 100.0%