python main.py --model decapoda-research/llama-7b-hf \
--prune_method wanda \
--density 0.5 \
--sparsity_type unstructured \
--learning_rate 0.0002 \
--eval_zero_shot \
Our implementation partially reuses Wanda's code.
Code for paper "EBFT: Effective and Block-Wise Fine-Tuning for Sparse LLMs"
python main.py --model decapoda-research/llama-7b-hf \
--prune_method wanda \
--density 0.5 \
--sparsity_type unstructured \
--learning_rate 0.0002 \
--eval_zero_shot \
Our implementation partially reuses Wanda's code.
Code for paper "EBFT: Effective and Block-Wise Fine-Tuning for Sparse LLMs"