This is the repository for the paper "StructLM: Towards Building Generalist Models for Structured Knowledge Grounding".
You can use this repository to evaluate the models. To reproduce the models, use SKGInstruct in your preferred finetuning framework.
The processed test data is already provided, but the prompts used for training and testing can be found in /prompts
- Arxiv Link: https://arxiv.org/abs/2402.16671
- Website: https://tiger-ai-lab.github.io/StructLM/
We added StructLM-7B-mistral, which is stronger than both Llama-based 7B and 13B models in many cases.
Requirements:
- Python 3.10
- Linux
- support for CUDA 11.8
pip install -r requirements.txt
./download.sh
this will download
- StructLM-7B
- The raw data required for executing evaluation
- The processed test data splits ready for evaluation
./run_test_eval.sh StructLM-7B
this will generate the results in
outputs/StructLM-7B/
You can also replace StructLM-7B
with StructLM-13B
or StructLM-34B
, i.e.
./run_test_eval.sh StructLM-13B`
./run_test_eval.sh StructLM-34B
and download those models separately.
The evaluation metrics in this repository were adapted and modified from the evaluation files found in https://github.com/HKUNLP/UnifiedSKG
@misc{zhuang2024structlm,
title={StructLM: Towards Building Generalist Models for Structured Knowledge Grounding},
author={Alex Zhuang and Ge Zhang and Tianyu Zheng and Xinrun Du and Junjie Wang and Weiming Ren and Stephen W. Huang and Jie Fu and Xiang Yue and Wenhu Chen},
year={2024},
eprint={2402.16671},
archivePrefix={arXiv},
primaryClass={cs.CL}
}