dongguanting / INTERS

This is the repository for our paper "INTERS: Unlocking the Power of Large Language Models in Search with Instruction Tuning"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

INTERS: Unlocking the Power of Large Language Models in Search with Instruction Tuning

license

📃 ArXiv Paper • 🤗 HuggingFace Model • 📚 Dataset

Authors: Yutao Zhu, Peitian Zhang, Chenghao Zhang, Yifei Chen, Binyu Xie, Zhicheng Dou, Zheng Liu, and Ji-Rong Wen

We will release the datasets, models, templates, and codes within a month (before Feb. 15th). Thanks for your attention!

Introduction

Large language models (LLMs) have demonstrated impressive capabilities in various natural language processing tasks. Despite this, their application to information retrieval (IR) tasks is still challenging due to the infrequent occurrence of many IR-specific concepts in natural language. While prompt-based methods can provide task descriptions to LLMs, they often fall short in facilitating comprehensive understanding and execution of IR tasks, thereby limiting LLMs' applicability. To address this gap, in this work, we explore the potential of instruction tuning to enhance LLMs' proficiency in IR tasks. We introduce a novel instruction tuning dataset, \ourdata{}, encompassing 20 tasks across three fundamental IR categories: query understanding, document understanding, and query-document relationship understanding. The data are derived from 43 distinct datasets with manually written templates. Our empirical results reveal that \ourdata{} significantly boosts the performance of various publicly available LLMs, such as LLaMA, Mistral, and Phi, in search-related tasks. Furthermore, we conduct a comprehensive analysis to ascertain the effects of base model selection, instruction design, volume of instructions, and task variety on performance.

Our dataset and the models fine-tuned on it will be released soon!

Citation

Please kindly cite our paper if it helps your research:

@article{Inters,
    author={Yutao Zhu and
            Peitian Zhang and
            Chenghao Zhang and
            Yifei Chen and
            Binyu Xie and
            Zhicheng Dou and
            Zheng Liu and
            Ji-Rong Wen},
    title={INTERS: Unlocking the Power of Large Language Models in Search with Instruction Tuning},
    journal={CoRR},
    volume={abs/2401.06532},
    year={2024},
    url={https://arxiv.org/abs/2401.06532},
    eprinttype={arXiv},
    eprint={}
}

About

This is the repository for our paper "INTERS: Unlocking the Power of Large Language Models in Search with Instruction Tuning"

License:MIT License