mutonix / RefGPT

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

RefGPT

[Paper] RefGPT

General chat models, like ChatGPT, have attained impressive capability to resolve a wide range of NLP tasks by tuning Large Language Models (LLMs) with high-quality instruction data. However, collecting human-written high-quality data, especially multi-turn dialogues, is expensive and unattainable for most people. Though previous studies have used powerful LLMs to generate the dialogues automatically, but they all suffer from generating untruthful dialogues because of the LLMs hallucination. Therefore, we propose a method called RefGPT to generate enormous truthful and customized dialogues without worrying about factual errors caused by the model hallucination. RefGPT solves the model hallucination in dialogue generation by restricting the LLMs to leverage the given reference instead of reciting their own knowledge to generate dialogues. Additionally, RefGPT adds detailed controls on every utterances to enable highly customization capability, which previous studies have ignored. On the basis of RefGPT, we also propose two high-quality dialogue datasets generated by GPT-4, namely RefGPT-Fact and RefGPT-Code. RefGPT-Fact is 100k multi-turn dialogue datasets based on factual knowledge and RefGPT-Code is 76k multi-turn dialogue dataset covering a wide range of coding scenarios.

News


  • ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅFeb 2024, we updated two new datasets to RefGPT dataset series.

    • We supplement RefGPT-Fact with 60.7k Chinese and 58.6k English factual multi-turn dialogues using the RefGPT method, denoted as RefGPT-Fact-v2 [download].
    • We also construct a new reasoning dataset with 288k Chinese reasoning QAs, denoted as RefGPT-Reason [download].
  • Oct, Congratulations!๐ŸŽ‰๐ŸŽ‰๐ŸŽ‰ Our paper has been accepted by EMNLP 2023 findings.

  • June 9th, we released the code of generating the data in RefGPT-Fact and provided an efficient implementation of calling the OpenAI API in parallel even with multiple API keys.

  • May 26th, we released the RefGPT-Fact dataset and the RefGPT-Code dataset on the huggingface hub. [download]

  • May 25th, we released the paper and annouced the new RefGPT-Fact dataset and the RefGPT-Code dataset.

  • April 15th, we released the initial version of RefGPT-Fact-CN dataset.

Proposed Method

RefGPT

Generate Your Own Dataset with RefGPT

pip install requirements.txt

python parallel_generate-refgpt-fact.py \
      --reference_filepaths wiki_demo.jsonl \
      --api_config api_config.jsonl \
      --save_filepath output.jsonl \
      --num_chat_to_generate 20000 \
      --language en \
      --assistant_word_count 300 \
      --human_word_count 100 \
      --num_turn_ratios 0 1 2 2 0 \
      --max_attempts 5 \
      --max_requests_per_minute 1000 \
      --max_tokens_per_minute 40000 

You can tune the parameters of max_requests_per_minute or max_tokens_per_minute to adjust the calling speed. You can see a demo of setting API request format in api_config.jsonl file.

Released Datasets

RefGPT-Fact [download]

RefGPT-Fact is a datasets containing 100k multi-turn dialogues about factual knowledge with 50k English and 50k Chinese. The English version uses the English Wikipedia as the reference and the Chinese version uses the frequently-used Chinese online encyclopedia website, Baidu Baike.

RefGPT-Fact-example

RefGPT-Fact-example

RefGPT-Code [download[ds][cr][bg]]

RefGPT-Code is a dataset containing 76k multi-turn dialogues about programming with 37k English and 39k Chinese, which has covered most aspects of code usage scenarios and multiple types of programming languages, as shown in Figure 2. Both the English version and Chinese version use the public GitHub dataset on Google BiqQuery with no overlap in these two languages. RefGPT-Code has derived various ways of leveraging the program code as the reference to enable different scenarios. We consider three perspectives of code discussion, code creation and bug fixing in RefGPT-Code

RefGPT-Code

RefGPT-Code-ds (discussion)

RefGPT-Fact-example

RefGPT-Fact-example

RefGPT-Code-cr (creation)

RefGPT-Fact-example

RefGPT-Fact-example

RefGPT-Code-bg (bug fixing)

RefGPT-Fact-example

RefGPT-Fact-example

Attention

As the datasets RefGPT-Fact and RefGPT-Code are collected by using the references like Wikipedia and Github repositories, it can not be avoided that the reference itself has factual errors, typos, or bugs and malicious code if it is from Github repositories. The datasets may also reflect the biases of the selected references and GPT-3.5/GPT-4 model

Please pay attention that RefGPT Datasets, including RefGPT-Fact and RefGPT-Code, have not undergone manual verification, and as such, their security cannot be strictly guaranteed. Users should be aware that they are responsible for the results generated using this data.

Contributors

  1. Dongjie Yang, djyang.tony@sjtu.edu.cn
  2. Ruifeng Yuan, ruifeng.yuan@connect.polyu.hk
  3. Yuantao Fan, yuantaofan@bupt.edu.cn
  4. Yifei Yang, yifeiyang@sjtu.edu.cn
  5. Zili Wang, ziliwang.do@gmail.com
  6. Shusen Wang, wssatzju@gmail.com
  7. Hai Zhao,zhaohai@cs.sjtu.edu.cn

Citation

If your work take inspiration from or make use of our method or data, we kindly request that you acknowledge and cite our paper as a reference.

@misc{yang2023refgpt,
      title={RefGPT: Reference -> Truthful & Customized Dialogues Generation by GPTs and for GPTs}, 
      author={Dongjie Yang and Ruifeng Yuan and YuanTao Fan and YiFei Yang and Zili Wang and Shusen Wang and Hai Zhao},
      year={2023},
      eprint={2305.14994},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

About

License:Apache License 2.0


Languages

Language:Python 100.0%