Data and code for our paper FreshLLMs: Refreshing Large Language Models with Search Engine Augmentation.
Next update: Jan 19, 2024
We update our dataset weekly or upon request. If you find any updates or misclassifications in our FreshQA
questions or answers that we may have overlooked, please notify us by commenting on the dataset spreadsheet above or sending an email to freshllms@google.com.
To facilitate future evaluations, we have developed FreshEval
, a simple automatic metric that uses few-shot in-context learning to teach an LLM to judge model responses, which achieved high agreement with human raters (see Appendix B in our paper for details).
To use FreshEval
under a specific evaluation mode (Relaxed
or Strict
), please follow the instructions below:
- Make a copy of our latest data spreadsheet and store it in your Google Drive with a new filename (e.g.,
fresheval_relaxed
orfresheval_strict
). - Insert 3 new columns
D
,E
,F
in the new spreadsheet for model responses, evaluation rating, evaluation explanation, respectively and save your model's responses in columnD
(see our sample evaluation spreadsheet below). - Run the associated
FreshEval
notebook with the evaluation mode. Note that for demonstration purposes, we evaluated only the first 10 model responses. You can adjust the number as needed.
Here are our FreshEval
notebooks.
After obtaining TRUE/FALSE ratings for model responses, follow the instructions below to calculate the accuracy for each question category:
- Download our latest data spreadsheet in a Comma Separated Values (.csv) format and store it as
freshqa.csv
. - Open this notebook and upload the
freshqa.csv
file to the session storage (Files > Upload file to session storage
). - Replace the existing ratings in the notebook with your ratings and run the notebook.
We thank Peter Hart and Filipe Mesquita for their help in updating our FreshQA
questions/answers.
We are grateful to the following people for their contributions to creating our original FreshQA dataset: Marzena Karpinska, Dustin Tran, Daniel Cer, Sam Fullerton, Elizabeth Clark, Nishant Raj, Xiaoyu Song, Yapei Chang, Yixiao Song, Nader Akoury, Ankita Gupta, Bill Ray, Chau Pham, Wenlong Zhao, Maximilian Mozes, Simeng Sun, Ronan Salz, Kalpesh Krishna, Katherine Thai, Kanishka Misra, Salaheddin Alzu'bi, Erica Cai, Thibault Sellam, Jiao Sun, Dhruv Agarwal, Tessa Masis, Andrew Drozdov, Brian Lester, George Wei, Naveen Jafer Nizar, Shufan Wang, Youngwoo Kim, and Shib Sankar Dasgupta.
We are grateful to SerpApi for their generous sponsorship of 20,000 searches for FreshPrompt
.
If you use our data or method, please cite our paper:
@misc{vu2023freshllms,
title={FreshLLMs: Refreshing Large Language Models with Search Engine Augmentation},
author={Tu Vu and Mohit Iyyer and Xuezhi Wang and Noah Constant and Jerry Wei and Jason Wei and Chris Tar and Yun-Hsuan Sung and Denny Zhou and Quoc Le and Thang Luong},
year={2023},
eprint={2310.03214},
archivePrefix={arXiv},
primaryClass={cs.CL}
}