Paper | Corpora | Human Data | Retrieval | Generation | Synthetic Data | Getting Started | Contact
We present MTRAG, a comprehensive and diverse human-generated multi-turn RAG dataset, accompanied by four document corpora. To the best of our knowledge, MTRAG is the first end-to-end human-generated multi-turn RAG benchmark that reflects real-world properties of multi-turn conversations.
The paper describing the benchmark and experiments is available on Arxiv:
MTRAG: A Multi-Turn Conversational Benchmark for Evaluating Retrieval-Augmented Generation Systems
Yannis Katsis, Sara Rosenthal, Kshitij Fadnis, Chulaka Gunasekara, Young-Suk Lee, Lucian Popa, Vraj Shah, Huaiyu Zhu, Danish Contractor, Marina Danilevsky
If you use MTRAG, please cite the paper as follows:
@misc{katsis2025mtrag,
title={MTRAG: A Multi-Turn Conversational Benchmark for Evaluating Retrieval-Augmented Generation Systems},
author={Yannis Katsis and Sara Rosenthal and Kshitij Fadnis and Chulaka Gunasekara and Young-Suk Lee and Lucian Popa and Vraj Shah and Huaiyu Zhu and Danish Contractor and Marina Danilevsky},
year={2025},
eprint={2501.03468},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2501.03468},
}
Our benchmark is built on document corpora from 4 domains: ClapNQ, Cloud, FiQA and Govt. ClapNQ and FiQA are existing corpora from QA/IR datasets, while Govt and Cloud are new corpora assembled specifically for this benchmark.
Important
Download and uncompress the files to use the corpora.
Corpus | Domain | Data | # Documents | # Passages |
---|---|---|---|---|
ClapNQ [1] | Wikipedia | Corpus | 4,293 | 183,408 |
Cloud | Technical Documentation | Corpus | 57,638 | 61,022 |
FiQA [2] | Finance | Corpus | 7,661 | 49,607 |
Govt | Government | Corpus | 8,578 | 72,422 |
MTRAG consists of 110 multi-turn conversations that are converted to 842 evaluation tasks.
- diverse question types
- answerable, unanswerable, partial, and conversational questions
- multi-turn: follow-up and clarification
- four domains
- relevant and irrelevant passages (irrelevant passages were not enforced but ones that exist can be used as hard negatives)
We provide our benchmark of 110 conversations in conversation format HERE.
They average 7.7 turns per conversation. Each conversation is on a single corpus domain and includes a variety of question types, answerability and multi-turn dimensions. All conversations created by our annotators have gone through a review phase to ensure they are of high quality. During the review phase annotators could accept or reject conversations, and repair responses, passage relevance, and enrichments as needed. They were not allowed to edit the questions or passages as such changes could negatively affect the conversation flow.
The retrieval task per domain in BEIR format on the Answerable and Partial tasks only. Additional details are available in the retrieval README.
Name | Corpus | Queries |
---|---|---|
ClapNQ | Corpus | Queries |
Cloud | Corpus | Queries |
FiQA | Corpus | Queries |
Govt | Corpus | Queries |
The conversations are converted into 842 tasks. A task is a conversation turn containing all previous turns together with the last user question (e.g., the task created for turn
Setting | Description | File |
---|---|---|
Reference | Generation using reference passages | reference.jsonl |
Reference + RAG | Retrieval followed by generation but with the reference passages kept in the top 5 passages | reference+RAG.jsonl |
Full RAG | Retrieval followed by generation where retrieval results consist of the top 5 passages | RAG.jsonl |
We provide generation results in the analytics files for the experiments provided in our paper.
Setting | Description | File |
---|---|---|
Reference | Generation using reference passages | reference.json |
Reference + RAG | Retrieval followed by generation but with the reference passages kept in the top 5 passages | reference+RAG.json |
Full RAG | Retrieval followed by generation where retrieval results consist of the top 5 passages | RAG.json |
Human Evaluation Reference | Generation using reference passages on a subset with human evaluation | reference_subset_with_human_evaluations.json |
Manually creating data is an expensive and time consuming process that does not scale well. Automating this process has become popular via synthetic data generation. For further details of creation see this paper.
We provide 200 synthetically generated conversations that follow the properties of the human data. The conversations are available HERE
Setting | Description | File |
---|---|---|
Reference | Generation using reference passages. | synthetic.jsonl |
Retrieval experiments can be run using the BEIR codebase as described in the retrieval README. The corpus will need to be ingested to run experiments.
Generation experiments can be run using any desired models (e.g. available on HuggingFace) and settings as described in the generation README.
We provide analytics files in InspectorRAGet format, which can be used to inspect the evaluation results and perform further analysis. Load any of the analytics files in InspectorRAGet by clicking "Visualize" and follow the instructions shown on the screen.
- We'd like to thank our internal annotators for their considerable effort in creating these conversations: Mohamed Nasr, Joekie Gurski, Tamara Henderson, Hee Dong Lee, Roxana Passaro, Chie Ugumori, Marina Variano, Eva-Maria Wolfe
- We'd like to thank Krishnateja Killamsetty for question classification
- We'd like to thank Lihong He for corpus ingestion
- We'd like to thank Aditya Gaur for deployment help
Sara Rosenthal, Yannis Katsis, Kshitij Fadnis, Chulaka Gunasekara, Young-Suk Lee, Lucian Popa, Vraj Shah, Huaiyu Zhu, Danish Contractor, Marina Danilevsky
- Sara Rosenthal sjrosenthal@us.ibm.com
- Yannis Katsis yannis.katsis@ibm.com
- Marina Danilevsky mdanile@us.ibm.com