Top Research Papers on RAG
Explore our curated list of top research papers on RAG. This comprehensive collection delves into various aspects and innovations surrounding RAG. Whether you're a student, researcher, or enthusiast, these studies will provide valuable insights and deepen your understanding of RAG. Dive into the world of research and discover the breakthroughs shaping the future of RAG.
Looking for research-backed answers?Try AI Search
Improving the Domain Adaptation of Retrieval Augmented Generation (RAG) Models for Open Domain Question Answering
214 Citations 2023Shamane Siriwardhana, Rivindu Weerasekera, Elliott Wen + 3 more
Transactions of the Association for Computational Linguistics
This paper proposes RAG-end2end, an extension to RAG that can adapt to a domain- specific knowledge base by updating all components of the external knowledge base during training and introduces an auxiliary training signal to inject more domain-specific knowledge.
Evaluating Retrieval Quality in Retrieval-Augmented Generation
100 Citations 2024Alireza Salemi, Hamed Zamani
journal unavailable
Evaluating retrieval-augmented generation with a novel evaluation approach, eRAG, where each document in the retrieval list is individually utilized by the large language model within the RAG system, and the downstream performance for each document serves as its relevance label.
Active Retrieval Augmented Generation
271 Citations 2023Zhengbao Jiang, Frank F. Xu, Luyu Gao + 6 more
journal unavailable
Zhengbao Jiang, Frank Xu, Luyu Gao, Zhiqing Sun, Qian Liu, Jane Dwivedi-Yu, Yiming Yang, Jamie Callan, Graham Neubig. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. 2023.
A Survey on RAG Meeting LLMs: Towards Retrieval-Augmented Large Language Models
400 Citations 2024Wenqi Fan, Yujuan Ding, Liangbo Ning + 5 more
journal unavailable
This survey comprehensively review existing research studies in RA-LLMs, covering three primary technical perspectives: architectures, training strategies, and applications, and systematically review mainstream relevant work by their architectures, training strategies, and application areas.
RAGAs: Automated Evaluation of Retrieval Augmented Generation
162 Citations 2024Shahul Es, Jithin James, Luis Espinosa-Anke + 1 more
journal unavailable
Shahul Es, Jithin James, Luis Espinosa Anke, Steven Schockaert. Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics: System Demonstrations. 2024.
Touché-25-Advertisement-in-Retrieval-Augmented-Generation
1283 Citations 2024Payal Bajaj, Daniel Campos, Nick Craswell + 12 more
arXiv (Cornell University)
The size of the dataset and the fact that the questions are derived from real user search queries distinguishes MS MARCO from other well-known publicly available datasets for machine reading comprehension and question-answering.
Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language Models through Question Complexity
108 Citations 2024Soyeong Jeong, Jinheon Baek, Sukmin Cho + 2 more
journal unavailable
Soyeong Jeong, Jinheon Baek, Sukmin Cho, Sung Ju Hwang, Jong Park. Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). 2024.
Benchmarking Large Language Models in Retrieval-Augmented Generation
272 Citations 2024Jiawei Chen, Hongyu Lin, Xianpei Han + 1 more
Proceedings of the AAAI Conference on Artificial Intelligence
Retrieval-Augmented Generation (RAG) is a promising approach for mitigating the hallucination of large language models (LLMs). However, existing research lacks rigorous evaluation of the impact of retrieval-augmented generation on different large language models, which make it challenging to identify the potential bottlenecks in the capabilities of RAG for different LLMs. In this paper, we systematically investigate the impact of Retrieval-Augmented Generation on large language models. We analyze the performance of different large language models in 4 fundamental abilities required for RAG, in...
Generation-Augmented Retrieval for Open-Domain Question Answering
135 Citations 2021Yuning Mao, Pengcheng He, Xiaodong Liu + 4 more
journal unavailable
It is shown that generating diverse contexts for a query is beneficial as fusing their results consistently yields better retrieval accuracy, and as sparse and dense representations are often complementary, GAR can be easily combined with DPR to achieve even better performance.
Retrieval-Augmented Generation for Large Language Models: A Survey
584 Citations 2023Yunfan Gao, Yun Xiong, Xinyu Gao + 7 more
arXiv (Cornell University)
Large Language Models (LLMs) showcase impressive capabilities but encounter challenges like hallucination, outdated knowledge, and non-transparent, untraceable reasoning processes. Retrieval-Augmented Generation (RAG) has emerged as a promising solution by incorporating knowledge from external databases. This enhances the accuracy and credibility of the generation, particularly for knowledge-intensive tasks, and allows for continuous knowledge updates and integration of domain-specific information. RAG synergistically merges LLMs' intrinsic knowledge with the vast, dynamic repositories of exte...
Enhancing Retrieval-Augmented Large Language Models with Iterative Retrieval-Generation Synergy
107 Citations 2023Zhihong Shao, Yeyun Gong, Yelong Shen + 3 more
journal unavailable
Retrieval-augmented generation has raise extensive attention as it is promising to address the limitations of large language models including outdated knowledge and hallucinations. However, retrievers struggle to capture relevance, especially for queries with complex information needs. Recent work has proposed to improve relevance modeling by having large language models actively involved in retrieval, i.e., to guide retrieval with generation. In this paper, we show that strong performance can be achieved by a method we call Iter-RetGen, which synergizes retrieval and generation in an iterativ...
The Power of Noise: Redefining Retrieval for RAG Systems
117 Citations 2024Florin Cuconasu, Giovanni Trappolini, Federico Siciliano + 5 more
journal unavailable
It is argued here that the retrieval component of RAG systems, be it dense or sparse, deserves increased attention from the research community, and accordingly, the first comprehensive and systematic examination of the retrieval strategy of RAG systems is conducted.
In-Context Retrieval-Augmented Language Models
283 Citations 2023Ori Ram, Yoav Levine, Itay Dalmedigos + 4 more
Transactions of the Association for Computational Linguistics
It is shown that In-Context RALM that builds on off-the-shelf general purpose retrievers provides surprisingly large LM gains across model sizes and diverse corpora, and that the document retrieval and ranking mechanism can be specialized to the RALm setting to further boost performance.
Retrieval Augmentation Reduces Hallucination in Conversation
412 Citations 2021Kurt Shuster, Spencer Poff, Moya Chen + 2 more
journal unavailable
The use of neural-retrieval-in-the-loop architectures - recently shown to be effective in open-domain QA - is explored for knowledge-grounded dialogue, a task that is arguably more challenging as it requires querying based on complex multi-turn dialogue context and generating conversationally coherent responses.
Development of a liver disease–specific large language model chat interface using retrieval-augmented generation
102 Citations 2024Jin Ge, Steve Sun, Joseph F. Owens + 5 more
Hepatology
While LiVersa demonstrated higher accuracy in answering questions related to hepatology - there were some deficiencies due to limitations set by the number of documents used for RAG, the LiVersa prototype is a proof of concept for utilizing RAG to customize LLMs for clinical use cases.
Optimization of hepatological clinical guidelines interpretation by large language models: a retrieval augmented generation-based framework
154 Citations 2024Simone Kresevic, Mauro Giuffrè, Miloš Ajčević + 3 more
npj Digital Medicine
The study highlights that structured guideline reformatting and advanced prompt engineering can enhance the efficacy of LLM integrations to CDSSs for guideline delivery.
REALM: Retrieval-Augmented Language Model Pre-Training
512 Citations 2020Kelvin Guu, Kenton Lee, Zora Tung + 2 more
arXiv (Cornell University)
The effectiveness of Retrieval-Augmented Language Model pre-training (REALM) is demonstrated by fine-tuning on the challenging task of Open-domain Question Answering (Open-QA) and is found to outperform all previous methods by a significant margin, while also providing qualitative benefits such as interpretability and modularity.
REPLUG: Retrieval-Augmented Black-Box Language Models
109 Citations 2024Weijia Shi, Sewon Min, Michihiro Yasunaga + 5 more
journal unavailable
Weijia Shi, Sewon Min, Michihiro Yasunaga, Minjoon Seo, Richard James, Mike Lewis, Luke Zettlemoyer, Wen-tau Yih. Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). 2024.
Almanac — Retrieval-Augmented Language Models for Clinical Medicine
287 Citations 2024Cyril Zakka, Rohan Shad, Akash Chaurasia + 19 more
NEJM AI
Almanac, an LLM framework augmented with retrieval capabilities from curated medical resources for medical guideline and treatment recommendations, showed a significant improvement in performance compared with the standard LLMs across axes of factuality, completeness, user preference, and adversarial safety.
Query Rewriting in Retrieval-Augmented Large Language Models
158 Citations 2023Xinbei Ma, Yeyun Gong, Pengcheng He + 2 more
journal unavailable
Large Language Models (LLMs) play powerful, black-box readers in the retrieve-then-read pipeline, making remarkable progress in knowledge-intensive tasks. This work introduces a new framework, Rewrite-Retrieve-Read instead of the previous retrieve-then-read for the retrieval-augmented LLMs from the perspective of the query rewriting. Unlike prior studies focusing on adapting either the retriever or the reader, our approach pays attention to the adaptation of the search query itself, for there is inevitably a gap between the input text and the needed knowledge in retrieval. We first prompt an L...