Top Research Papers on Retrieval Augmented Generation
Dive into the top research papers on Retrieval Augmented Generation, a cutting-edge field at the intersection of information retrieval and natural language generation. This compilation highlights pioneering studies and provides valuable insights for researchers, professionals, and enthusiasts. Enhance your knowledge and stay current with the latest advancements in Retrieval Augmented Generation.
Looking for research-backed answers?Try AI Search
Improving the Domain Adaptation of Retrieval Augmented Generation (RAG) Models for Open Domain Question Answering
214 Citations 2023Shamane Siriwardhana, Rivindu Weerasekera, Elliott Wen + 3 more
Transactions of the Association for Computational Linguistics
This paper proposes RAG-end2end, an extension to RAG that can adapt to a domain- specific knowledge base by updating all components of the external knowledge base during training and introduces an auxiliary training signal to inject more domain-specific knowledge.
Evaluating Retrieval Quality in Retrieval-Augmented Generation
100 Citations 2024Alireza Salemi, Hamed Zamani
journal unavailable
Evaluating retrieval-augmented generation with a novel evaluation approach, eRAG, where each document in the retrieval list is individually utilized by the large language model within the RAG system, and the downstream performance for each document serves as its relevance label.
Active Retrieval Augmented Generation
271 Citations 2023Zhengbao Jiang, Frank F. Xu, Luyu Gao + 6 more
journal unavailable
Zhengbao Jiang, Frank Xu, Luyu Gao, Zhiqing Sun, Qian Liu, Jane Dwivedi-Yu, Yiming Yang, Jamie Callan, Graham Neubig. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. 2023.
A Survey on RAG Meeting LLMs: Towards Retrieval-Augmented Large Language Models
400 Citations 2024Wenqi Fan, Yujuan Ding, Liangbo Ning + 5 more
journal unavailable
This survey comprehensively review existing research studies in RA-LLMs, covering three primary technical perspectives: architectures, training strategies, and applications, and systematically review mainstream relevant work by their architectures, training strategies, and application areas.
RAGAs: Automated Evaluation of Retrieval Augmented Generation
162 Citations 2024Shahul Es, Jithin James, Luis Espinosa-Anke + 1 more
journal unavailable
Shahul Es, Jithin James, Luis Espinosa Anke, Steven Schockaert. Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics: System Demonstrations. 2024.
Touché-25-Advertisement-in-Retrieval-Augmented-Generation
1283 Citations 2024Payal Bajaj, Daniel Campos, Nick Craswell + 12 more
arXiv (Cornell University)
The size of the dataset and the fact that the questions are derived from real user search queries distinguishes MS MARCO from other well-known publicly available datasets for machine reading comprehension and question-answering.
Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language Models through Question Complexity
108 Citations 2024Soyeong Jeong, Jinheon Baek, Sukmin Cho + 2 more
journal unavailable
Soyeong Jeong, Jinheon Baek, Sukmin Cho, Sung Ju Hwang, Jong Park. Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). 2024.
Benchmarking Large Language Models in Retrieval-Augmented Generation
272 Citations 2024Jiawei Chen, Hongyu Lin, Xianpei Han + 1 more
Proceedings of the AAAI Conference on Artificial Intelligence
Retrieval-Augmented Generation (RAG) is a promising approach for mitigating the hallucination of large language models (LLMs). However, existing research lacks rigorous evaluation of the impact of retrieval-augmented generation on different large language models, which make it challenging to identify the potential bottlenecks in the capabilities of RAG for different LLMs. In this paper, we systematically investigate the impact of Retrieval-Augmented Generation on large language models. We analyze the performance of different large language models in 4 fundamental abilities required for RAG, in...
Generation-Augmented Retrieval for Open-Domain Question Answering
135 Citations 2021Yuning Mao, Pengcheng He, Xiaodong Liu + 4 more
journal unavailable
It is shown that generating diverse contexts for a query is beneficial as fusing their results consistently yields better retrieval accuracy, and as sparse and dense representations are often complementary, GAR can be easily combined with DPR to achieve even better performance.
Retrieval-Augmented Generation for Large Language Models: A Survey
584 Citations 2023Yunfan Gao, Yun Xiong, Xinyu Gao + 7 more
arXiv (Cornell University)
Large Language Models (LLMs) showcase impressive capabilities but encounter challenges like hallucination, outdated knowledge, and non-transparent, untraceable reasoning processes. Retrieval-Augmented Generation (RAG) has emerged as a promising solution by incorporating knowledge from external databases. This enhances the accuracy and credibility of the generation, particularly for knowledge-intensive tasks, and allows for continuous knowledge updates and integration of domain-specific information. RAG synergistically merges LLMs' intrinsic knowledge with the vast, dynamic repositories of exte...
Enhancing Retrieval-Augmented Large Language Models with Iterative Retrieval-Generation Synergy
107 Citations 2023Zhihong Shao, Yeyun Gong, Yelong Shen + 3 more
journal unavailable
Retrieval-augmented generation has raise extensive attention as it is promising to address the limitations of large language models including outdated knowledge and hallucinations. However, retrievers struggle to capture relevance, especially for queries with complex information needs. Recent work has proposed to improve relevance modeling by having large language models actively involved in retrieval, i.e., to guide retrieval with generation. In this paper, we show that strong performance can be achieved by a method we call Iter-RetGen, which synergizes retrieval and generation in an iterativ...
The Power of Noise: Redefining Retrieval for RAG Systems
117 Citations 2024Florin Cuconasu, Giovanni Trappolini, Federico Siciliano + 5 more
journal unavailable
It is argued here that the retrieval component of RAG systems, be it dense or sparse, deserves increased attention from the research community, and accordingly, the first comprehensive and systematic examination of the retrieval strategy of RAG systems is conducted.
In-Context Retrieval-Augmented Language Models
283 Citations 2023Ori Ram, Yoav Levine, Itay Dalmedigos + 4 more
Transactions of the Association for Computational Linguistics
It is shown that In-Context RALM that builds on off-the-shelf general purpose retrievers provides surprisingly large LM gains across model sizes and diverse corpora, and that the document retrieval and ranking mechanism can be specialized to the RALm setting to further boost performance.
Retrieval Augmentation Reduces Hallucination in Conversation
412 Citations 2021Kurt Shuster, Spencer Poff, Moya Chen + 2 more
journal unavailable
The use of neural-retrieval-in-the-loop architectures - recently shown to be effective in open-domain QA - is explored for knowledge-grounded dialogue, a task that is arguably more challenging as it requires querying based on complex multi-turn dialogue context and generating conversationally coherent responses.
Development of a liver disease–specific large language model chat interface using retrieval-augmented generation
102 Citations 2024Jin Ge, Steve Sun, Joseph F. Owens + 5 more
Hepatology
While LiVersa demonstrated higher accuracy in answering questions related to hepatology - there were some deficiencies due to limitations set by the number of documents used for RAG, the LiVersa prototype is a proof of concept for utilizing RAG to customize LLMs for clinical use cases.
Optimization of hepatological clinical guidelines interpretation by large language models: a retrieval augmented generation-based framework
154 Citations 2024Simone Kresevic, Mauro Giuffrè, Miloš Ajčević + 3 more
npj Digital Medicine
The study highlights that structured guideline reformatting and advanced prompt engineering can enhance the efficacy of LLM integrations to CDSSs for guideline delivery.
REALM: Retrieval-Augmented Language Model Pre-Training
512 Citations 2020Kelvin Guu, Kenton Lee, Zora Tung + 2 more
arXiv (Cornell University)
The effectiveness of Retrieval-Augmented Language Model pre-training (REALM) is demonstrated by fine-tuning on the challenging task of Open-domain Question Answering (Open-QA) and is found to outperform all previous methods by a significant margin, while also providing qualitative benefits such as interpretability and modularity.
REPLUG: Retrieval-Augmented Black-Box Language Models
109 Citations 2024Weijia Shi, Sewon Min, Michihiro Yasunaga + 5 more
journal unavailable
Weijia Shi, Sewon Min, Michihiro Yasunaga, Minjoon Seo, Richard James, Mike Lewis, Luke Zettlemoyer, Wen-tau Yih. Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers). 2024.
Almanac — Retrieval-Augmented Language Models for Clinical Medicine
287 Citations 2024Cyril Zakka, Rohan Shad, Akash Chaurasia + 19 more
NEJM AI
Almanac, an LLM framework augmented with retrieval capabilities from curated medical resources for medical guideline and treatment recommendations, showed a significant improvement in performance compared with the standard LLMs across axes of factuality, completeness, user preference, and adversarial safety.
Query Rewriting in Retrieval-Augmented Large Language Models
158 Citations 2023Xinbei Ma, Yeyun Gong, Pengcheng He + 2 more
journal unavailable
Large Language Models (LLMs) play powerful, black-box readers in the retrieve-then-read pipeline, making remarkable progress in knowledge-intensive tasks. This work introduces a new framework, Rewrite-Retrieve-Read instead of the previous retrieve-then-read for the retrieval-augmented LLMs from the perspective of the query rewriting. Unlike prior studies focusing on adapting either the retriever or the reader, our approach pays attention to the adaptation of the search query itself, for there is inevitably a gap between the input text and the needed knowledge in retrieval. We first prompt an L...
Internet-Augmented Dialogue Generation
166 Citations 2022Mojtaba Komeili, Kurt Shuster, Jason Weston
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
An approach that learns to generate an internet search query based on the context, and then conditions on the search results to finally generate a response, a method that can employ up-to-the-minute relevant information.
Generating Clarifying Questions for Information Retrieval
205 Citations 2020Hamed Zamani, Susan Dumais, Nick Craswell + 2 more
journal unavailable
A taxonomy of clarification for open-domain search queries is identified by analyzing large-scale query reformulation data sampled from Bing search logs, and supervised and reinforcement learning models for generating clarifying questions learned from weak supervision data are proposed.
Atlas: Few-shot Learning with Retrieval Augmented Language Models
195 Citations 2022Gautier Izacard, Patrick Lewis, María Lomelí + 7 more
arXiv (Cornell University)
Large language models have shown impressive few-shot results on a wide range of tasks. However, when knowledge is key for such results, as is the case for tasks such as question answering and fact checking, massive parameter counts to store knowledge seem to be needed. Retrieval augmented models are known to excel at knowledge intensive tasks without the need for as many parameters, but it is unclear whether they work in few-shot settings. In this work we present Atlas, a carefully designed and pre-trained retrieval augmented language model able to learn knowledge intensive tasks with very few...
Tree-Augmented Cross-Modal Encoding for Complex-Query Video Retrieval
128 Citations 2020Xun Yang, Jianfeng Dong, Yixin Cao + 3 more
journal unavailable
A Tree-augmented Cross-modal Encoding method by jointly learning the linguistic structure of queries and the temporal representation of videos to facilitate video retrieval with complex queries, thereby achieving a better video retrieval performance.
Enhancing Financial Sentiment Analysis via Retrieval Augmented Large Language Models
143 Citations 2023Boyu Zhang, Hongyang Yang, Tianyu Zhou + 2 more
journal unavailable
This work introduces a retrieval-augmented LLMs framework for financial sentiment analysis, which includes an instruction-tuned LL Ms module, which ensures LLMs behave as predictors of sentiment labels, and a retrieval -augmentation module which retrieves additional context from reliable external sources.
A Survey on RAG with LLMs
113 Citations 2024Muhammad Arslan, Hussam Ghanem, Saba Munawar + 1 more
Procedia Computer Science
In the fast-paced realm of digital transformation, businesses are increasingly pressured to innovate and boost efficiency to remain competitive and foster growth. Large Language Models (LLMs) have emerged as game-changers across industries, revolutionizing various sectors by harnessing extensive text data to analyze and generate human-like text. Despite their impressive capabilities, LLMs often encounter challenges when dealing with domain-specific queries, potentially leading to inaccuracies in their outputs. In response, Retrieval-Augmented Generation (RAG) has emerged as a viable solution. ...
Generative Adversarial Networks in Medical Image augmentation: A review
367 Citations 2022Yizhou Chen, Xu-Hua Yang, Zihan Wei + 7 more
Computers in Biology and Medicine
A comprehensive and systematic review and analysis of medical image augmentation work are carried out, and the existing limitations of this type of model are discussed and possible research directions are suggested.
How Generative AI Can Augment Human Creativity
114 Citations 2024Tojin Eapen, Daniel J. Finkenstadt, Josh Folk + 1 more
SSRN Electronic Journal
The advent of generative AI has sparked apprehension regarding its capacity to replace human workers in many jobs. However, the most significant opportunity for this technology lies in its potential to assist businesses and governments in addressing the challenges associated with democratizing innovation by augmenting human creativity. These challenges include the influx of ideas, the inability of domain experts to embrace novel concepts, the difficulty in refining vague ideas into actionable plans, and the struggle to integrate diverse customer requirements into viable solutions. Generative A...
Open-book Video Captioning with Retrieve-Copy-Generate Network
102 Citations 2021Ziqi Zhang, Zhongang Qi, Chunfeng Yuan + 4 more
journal unavailable
A novel Retrieve-Copy-Generate network is proposed, where a pluggable video-to-text retriever is constructed to retrieve sentences as hints from the training corpus effectively, and a copy-mechanism generator is introduced to extract expressions from multi-retrieved sentences dynamically.
JamLab: augmenting sensornet testbeds with realistic and controlled interference generation
130 Citations 2022Carlo Alberto Boano, Thiemo Voigt, Claro Noda + 2 more
Publications (Konstfack University of Arts, Crafts, and Design)
This paper uses off-the-shelf sensor motes to record and playback interference patterns as well as to generate customizable and repeat-able interference in real-time, and proposes and develops JamLab: a low-cost infrastructure to augment existing sensornet testbeds with accurate interference generation while limiting the overhead to a simple upload of the appropriate software.
Generative Adversarial Networks-Based Data Augmentation for Brain–Computer Interface
207 Citations 2020Fatemeh Fahimi, Strahinja Došen, Kai Keng Ang + 2 more
IEEE Transactions on Neural Networks and Learning Systems
This study shows that using GANs for EEG augmentation can significantly improve BCI performance, especially in real-life applications, whereby users' attention may be diverted.
Realistic in silico generation and augmentation of single-cell RNA-seq data using generative adversarial networks
241 Citations 2020Mohamed Marouf, Pierre Machart, V. Bansal + 4 more
Nature Communications
A method to generate realistic scRNA-seq data using GANs that learn gene expression dependencies from complex samples, and shows that augmenting spare cell populations improves downstream analyses.
Generative adversarial networks (GANs) for image augmentation in agriculture: A systematic review
287 Citations 2022Yuzhen Lu, Dong Chen, Ebenezer O. Olaniyi + 1 more
Computers and Electronics in Agriculture
In agricultural image analysis, optimal model performance is keenly pursued for better fulfilling visual recognition tasks (e.g., image classification, segmentation, object detection and localization), in the presence of challenges with biological variability and unstructured environments. Large-scale, balanced and ground-truthed image datasets are tremendously beneficial but most often difficult to obtain to fuel the development of highly performant models. As artificial intelligence through deep learning is impacting analysis and modeling of agricultural images, image augmentation plays a cr...
Skin Electronics: Next‐Generation Device Platform for Virtual and Augmented Reality
186 Citations 2021Jae Joon Kim, Yan Wang, Haoyang Wang + 3 more
Advanced Functional Materials
The current skin electronics are summarized as one of the most promising device solutions for future VR/AR devices, especially focusing on the recent materials and structures.
DAGA: Data Augmentation with a Generation Approach for Low-resource Tagging Tasks
132 Citations 2020Bosheng Ding, Linlin Liu, Lidong Bing + 5 more
journal unavailable
To generate high quality synthetic data for low-resource tagging tasks, a novel augmentation method with language models trained on the linearized labeled sentences is proposed that can consistently outperform the baselines.
KG-BART: Knowledge Graph-Augmented BART for Generative Commonsense Reasoning
155 Citations 2021Liu Ye, Yao Wan, Lifang He + 2 more
Proceedings of the AAAI Conference on Artificial Intelligence
A novel knowledge graph augmented pre-trained language generation model KG-BART is proposed, which encompasses the complex relations of concepts through the knowledge graph and produces more logical and natural sentences as output and can leverage the graph attention to aggregate the rich concept semantics that enhances the model generalization on unseen concept sets.
Conditional Augmentation for Aspect Term Extraction via Masked Sequence-to-Sequence Generation
100 Citations 2020Kun Li, Chengbo Chen, Xiaojun Quan + 2 more
journal unavailable
This paper proposes a masked sequence-to-sequence method for conditional augmentation of aspect term extraction that is controllable and allows to generate more diversified sentences, and effectively boosts the performances of several current models for aspect terms extraction.
Data augmentation for enhancing EEG-based emotion recognition with deep generative models
149 Citations 2020Yun Luo, Lizhen Zhu, Ziyu Wan + 1 more
Journal of Neural Engineering
The experimental results demonstrate that the proposed data augmentation methods based on generative models outperform the existingData augmentation approaches such as conditional VAE, Gaussian noise, and rotational data augmented.
Defect modified zinc oxide with augmenting sonodynamic reactive oxygen species generation
185 Citations 2020Yang Liu, Ying Wang, Wenyao Zhen + 8 more
Biomaterials
The mechanism of how the oxygen deficiency enhanced the sonodynamic efficacy of zinc oxide is revealed, providing a new application of defect engineering in the field of cancer therapy.
ARShadowGAN: Shadow Generative Adversarial Network for Augmented Reality in Single Light Scenes
117 Citations 2020Daquan Liu, Chengjiang Long, Hongpan Zhang + 3 more
journal unavailable
The extensive experimental results show that the proposed ARShadowGAN is capable of directly generating plausible virtual object shadows in single light scenes.