SOTAVerified

RAG

Retrieval-Augmented Generation (RAG) is a task that combines the strengths of both retrieval-based models and generation-based models. In this approach, a retrieval system selects relevant documents or passages from a large corpus, and a generation model, typically a neural language model, uses the retrieved information to generate a response. This method enhances the accuracy and coherence of generated text, especially in tasks requiring detailed knowledge or long context handling.

RAG is particularly useful in open-domain question answering, knowledge-grounded dialogue, and summarization tasks. The retrieval step helps the model to access and incorporate external information, making it less reliant on memorized knowledge and better suited for generating responses based on the latest or domain-specific information.

The performance of RAG systems is usually measured using metrics such as precision, recall, F1 score, BLEU score, and exact match. Some popular datasets for evaluating RAG models include Natural Questions, MS MARCO, TriviaQA, and SQuAD.

Papers

Showing 176200 of 2111 papers

TitleStatusHype
Retrieval Augmented Generation-based Large Language Models for Bridging Transportation Cybersecurity Legal Knowledge Gaps0
Search Wisely: Mitigating Sub-optimal Agentic Searches By Reducing Uncertainty0
Personalizing Student-Agent Interactions Using Log-Contextualized Retrieval Augmented Generation (RAG)0
DailyQA: A Benchmark to Evaluate Web Retrieval Augmented LLMs Based on Capturing Real-World Changes0
VoxRAG: A Step Toward Transcription-Free RAG Systems in Spoken Question Answering0
BAGELS: Benchmarking the Automated Generation and Extraction of Limitations from Scholarly Text0
Chain-of-Thought Poisoning Attacks against R1-based Retrieval-Augmented Generation Systems0
Augmenting LLM Reasoning with Dynamic Notes Writing for Complex QA0
Continually Self-Improving Language Models for Bariatric Surgery Question--Answering0
MuseRAG: Idea Originality Scoring At ScaleCode0
R1-Searcher++: Incentivizing the Dynamic Knowledge Acquisition of LLMs via Reinforcement LearningCode4
CUB: Benchmarking Context Utilisation Techniques for Language Models0
HASH-RAG: Bridging Deep Hashing with Retriever for Efficient, Fine Retrieval and Augmented Generation0
SimpleDeepSearcher: Deep Information Seeking via Web-Powered Reasoning Trajectory SynthesisCode4
Attributing Response to Context: A Jensen-Shannon Divergence Driven Mechanistic Study of Context Attribution in Retrieval-Augmented Generation0
Tools in the Loop: Quantifying Uncertainty of LLM Question Answering Systems That Use Tools0
Align-GRAG: Reasoning-Guided Dual Alignment for Graph Retrieval-Augmented Generation0
Ask, Retrieve, Summarize: A Modular Pipeline for Scientific Literature SummarizationCode0
Walk&Retrieve: Simple Yet Effective Zero-shot Retrieval-Augmented Generation via Knowledge Graph WalksCode0
Code Graph Model (CGM): A Graph-Integrated Large Language Model for Repository-Level Software Engineering Tasks0
After Retrieval, Before Generation: Enhancing the Trustworthiness of Large Language Models in RAG0
InfoDeepSeek: Benchmarking Agentic Information Seeking for Retrieval-Augmented Generation0
BR-TaxQA-R: A Dataset for Question Answering with References for Brazilian Personal Income Tax Law, including case law0
Ranking Free RAG: Replacing Re-ranking with Selection in RAG for Sensitive Domains0
Reranking with Compressed Document Representation0
Show:102550
← PrevPage 8 of 85Next →

No leaderboard results yet.