SOTAVerified

RAG

Retrieval-Augmented Generation (RAG) is a task that combines the strengths of both retrieval-based models and generation-based models. In this approach, a retrieval system selects relevant documents or passages from a large corpus, and a generation model, typically a neural language model, uses the retrieved information to generate a response. This method enhances the accuracy and coherence of generated text, especially in tasks requiring detailed knowledge or long context handling.

RAG is particularly useful in open-domain question answering, knowledge-grounded dialogue, and summarization tasks. The retrieval step helps the model to access and incorporate external information, making it less reliant on memorized knowledge and better suited for generating responses based on the latest or domain-specific information.

The performance of RAG systems is usually measured using metrics such as precision, recall, F1 score, BLEU score, and exact match. Some popular datasets for evaluating RAG models include Natural Questions, MS MARCO, TriviaQA, and SQuAD.

Papers

Showing 526550 of 2111 papers

TitleStatusHype
GINGER: Grounded Information Nugget-Based Generation of ResponsesCode0
Experience Retrieval-Augmentation with Electronic Health Records Enables Accurate Discharge QACode0
MedAgent-Pro: Towards Evidence-based Multi-modal Medical Diagnosis via Reasoning Agentic WorkflowCode2
RustEvo^2: An Evolving Benchmark for API Evolution in LLM-based Rust Code GenerationCode0
An LLM-Powered Clinical Calculator Chatbot Backed by Verifiable Clinical Calculators and their Metadata0
Autonomous Radiotherapy Treatment Planning Using DOLA: A Privacy-Preserving, LLM-Based Optimization Agent0
Financial Analysis: Intelligent Financial Data Analysis System Based on LLM-RAG0
FutureGen: LLM-RAG Approach to Generate the Future Work of Scientific ArticleCode0
Parameters vs. Context: Fine-Grained Control of Knowledge Reliance in Language ModelsCode1
Tuning LLMs by RAG Principles: Towards LLM-native MemoryCode1
Towards Lighter and Robust Evaluation for Retrieval Augmented GenerationCode0
Typed-RAG: Type-aware Multi-Aspect Decomposition for Non-Factoid Question AnsweringCode0
Graph-Based Re-ranking: Emerging Techniques, Limitations, and Opportunities0
Enhancing Pancreatic Cancer Staging with Large Language Models: The Role of Retrieval-Augmented Generation0
Bias Evaluation and Mitigation in Retrieval-Augmented Medical Question-Answering Systems0
Optimizing Retrieval Strategies for Financial Question Answering Documents in Retrieval-Augmented Generation SystemsCode1
When Pigs Get Sick: Multi-Agent AI for Swine Disease Detection0
Does Context Matter? ContextualJudgeBench for Evaluating LLM-based Judges in Contextual SettingsCode0
RAG-based User Profiling for Precision Planning in Mixed-precision Over-the-Air Federated Learning0
Enhancing LLM Generation with Knowledge Hypergraph for Evidence-Based Medicine0
RAGO: Systematic Performance Optimization for Retrieval-Augmented Generation ServingCode1
JuDGE: Benchmarking Judgment Document Generation for Chinese Legal SystemCode1
MoK-RAG: Mixture of Knowledge Paths Enhanced Retrieval-Augmented Generation for Embodied AI Environments0
From "Hallucination" to "Suture": Insights from Language Philosophy to Enhance Large Language Models0
Good/Evil Reputation Judgment of Celebrities by LLMs via Retrieval Augmented Generation0
Show:102550
← PrevPage 22 of 85Next →

No leaderboard results yet.