SOTAVerified

RAG

Retrieval-Augmented Generation (RAG) is a task that combines the strengths of both retrieval-based models and generation-based models. In this approach, a retrieval system selects relevant documents or passages from a large corpus, and a generation model, typically a neural language model, uses the retrieved information to generate a response. This method enhances the accuracy and coherence of generated text, especially in tasks requiring detailed knowledge or long context handling.

RAG is particularly useful in open-domain question answering, knowledge-grounded dialogue, and summarization tasks. The retrieval step helps the model to access and incorporate external information, making it less reliant on memorized knowledge and better suited for generating responses based on the latest or domain-specific information.

The performance of RAG systems is usually measured using metrics such as precision, recall, F1 score, BLEU score, and exact match. Some popular datasets for evaluating RAG models include Natural Questions, MS MARCO, TriviaQA, and SQuAD.

Papers

Showing 15761600 of 2111 papers

TitleStatusHype
Bridging the Language Gap: Dynamic Learning Strategies for Improving Multilingual Performance in LLMs0
Bridging Industrial Expertise and XR with LLM-Powered Conversational Agents0
Bridging Relevance and Reasoning: Rationale Distillation in Retrieval-Augmented Generation0
Bridging the Gap: Dynamic Learning Strategies for Improving Multilingual Performance in LLMs0
Bridging the Gap: Enabling Natural Language Queries for NoSQL Databases through Text-to-NoSQL Translation0
Automating Pharmacovigilance Evidence Generation: Using Large Language Models to Produce Context-Aware SQL0
Bridging the Preference Gap between Retrievers and LLMs0
BRIT: Bidirectional Retrieval over Unified Image-Text Graph0
BR-TaxQA-R: A Dataset for Question Answering with References for Brazilian Personal Income Tax Law, including case law0
BSharedRAG: Backbone Shared Retrieval-Augmented Generation for the E-commerce Domain0
Building Trustworthy AI: Transparent AI Systems via Large Language Models, Ontologies, and Logical Reasoning (TranspNet)0
C-3PO: Compact Plug-and-Play Proxy Optimization to Achieve Human-like Retrieval-Augmented Generation0
Cache-Craft: Managing Chunk-Caches for Efficient Retrieval-Augmented Generation0
CAFE: Retrieval Head-based Coarse-to-Fine Information Seeking to Enhance Multi-Document QA Capability0
CAISSON: Concept-Augmented Inference Suite of Self-Organizing Neural Networks0
Calibrated Decision-Making through LLM-Assisted Retrieval0
CALLM: Understanding Cancer Survivors' Emotions and Intervention Opportunities via Mobile Diaries and Context-Aware Language Models0
CancerKG.ORG A Web-scale, Interactive, Verifiable Knowledge Graph-LLM Hybrid for Assisting with Optimal Cancer Treatment and Care0
Can GPT Redefine Medical Understanding? Evaluating GPT on Biomedical Machine Reading Comprehension0
Can Language Models Enable In-Context Database?0
Can LLMs Be Trusted for Evaluating RAG Systems? A Survey of Methods and Datasets0
Can We Further Elicit Reasoning in LLMs? Critic-Guided Planning with Retrieval-Augmentation for Solving Challenging Tasks0
Can we Retrieve Everything All at Once? ARM: An Alignment-Oriented LLM-based Retrieval Method0
Capability-Driven Skill Generation with LLMs: A RAG-Based Approach for Reusing Existing Libraries and Interfaces0
CAPRAG: A Large Language Model Solution for Customer Service and Automatic Reporting using Vector and Graph Retrieval-Augmented Generation0
Show:102550
← PrevPage 64 of 85Next →

No leaderboard results yet.