SOTAVerified

RAG

Retrieval-Augmented Generation (RAG) is a task that combines the strengths of both retrieval-based models and generation-based models. In this approach, a retrieval system selects relevant documents or passages from a large corpus, and a generation model, typically a neural language model, uses the retrieved information to generate a response. This method enhances the accuracy and coherence of generated text, especially in tasks requiring detailed knowledge or long context handling.

RAG is particularly useful in open-domain question answering, knowledge-grounded dialogue, and summarization tasks. The retrieval step helps the model to access and incorporate external information, making it less reliant on memorized knowledge and better suited for generating responses based on the latest or domain-specific information.

The performance of RAG systems is usually measured using metrics such as precision, recall, F1 score, BLEU score, and exact match. Some popular datasets for evaluating RAG models include Natural Questions, MS MARCO, TriviaQA, and SQuAD.

Papers

Showing 10011025 of 2111 papers

TitleStatusHype
RAG-Star: Enhancing Deliberative Reasoning with Retrieval Augmented Verification and Refinement0
RemoteRAG: A Privacy-Preserving LLM Cloud RAG Service0
Adaptations of AI models for querying the LandMatrix database in natural languageCode0
OmniEval: An Omnidirectional and Automatic RAG Evaluation Benchmark in Financial DomainCode2
What External Knowledge is Preferred by LLMs? Characterizing and Exploring Chain of Evidence in Imperfect Context0
Advanced ingestion process powered by LLM parsing for RAG system0
RAG Playground: A Framework for Systematic Evaluation of Retrieval Strategies and Prompt Engineering in RAG SystemsCode1
LogBabylon: A Unified Framework for Cross-Log File Integration and Analysis0
Unanswerability Evaluation for Retrieval Augmented Generation0
BioRAGent: A Retrieval-Augmented Generation System for Showcasing Generative Query Expansion and Domain-Specific Search for Scientific Q&ACode0
Let your LLM generate a few tokens and you will reduce the need for retrieval0
Attention with Dependency Parsing Augmentation for Fine-Grained Attribution0
RetroLLM: Empowering Large Language Models to Retrieve Fine-grained Evidence within GenerationCode2
Agentic AI-Driven Technical Troubleshooting for Enterprise Systems: A Novel Weighted Retrieval-Augmented Generation Paradigm0
RAC3: Retrieval-Augmented Corner Case Comprehension for Autonomous Driving with Vision-Language Models0
One-Shot Multilingual Font Generation Via ViT0
Streamlining Systematic Reviews: A Novel Application of Large Language Models0
Accelerating Retrieval-Augmented Generation0
Inference Scaling for Bridging Retrieval and Augmented Generation0
SusGen-GPT: A Data-Centric LLM for Financial NLP and Sustainability Report GenerationCode1
VisDoM: Multi-Document QA with Visually Rich Elements Using Multimodal Retrieval-Augmented Generation0
Evidence Contextualization and Counterfactual Attribution for Conversational QA over Heterogeneous Data with RAG Systems0
RAGServe: Fast Quality-Aware RAG Systems with Configuration Adaptation0
CaLoRAify: Calorie Estimation with Visual-Text Pairing and LoRA-Driven Visual Language ModelsCode1
VLR-Bench: Multilingual Benchmark Dataset for Vision-Language Retrieval Augmented Generation0
Show:102550
← PrevPage 41 of 85Next →

No leaderboard results yet.