SOTAVerified

RAG

Retrieval-Augmented Generation (RAG) is a task that combines the strengths of both retrieval-based models and generation-based models. In this approach, a retrieval system selects relevant documents or passages from a large corpus, and a generation model, typically a neural language model, uses the retrieved information to generate a response. This method enhances the accuracy and coherence of generated text, especially in tasks requiring detailed knowledge or long context handling.

RAG is particularly useful in open-domain question answering, knowledge-grounded dialogue, and summarization tasks. The retrieval step helps the model to access and incorporate external information, making it less reliant on memorized knowledge and better suited for generating responses based on the latest or domain-specific information.

The performance of RAG systems is usually measured using metrics such as precision, recall, F1 score, BLEU score, and exact match. Some popular datasets for evaluating RAG models include Natural Questions, MS MARCO, TriviaQA, and SQuAD.

Papers

Showing 14311440 of 2111 papers

TitleStatusHype
Exploring Information Retrieval Landscapes: An Investigation of a Novel Evaluation Techniques and Comparative Document Splitting MethodsCode0
Retro-li: Small-Scale Retrieval Augmented Generation Supporting Noisy Similarity Searches and Domain Shift GeneralizationCode0
On the Vulnerability of Applying Retrieval-Augmented Generation within Knowledge-Intensive Application Domains0
Enhancing Q&A Text Retrieval with Ranking Models: Benchmarking, fine-tuning and deploying Rerankers for RAG0
OmniQuery: Contextually Augmenting Captured Multimodal Memory to Enable Personal Question Answering0
Unleashing Worms and Extracting Data: Escalating the Outcome of Attacks against RAG-based Inference in Scale and Severity Using JailbreakingCode0
Bio-Eng-LMM AI Assist chatbot: A Comprehensive Tool for Research and EducationCode0
KAG: Boosting LLMs in Professional Domains via Knowledge Augmented GenerationCode9
Knowing When to Ask -- Bridging Large Language Models and Data0
GroUSE: A Benchmark to Evaluate Evaluators in Grounded Question AnsweringCode1
Show:102550
← PrevPage 144 of 212Next →

No leaderboard results yet.