SOTAVerified

Hallucination

Papers

Showing 801850 of 1816 papers

TitleStatusHype
FRAG: A Flexible Modular Framework for Retrieval-Augmented Generation based on Knowledge Graphs0
ArxEval: Evaluating Retrieval and Generation in Language Models for Scientific Literature0
A Survey on Responsible LLMs: Inherent Risk, Malicious Use, and Mitigation Strategy0
HALoGEN: Fantastic LLM Hallucinations and Where to Find Them0
GPT as a Monte Carlo Language Tree: A Probabilistic Perspective0
MedCT: A Clinical Terminology Graph for Generative AI Applications in Healthcare0
Fine-tuning Large Language Models for Improving Factuality in Legal Question AnsweringCode0
Hermit Kingdom Through the Lens of Multiple Perspectives: A Case Study of LLM Hallucination on North Korea0
Seeing with Partial Certainty: Conformal Prediction for Robotic Scene Recognition in Built Environments0
Feedback-Driven Vision-Language Alignment with Minimal Human Supervision0
RAG-Check: Evaluating Multimodal Retrieval Augmented Generation Performance0
EAGLE: Enhanced Visual Grounding Minimizes Hallucinations in Instructional Multimodal Models0
Socratic Questioning: Learn to Self-guide Multimodal Reasoning in the WildCode0
Foundations of GenIR0
FlippedRAG: Black-Box Opinion Manipulation Adversarial Attacks to Retrieval-Augmented Generation Models0
CHAIR -- Classifier of Hallucination as ImproverCode0
LLMs & Legal Aid: Understanding Legal Needs Exhibited Through User Queries0
CarbonChat: Large Language Model-Based Corporate Carbon Emission Analysis and Climate Knowledge Q&A System0
Large Language Model-Enhanced Symbolic Reasoning for Knowledge Base Completion0
Think More, Hallucinate Less: Mitigating Hallucinations via Dual Process of Fast and Slow Thinking0
Enhancing Uncertainty Modeling with Semantic Graph for Hallucination Detection0
RRHF-V: Ranking Responses to Mitigate Hallucinations in Multimodal Large Language Models with Human FeedbackCode0
IllusionBench: A Large-scale and Comprehensive Benchmark for Visual Illusion Understanding in Vision-Language Models0
Stop Learning it all to Mitigate Visual Hallucination, Focus on the Hallucination Target.0
POPEN: Preference-Based Optimization and Ensemble for LVLM-Based Reasoning Segmentation0
VL-RewardBench: A Challenging Benchmark for Vision-Language Generative Reward Models0
A review of faithfulness metrics for hallucination assessment in Large Language Models0
Distilling Desired Comments for Enhanced Code Review with Large Language Models0
HALLUCINOGEN: A Benchmark for Evaluating Object Hallucination in Large Visual-Language ModelsCode0
Is Your Text-to-Image Model Robust to Caption Noise?0
An End-to-End Depth-Based Pipeline for Selfie Image Rectification0
MedHallBench: A New Benchmark for Assessing Hallucination in Medical Large Language Models0
Improving Factuality with Explicit Working Memory0
From Hallucinations to Facts: Enhancing Language Models with Curated Knowledge Graphs0
Multimodal Preference Data Synthetic Alignment with Reward ModelCode0
CiteBART: Learning to Generate Citations for Local Citation RecommendationCode0
AlzheimerRAG: Multimodal Retrieval Augmented Generation for PubMed articles0
Logical Consistency of Large Language Models in Fact-checking0
Toward Robust Hyper-Detailed Image Captioning: A Multiagent Approach and Dual Evaluation Metrics for Factuality and Coverage0
Token Preference Optimization with Self-Calibrated Visual-Anchored Rewards for Hallucination Mitigation0
Dehallucinating Parallel Context Extension for Retrieval-Augmented Generation0
Think&Cite: Improving Attributed Text Generation with Self-Guided Tree Search and Progress Reward Modeling0
Query pipeline optimization for cancer patient question answering systems0
A Comparative Study of DSPy Teleprompter Algorithms for Aligning Large Language Models Evaluation Metrics to Human Evaluation0
Cracking the Code of Hallucination in LVLMs with Vision-aware Head Divergence0
Are LLMs Good Literature Review Writers? Evaluating the Literature Review Writing Ability of Large Language Models0
When to Speak, When to Abstain: Contrastive Decoding with Abstention0
A MapReduce Approach to Effectively Utilize Long Context Information in Retrieval Augmented Language Models0
What External Knowledge is Preferred by LLMs? Characterizing and Exploring Chain of Evidence in Imperfect Context0
ReXTrust: A Model for Fine-Grained Hallucination Detection in AI-Generated Radiology Reports0
Show:102550
← PrevPage 17 of 37Next →

No leaderboard results yet.