SOTAVerified

Hallucination

Papers

Showing 51100 of 1816 papers

TitleStatusHype
Mitigating Manipulation and Enhancing Persuasion: A Reflective Multi-Agent Approach for Legal Argument Generation0
Machine Mirages: Defining the Undefined0
FlySearch: Exploring how vision-language models exploreCode1
Tomographic Foundation Model -- FORCE: Flow-Oriented Reconstruction Conditioning Engine0
TRUST -- Transformer-Driven U-Net for Sparse Target Recovery0
Generative AI and Organizational Structure in the Knowledge Economy0
Measuring Faithfulness and Abstention: An Automated Pipeline for Evaluating LLM-Generated 3-ply Case-Based Legal Arguments0
An AI-powered Knowledge Hub for Potato Functional Genomics0
BIMA: Bijective Maximum Likelihood Learning Approach to Hallucination Prediction and Mitigation in Large Vision-Language Models0
Improving Reliability and Explainability of Medical Question Answering through Atomic Fact Checking in Retrieval-Augmented LLMs0
MIRAGE: Assessing Hallucination in Multimodal Reasoning Chains of MLLM0
FinMME: Benchmark Dataset for Financial Multi-Modal Reasoning EvaluationCode2
LLM Inference Enhanced by External Knowledge: A SurveyCode0
The Hallucination Dilemma: Factuality-Aware Reinforcement Learning for Large Reasoning ModelsCode1
Preemptive Hallucination Reduction: An Input-Level Approach for Multimodal Language Model0
Reinforcement Learning for Better Verbalized Confidence in Long-Form Generation0
Map&Make: Schema Guided Text to Table Generation0
MMBoundary: Advancing MLLM Knowledge Boundary Awareness through Reasoning Step Confidence CalibrationCode0
Are Reasoning Models More Prone to Hallucination?0
Active Layer-Contrastive Decoding Reduces Hallucination in Large Language Model Generation0
Qwen Look Again: Guiding Vision-Language Reasoning Models to Re-attention Visual InformationCode0
Data-efficient Meta-models for Evaluation of Context-based Questions and Answers in LLMs0
Evaluation Hallucination in Multi-Round Incomplete Information Lateral-Driven Reasoning Tasks0
SkewRoute: Training-Free LLM Routing for Knowledge Graph Retrieval-Augmented Generation via Score Skewness of Retrieved Context0
CogniBench: A Legal-inspired Framework and Dataset for Assessing Cognitive Faithfulness of Large Language ModelsCode1
A Lightweight Multi-Expert Generative Language Model System for Engineering Information and Knowledge Extraction0
Mitigating Hallucination in Large Vision-Language Models via Adaptive Attention Calibration0
R3-RAG: Learning Step-by-Step Reasoning and Retrieval for LLMs via Reinforcement LearningCode1
Retrieval Visual Contrastive Decoding to Mitigate Object Hallucinations in Large Vision-Language ModelsCode0
Causal-LLaVA: Causal Disentanglement for Mitigating Hallucination in Multimodal Large Language ModelsCode0
Attention! You Vision Language Model Could Be Maliciously Manipulated0
Error Typing for Smarter Rewards: Improving Process Reward Models with Error-Aware Hierarchical SupervisionCode0
Enhancing Visual Reliance in Text Generation: A Bayesian Perspective on Mitigating Hallucination in Large Vision-Language Models0
Omni-R1: Reinforcement Learning for Omnimodal Reasoning via Two-System CollaborationCode2
Grounding Language with Vision: A Conditional Mutual Information Calibrated Decoding Strategy for Reducing Hallucinations in LVLMs0
Uncertainty-Aware Attention Heads: Efficient Unsupervised Uncertainty Quantification for LLMs0
LLLMs: A Data-Driven Survey of Evolving Research on Limitations of Large Language Models0
CCHall: A Novel Benchmark for Joint Cross-Lingual and Cross-Modal Hallucinations Detection in Large Language ModelsCode0
GUARDIAN: Safeguarding LLM Multi-Agent Collaborations with Temporal Graph Modeling0
Removal of Hallucination on Hallucination: Debate-Augmented RAGCode1
MedScore: Factuality Evaluation of Free-Form Medical AnswersCode0
More Thinking, Less Seeing? Assessing Amplified Hallucination in Multimodal Reasoning Models0
Teaching with Lies: Curriculum DPO on Synthetic Negatives for Hallucination Detection0
keepitsimple at SemEval-2025 Task 3: LLM-Uncertainty based Approach for Multilingual Hallucination Span DetectionCode0
LLM-Powered Agents for Navigating Venice's Historical Cadastre0
Chain-of-Thought Poisoning Attacks against R1-based Retrieval-Augmented Generation Systems0
Shadows in the Attention: Contextual Perturbation and Representation Drift in the Dynamics of Hallucination in LLMs0
Mitigating Hallucinations in Vision-Language Models through Image-Guided Head SuppressionCode1
Steering LVLMs via Sparse Autoencoder for Hallucination Mitigation0
Seeing Far and Clearly: Mitigating Hallucinations in MLLMs with Attention Causal Decoding0
Show:102550
← PrevPage 2 of 37Next →

No leaderboard results yet.