SOTAVerified

Hallucination

Papers

Showing 150 of 1816 papers

TitleStatusHype
Mitigating Object Hallucinations via Sentence-Level Early InterventionCode1
ByDeWay: Boost Your multimodal LLM with DEpth prompting in a Training-Free Way0
UQLM: A Python Package for Uncertainty Quantification in Large Language ModelsCode5
ReLoop: "Seeing Twice and Thinking Backwards" via Closed-loop Training to Mitigate Hallucinations in Multimodal understanding0
DeepRetro: Retrosynthetic Pathway Discovery using Iterative LLM Reasoning0
The Future is Agentic: Definitions, Perspectives, and Open Challenges of Multi-Agent Recommender Systems0
GAF-Guard: An Agentic Framework for Risk Management and Governance in Large Language ModelsCode0
HalluSegBench: Counterfactual Visual Reasoning for Segmentation Hallucination Evaluation0
Mitigating Hallucination of Large Vision-Language Models via Dynamic Logits CalibrationCode0
Seeing is Believing? Mitigating OCR Hallucinations in Multimodal Large Language Models0
Feature Hallucination for Self-supervised Action Recognition0
KnowRL: Exploring Knowledgeable Reinforcement Learning for FactualityCode1
DiscoSG: Towards Discourse-Level Text Scene Graph Parsing through Iterative Graph RefinementCode2
Robust Instant Policy: Leveraging Student's t-Regression Model for Robust In-context Imitation Learning of Robot Manipulation0
HEAL: An Empirical Study on Hallucinations in Embodied Agents Driven by Large Language Models0
ASCD: Attention-Steerable Contrastive Decoding for Reducing Hallucination in MLLM0
Abstract Meaning Representation for Hospital Discharge SummarizationCode0
DREAM: On hallucinations in AI-generated content for nuclear medicine imaging0
Stress-Testing Multimodal Foundation Models for Crystallographic ReasoningCode0
A Regret Perspective on Online Selective Generation0
VL-GenRM: Enhancing Vision-Language Verification via Vision Experts and Iterative Training0
HKD4VLM: A Progressive Hybrid Knowledge Distillation Framework for Robust Multimodal Hallucination and Factuality Detection in VLMs0
Second Order State Hallucinations for Adversarial Attack Mitigation in Formation Control of Multi-Agent Systems0
DiffFuSR: Super-Resolution of all Sentinel-2 Multispectral Bands using Diffusion ModelsCode1
HalLoc: Token-level Localization of Hallucinations for Vision Language ModelsCode0
Generalization or Hallucination? Understanding Out-of-Context Reasoning in Transformers0
Attention Head Embeddings with Trainable Deep Kernels for Hallucination Detection in LLMs0
Text-Aware Image Restoration with Diffusion Models0
Step-by-step Instructions and a Simple Tabular Output Format Improve the Dependency Parsing Accuracy of LLMsCode0
ViCrit: A Verifiable Reinforcement Learning Proxy Task for Visual Perception in VLMsCode1
Revisit What You See: Disclose Language Prior in Vision Tokens for Efficient Guided Decoding of LVLMsCode1
RHealthTwin: Towards Responsible and Multimodal Digital Twins for Personalized Well-being0
SECOND: Mitigating Perceptual Hallucination in Vision-Language Models via Selective and Contrastive DecodingCode0
MedChat: A Multi-Agent Framework for Multimodal Diagnosis with Large Language ModelsCode1
Uncertainty-o: One Model-agnostic Framework for Unveiling Uncertainty in Large Multimodal Models0
MEMOIR: Lifelong Model Editing with Minimal Overwrite and Informed Retention for LLMs0
ARGUS: Hallucination and Omission Evaluation in Video-LLMs0
Conservative Bias in Large Language Models: Measuring Relation Predictions0
Hallucination at a Glance: Controlled Visual Edits and Fine-Grained Multimodal Learning0
Reducing Object Hallucination in Large Audio-Language Models via Audio-Aware Decoding0
QuantMCP: Grounding Large Language Models in Verifiable Financial Reality0
Joint Evaluation of Answer and Reasoning Consistency for Hallucination Detection in Large Reasoning ModelsCode1
Safe: Enhancing Mathematical Reasoning in Large Language Models via Retrospective Step-aware Formal VerificationCode1
CLATTER: Comprehensive Entailment Reasoning for Hallucination Detection0
When Thinking LLMs Lie: Unveiling the Strategic Deception in Representations of Reasoning Models0
GOLFer: Smaller LM-Generated Documents Hallucination Filter & Combiner for Query Expansion in Information RetrievalCode0
On the Fundamental Impossibility of Hallucination Control in Large Language Models0
CHIME: Conditional Hallucination and Integrated Multi-scale Enhancement for Time Series Diffusion Model0
OWMM-Agent: Open World Mobile Manipulation With Multi-modal Agentic Data SynthesisCode1
Magic Mushroom: A Customizable Benchmark for Fine-grained Analysis of Retrieval Noise Erosion in RAG Systems0
Show:102550
← PrevPage 1 of 37Next →

No leaderboard results yet.