| Attention! You Vision Language Model Could Be Maliciously Manipulated | May 26, 2025 | Decision MakingHallucination | —Unverified | 0 |
| Error Typing for Smarter Rewards: Improving Process Reward Models with Error-Aware Hierarchical Supervision | May 26, 2025 | HallucinationMath | CodeCode Available | 0 |
| Omni-R1: Reinforcement Learning for Omnimodal Reasoning via Two-System Collaboration | May 26, 2025 | Domain GeneralizationHallucination | CodeCode Available | 2 |
| Enhancing Visual Reliance in Text Generation: A Bayesian Perspective on Mitigating Hallucination in Large Vision-Language Models | May 26, 2025 | HallucinationMME | —Unverified | 0 |
| Uncertainty-Aware Attention Heads: Efficient Unsupervised Uncertainty Quantification for LLMs | May 26, 2025 | HallucinationQuestion Answering | —Unverified | 0 |
| Grounding Language with Vision: A Conditional Mutual Information Calibrated Decoding Strategy for Reducing Hallucinations in LVLMs | May 26, 2025 | Hallucination | —Unverified | 0 |
| LLLMs: A Data-Driven Survey of Evolving Research on Limitations of Large Language Models | May 25, 2025 | Hallucinationknowledge editing | —Unverified | 0 |
| CCHall: A Novel Benchmark for Joint Cross-Lingual and Cross-Modal Hallucinations Detection in Large Language Models | May 25, 2025 | Hallucination | CodeCode Available | 0 |
| GUARDIAN: Safeguarding LLM Multi-Agent Collaborations with Temporal Graph Modeling | May 25, 2025 | DecoderHallucination | —Unverified | 0 |
| MedScore: Factuality Evaluation of Free-Form Medical Answers | May 24, 2025 | FormHallucination | CodeCode Available | 0 |