| Safe: Enhancing Mathematical Reasoning in Large Language Models via Retrospective Step-aware Formal Verification | Jun 5, 2025 | Automated Theorem ProvingHallucination | CodeCode Available | 1 |
| OWMM-Agent: Open World Mobile Manipulation With Multi-modal Agentic Data Synthesis | Jun 4, 2025 | Action GenerationDecision Making | CodeCode Available | 1 |
| FlySearch: Exploring how vision-language models explore | Jun 3, 2025 | HallucinationTask Planning | CodeCode Available | 1 |
| The Hallucination Dilemma: Factuality-Aware Reinforcement Learning for Large Reasoning Models | May 30, 2025 | HallucinationMathematical Reasoning | CodeCode Available | 1 |
| CogniBench: A Legal-inspired Framework and Dataset for Assessing Cognitive Faithfulness of Large Language Models | May 27, 2025 | HallucinationLanguage Modeling | CodeCode Available | 1 |
| R3-RAG: Learning Step-by-Step Reasoning and Retrieval for LLMs via Reinforcement Learning | May 26, 2025 | HallucinationRAG | CodeCode Available | 1 |
| Removal of Hallucination on Hallucination: Debate-Augmented RAG | May 24, 2025 | HallucinationRAG | CodeCode Available | 1 |
| Mitigating Hallucinations in Vision-Language Models through Image-Guided Head Suppression | May 22, 2025 | HallucinationImage Description | CodeCode Available | 1 |
| Know Or Not: a library for evaluating out-of-knowledge base robustness | May 19, 2025 | HallucinationRAG | CodeCode Available | 1 |
| Phare: A Safety Probe for Large Language Models | May 16, 2025 | DiagnosticHallucination | CodeCode Available | 1 |