| R2-KG: General-Purpose Dual-Agent Framework for Reliable Reasoning on Knowledge Graphs | Feb 18, 2025 | HallucinationKnowledge Graphs | CodeCode Available | 1 |
| How Much Do LLMs Hallucinate across Languages? On Multilingual Estimation of LLM Hallucination in the Wild | Feb 18, 2025 | ArticlesHallucination | CodeCode Available | 0 |
| Unveiling the Magic of Code Reasoning through Hypothesis Decomposition and Amendment | Feb 17, 2025 | HallucinationLogical Reasoning | CodeCode Available | 2 |
| Can Your Uncertainty Scores Detect Hallucinated Entity? | Feb 17, 2025 | HallucinationSentence | —Unverified | 0 |
| Smoothing Out Hallucinations: Mitigating LLM Hallucination with Smoothed Knowledge Distillation | Feb 16, 2025 | HallucinationKnowledge Distillation | —Unverified | 0 |
| Valuable Hallucinations: Realizable Non-realistic Propositions | Feb 16, 2025 | Hallucination | —Unverified | 0 |
| A Survey of LLM-based Agents in Medicine: How far are we from Baymax? | Feb 16, 2025 | HallucinationSurvey | —Unverified | 0 |
| Automated Hypothesis Validation with Agentic Sequential Falsifications | Feb 14, 2025 | Decision MakingHallucination | CodeCode Available | 3 |
| Enhancing RAG with Active Learning on Conversation Records: Reject Incapables and Answer Capables | Feb 13, 2025 | Active LearningHallucination | —Unverified | 0 |
| DeepSeek on a Trip: Inducing Targeted Visual Hallucinations via Representation Vulnerabilities | Feb 11, 2025 | HallucinationSSIM | —Unverified | 0 |