| Ornithologist: Towards Trustworthy "Reasoning" about Central Bank Communications | May 14, 2025 | HallucinationLanguage Modeling | —Unverified | 0 |
| Prioritizing Image-Related Tokens Enhances Vision-Language Pre-Training | May 13, 2025 | HallucinationLarge Language Model | CodeCode Available | 0 |
| Improving the Reliability of LLMs: Combining CoT, RAG, Self-Consistency, and Self-Verification | May 13, 2025 | HallucinationRAG | —Unverified | 0 |
| Adaptive Schema-aware Event Extraction with Retrieval-Augmented Generation | May 13, 2025 | Event ExtractionHallucination | —Unverified | 0 |
| A Head to Predict and a Head to Question: Pre-trained Uncertainty Quantification Heads for Hallucination Detection in LLM Outputs | May 13, 2025 | HallucinationUncertainty Quantification | CodeCode Available | 1 |
| SEReDeEP: Hallucination Detection in Retrieval-Augmented Models via Semantic Entropy and Context-Parameter Fusion | May 12, 2025 | HallucinationRAG | —Unverified | 0 |
| On the Cost and Benefits of Training Context with Utterance or Full Conversation Training: A Comparative Stud | May 12, 2025 | GPUHallucination | —Unverified | 0 |
| Multimodal Survival Modeling in the Age of Foundation Models | May 12, 2025 | HallucinationSurvival Prediction | CodeCode Available | 0 |
| Critique Before Thinking: Mitigating Hallucination through Rationale-Augmented Instruction Tuning | May 12, 2025 | HallucinationMultimodal Reasoning | —Unverified | 0 |
| TrumorGPT: Graph-Based Retrieval-Augmented Large Language Model for Fact-Checking | May 11, 2025 | Fact CheckingFew-Shot Learning | —Unverified | 0 |