| What does it take to get state of the art in simultaneous speech-to-speech translation? | Sep 2, 2024 | HallucinationManagement | —Unverified | 0 |
| What External Knowledge is Preferred by LLMs? Characterizing and Exploring Chain of Evidence in Imperfect Context | Dec 17, 2024 | HallucinationMisinformation | —Unverified | 0 |
| What Matters in Memorizing and Recalling Facts? Multifaceted Benchmarks for Knowledge Probing in Language Models | Jun 18, 2024 | DecoderHallucination | —Unverified | 0 |
| When Not to Answer: Evaluating Prompts on GPT Models for Effective Abstention in Unanswerable Math Word Problems | Oct 16, 2024 | HallucinationMath | —Unverified | 0 |
| When Thinking LLMs Lie: Unveiling the Strategic Deception in Representations of Reasoning Models | Jun 5, 2025 | HallucinationMisinformation | —Unverified | 0 |
| When to Speak, When to Abstain: Contrastive Decoding with Abstention | Dec 17, 2024 | HallucinationQuestion Answering | —Unverified | 0 |
| Whispers that Shake Foundations: Analyzing and Mitigating False Premise Hallucinations in Large Language Models | Feb 29, 2024 | Hallucination | —Unverified | 0 |
| Who Brings the Frisbee: Probing Hidden Hallucination Factors in Large Vision-Language Model via Causality Analysis | Dec 4, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| "Why is this misleading?": Detecting News Headline Hallucinations with Explanations | Feb 12, 2023 | HallucinationHeadline Generation | —Unverified | 0 |
| WildHallucinations: Evaluating Long-form Factuality in LLMs with Real-World Entity Queries | Jul 24, 2024 | ChatbotForm | —Unverified | 0 |