| Cracking the Code of Hallucination in LVLMs with Vision-aware Head Divergence | Dec 18, 2024 | HallucinationMultimodal Reasoning | —Unverified | 0 |
| Are LLMs Good Literature Review Writers? Evaluating the Literature Review Writing Ability of Large Language Models | Dec 18, 2024 | Hallucination | —Unverified | 0 |
| ReXTrust: A Model for Fine-Grained Hallucination Detection in AI-Generated Radiology Reports | Dec 17, 2024 | Hallucination | —Unverified | 0 |
| A MapReduce Approach to Effectively Utilize Long Context Information in Retrieval Augmented Language Models | Dec 17, 2024 | HallucinationRAG | —Unverified | 0 |
| What External Knowledge is Preferred by LLMs? Characterizing and Exploring Chain of Evidence in Imperfect Context | Dec 17, 2024 | HallucinationMisinformation | —Unverified | 0 |
| When to Speak, When to Abstain: Contrastive Decoding with Abstention | Dec 17, 2024 | HallucinationQuestion Answering | —Unverified | 0 |
| A Benchmark and Robustness Study of In-Context-Learning with Large Language Models in Music Entity Detection | Dec 16, 2024 | HallucinationIn-Context Learning | CodeCode Available | 0 |
| Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial Reasoning | Dec 16, 2024 | HallucinationRobot Manipulation | CodeCode Available | 2 |
| CG-Bench: Clue-grounded Question Answering Benchmark for Long Video Understanding | Dec 16, 2024 | HallucinationMultiple-choice | —Unverified | 0 |
| Task-Oriented Dialog Systems for the Senegalese Wolof Language | Dec 15, 2024 | ChatbotHallucination | —Unverified | 0 |