| A Survey on Responsible LLMs: Inherent Risk, Malicious Use, and Mitigation Strategy | Jan 16, 2025 | HallucinationSurvey | —Unverified | 0 |
| ChartInsighter: An Approach for Mitigating Hallucination in Time-series Chart Summary Generation with A Benchmark Dataset | Jan 16, 2025 | HallucinationSentence | CodeCode Available | 1 |
| Knowledge Graph-based Retrieval-Augmented Generation for Schema Matching | Jan 15, 2025 | HallucinationKnowledge Graphs | CodeCode Available | 1 |
| Multimodal LLMs Can Reason about Aesthetics in Zero-Shot | Jan 15, 2025 | BenchmarkingHallucination | CodeCode Available | 1 |
| HALoGEN: Fantastic LLM Hallucinations and Where to Find Them | Jan 14, 2025 | HallucinationWorld Knowledge | —Unverified | 0 |
| Tarsier2: Advancing Large Vision-Language Models from Detailed Video Description to Comprehensive Video Understanding | Jan 14, 2025 | Embodied Question AnsweringHallucination | CodeCode Available | 4 |
| GPT as a Monte Carlo Language Tree: A Probabilistic Perspective | Jan 13, 2025 | Hallucination | —Unverified | 0 |
| Fine-tuning Large Language Models for Improving Factuality in Legal Question Answering | Jan 11, 2025 | HallucinationQuestion Answering | CodeCode Available | 0 |
| VASparse: Towards Efficient Visual Hallucination Mitigation for Large Vision-Language Model via Visual-Aware Sparsification | Jan 11, 2025 | HallucinationLanguage Modeling | CodeCode Available | 1 |
| MedCT: A Clinical Terminology Graph for Generative AI Applications in Healthcare | Jan 11, 2025 | DiagnosticEntity Linking | —Unverified | 0 |
| Hermit Kingdom Through the Lens of Multiple Perspectives: A Case Study of LLM Hallucination on North Korea | Jan 10, 2025 | HallucinationMisinformation | —Unverified | 0 |
| ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition Benchmark | Jan 9, 2025 | FairnessHallucination | CodeCode Available | 1 |
| Seeing with Partial Certainty: Conformal Prediction for Robotic Scene Recognition in Built Environments | Jan 9, 2025 | Conformal PredictionHallucination | —Unverified | 0 |
| Feedback-Driven Vision-Language Alignment with Minimal Human Supervision | Jan 8, 2025 | HallucinationQuestion Answering | —Unverified | 0 |
| RAG-Check: Evaluating Multimodal Retrieval Augmented Generation Performance | Jan 7, 2025 | HallucinationRAG | —Unverified | 0 |
| FlippedRAG: Black-Box Opinion Manipulation Adversarial Attacks to Retrieval-Augmented Generation Models | Jan 6, 2025 | Adversarial AttackHallucination | —Unverified | 0 |
| EAGLE: Enhanced Visual Grounding Minimizes Hallucinations in Instructional Multimodal Models | Jan 6, 2025 | HallucinationVisual Grounding | —Unverified | 0 |
| Socratic Questioning: Learn to Self-guide Multimodal Reasoning in the Wild | Jan 6, 2025 | HallucinationMultimodal Reasoning | CodeCode Available | 0 |
| Foundations of GenIR | Jan 6, 2025 | HallucinationRetrieval-augmented Generation | —Unverified | 0 |
| CHAIR -- Classifier of Hallucination as Improver | Jan 5, 2025 | HallucinationMMLU | CodeCode Available | 0 |
| A Survey of State of the Art Large Vision Language Models: Alignment, Benchmark, Evaluations and Challenges | Jan 4, 2025 | FairnessHallucination | CodeCode Available | 4 |
| CarbonChat: Large Language Model-Based Corporate Carbon Emission Analysis and Climate Knowledge Q&A System | Jan 3, 2025 | ChunkingHallucination | —Unverified | 0 |
| Mitigating Hallucination for Large Vision Language Model by Inter-Modality Correlation Calibration Decoding | Jan 3, 2025 | HallucinationLanguage Modeling | CodeCode Available | 1 |
| LLMs & Legal Aid: Understanding Legal Needs Exhibited Through User Queries | Jan 3, 2025 | Hallucinationzero-shot-classification | —Unverified | 0 |
| Enhancing Uncertainty Modeling with Semantic Graph for Hallucination Detection | Jan 2, 2025 | HallucinationSentence | —Unverified | 0 |