| Beyond Words: On Large Language Models Actionability in Mission-Critical Risk Analysis | Jun 11, 2024 | HallucinationLanguage Modelling | —Unverified | 0 |
| A Debate-Driven Experiment on LLM Hallucinations and Accuracy | Oct 25, 2024 | Fact CheckingHallucination | —Unverified | 0 |
| Beyond Under-Alignment: Atomic Preference Enhanced Factuality Tuning for Large Language Models | Jun 18, 2024 | Hallucination | —Unverified | 0 |
| Beyond the Black Box: Interpretability of LLMs in Finance | May 14, 2025 | FairnessHallucination | —Unverified | 0 |
| DiDOTS: Knowledge Distillation from Large-Language-Models for Dementia Obfuscation in Transcribed Speech | Oct 5, 2024 | HallucinationKnowledge Distillation | —Unverified | 0 |
| An Automated Reinforcement Learning Reward Design Framework with Large Language Model for Cooperative Platoon Coordination | Apr 28, 2025 | Code GenerationHallucination | —Unverified | 0 |
| Cost-Effective Hallucination Detection for LLMs | Jul 31, 2024 | Decision MakingFact Checking | —Unverified | 0 |
| DHCP: Detecting Hallucinations by Cross-modal Attention Pattern in Large Vision-Language Models | Nov 27, 2024 | AttributeHallucination | —Unverified | 0 |
| Beyond Logit Lens: Contextual Embeddings for Robust Hallucination Detection & Grounding in VLMs | Nov 28, 2024 | AttributeHallucination | —Unverified | 0 |
| Anatomy of Industrial Scale Multilingual ASR | Apr 15, 2024 | AnatomyAutomatic Speech Recognition | —Unverified | 0 |
| Improving Zero-Shot ObjectNav with Generative Communication | Aug 3, 2024 | HallucinationNavigate | —Unverified | 0 |
| FRAME: Evaluating Rationale-Label Consistency Metrics for Free-Text Rationales | Jul 2, 2022 | HallucinationLanguage Modelling | —Unverified | 0 |
| Free-text Rationale Generation under Readability Level Control | Jul 1, 2024 | HallucinationText Generation | —Unverified | 0 |
| Developing a Reliable, Fast, General-Purpose Hallucination Detection and Mitigation Service | Jul 22, 2024 | Hallucinationnamed-entity-recognition | —Unverified | 0 |
| LLMAuditor: A Framework for Auditing Large Language Models Using Human-in-the-Loop | Feb 14, 2024 | HallucinationTruthfulQA | —Unverified | 0 |
| Detection and Mitigation of Hallucination in Large Reasoning Models: A Mechanistic Perspective | May 19, 2025 | Hallucination | —Unverified | 0 |
| An Analysis of Decoding Methods for LLM-based Agents for Faithful Multi-Hop Question Answering | Mar 30, 2025 | HallucinationMulti-hop Question Answering | —Unverified | 0 |
| Detecting LLM Hallucination Through Layer-wise Information Deficiency: Analysis of Unanswerable Questions and Ambiguous Prompts | Dec 13, 2024 | Hallucination | —Unverified | 0 |
| Detecting LLM Fact-conflicting Hallucinations Enhanced by Temporal-logic-based Reasoning | Feb 19, 2025 | Hallucination | —Unverified | 0 |
| Detecting Hallucinations in Virtual Histology with Neural Precursors | Nov 22, 2024 | HallucinationVirtual Staining | —Unverified | 0 |
| Detecting Hallucination and Coverage Errors in Retrieval Augmented Generation for Controversial Topics | Mar 13, 2024 | HallucinationRetrieval | —Unverified | 0 |
| Benchmarking Retrieval-Augmented Large Language Models in Biomedical NLP: Application, Robustness, and Self-Awareness | May 13, 2024 | Benchmarkingcounterfactual | —Unverified | 0 |
| Addressing Image Hallucination in Text-to-Image Generation through Factual Image Retrieval | Jul 15, 2024 | Common Sense ReasoningHallucination | —Unverified | 0 |
| Detecting Buggy Contracts via Smart Testing | Sep 6, 2024 | Hallucination | —Unverified | 0 |
| Analyzing LLM Behavior in Dialogue Summarization: Unveiling Circumstantial Hallucination Trends | Jun 5, 2024 | Hallucination | —Unverified | 0 |