| DiffMAC: Diffusion Manifold Hallucination Correction for High Generalization Blind Face Restoration | Mar 15, 2024 | AttributeBlind Face Restoration | —Unverified | 0 |
| AIGCs Confuse AI Too: Investigating and Explaining Synthetic Image-induced Hallucinations in Large Vision-Language Models | Mar 13, 2024 | Hallucination | CodeCode Available | 0 |
| Detecting Hallucination and Coverage Errors in Retrieval Augmented Generation for Controversial Topics | Mar 13, 2024 | HallucinationRetrieval | —Unverified | 0 |
| Investigating the performance of Retrieval-Augmented Generation and fine-tuning for the development of AI-driven knowledge-based systems | Mar 12, 2024 | Domain AdaptationHallucination | CodeCode Available | 0 |
| TRAWL: External Knowledge-Enhanced Recommendation with LLM Assistance | Mar 11, 2024 | Contrastive LearningDenoising | —Unverified | 0 |
| Put Myself in Your Shoes: Lifting the Egocentric Perspective from Exocentric Videos | Mar 11, 2024 | HallucinationTranslation | —Unverified | 0 |
| Guiding Clinical Reasoning with Large Language Models via Knowledge Seeds | Mar 11, 2024 | Hallucination | —Unverified | 0 |
| On the Benefits of Fine-Grained Loss Truncation: A Case Study on Factuality in Summarization | Mar 9, 2024 | HallucinationText Summarization | CodeCode Available | 0 |
| Tuning-Free Accountable Intervention for LLM Deployment -- A Metacognitive Approach | Mar 8, 2024 | Decision MakingHallucination | —Unverified | 0 |
| ERBench: An Entity-Relationship based Automatically Verifiable Hallucination Benchmark for Large Language Models | Mar 8, 2024 | AttributeHallucination | CodeCode Available | 0 |
| Can Large Language Models Play Games? A Case Study of A Self-Play Approach | Mar 8, 2024 | Decision MakingHallucination | —Unverified | 0 |
| Sora as an AGI World Model? A Complete Survey on Text-to-Video Generation | Mar 8, 2024 | ArticlesHallucination | —Unverified | 0 |
| ChatASU: Evoking LLM's Reflexion to Truly Understand Aspect Sentiment in Dialogues | Mar 8, 2024 | HallucinationQuestion Answering | —Unverified | 0 |
| Effectiveness Assessment of Recent Large Vision-Language Models | Mar 7, 2024 | Anomaly DetectionAttribute | —Unverified | 0 |
| Fact-Checking the Output of Large Language Models via Token-Level Uncertainty Quantification | Mar 7, 2024 | Fact CheckingHallucination | —Unverified | 0 |
| HaluEval-Wild: Evaluating Hallucinations of Language Models in the Wild | Mar 7, 2024 | HallucinationQuestion Answering | CodeCode Available | 0 |
| Benchmarking Hallucination in Large Language Models based on Unanswerable Math Word Problem | Mar 6, 2024 | BenchmarkingHallucination | CodeCode Available | 0 |
| German also Hallucinates! Inconsistency Detection in News Summaries with the Absinth Dataset | Mar 6, 2024 | HallucinationIn-Context Learning | CodeCode Available | 0 |
| The Claude 3 Model Family: Opus, Sonnet, Haiku | Mar 4, 2024 | 1 Image, 2*2 StitchingArithmetic Reasoning | —Unverified | 0 |
| Quantity Matters: Towards Assessing and Mitigating Number Hallucination in Large Vision-Language Models | Mar 3, 2024 | Hallucination | —Unverified | 0 |
| Right for Right Reasons: Large Language Models for Verifiable Commonsense Knowledge Graph Question Answering | Mar 3, 2024 | Claim VerificationGraph Question Answering | —Unverified | 0 |
| Self-Consistent Decoding for More Factual Open Responses | Mar 1, 2024 | HallucinationResponse Generation | CodeCode Available | 0 |
| MALTO at SemEval-2024 Task 6: Leveraging Synthetic Data for LLM Hallucination Detection | Mar 1, 2024 | Data AugmentationHallucination | —Unverified | 0 |
| Crimson: Empowering Strategic Reasoning in Cybersecurity through Large Language Models | Mar 1, 2024 | HallucinationRetrieval | —Unverified | 0 |
| Whispers that Shake Foundations: Analyzing and Mitigating False Premise Hallucinations in Large Language Models | Feb 29, 2024 | Hallucination | —Unverified | 0 |