| Don't Believe Everything You Read: Enhancing Summarization Interpretability through Automatic Identification of Hallucinations in Large Language Models | Dec 22, 2023 | HallucinationMachine Translation | —Unverified | 0 |
| Theory of Hallucinations based on Equivariance | Dec 22, 2023 | Hallucination | —Unverified | 0 |
| Context-aware Decoding Reduces Hallucination in Query-focused Summarization | Dec 21, 2023 | HallucinationLanguage Modelling | CodeCode Available | 1 |
| Reducing Hallucinations: Enhancing VQA for Flood Disaster Damage Assessment with Visual Contexts | Dec 21, 2023 | HallucinationQuestion Answering | —Unverified | 0 |
| Experimenting with Large Language Models and vector embeddings in NASA SciX | Dec 21, 2023 | Data AugmentationHallucination | —Unverified | 0 |
| Quantifying Bias in Text-to-Image Generative Models | Dec 20, 2023 | HallucinationMarketing | —Unverified | 0 |
| On Early Detection of Hallucinations in Factual Question Answering | Dec 19, 2023 | HallucinationOpen-Ended Question Answering | CodeCode Available | 1 |
| MELO: Enhancing Model Editing with Neuron-Indexed Dynamic LoRA | Dec 19, 2023 | Document ClassificationHallucination | CodeCode Available | 0 |
| "Knowing When You Don't Know": A Multilingual Relevance Assessment Dataset for Robust Retrieval-Augmented Generation | Dec 18, 2023 | HallucinationLanguage Modelling | CodeCode Available | 1 |
| Retrieval-Augmented Generation for Large Language Models: A Survey | Dec 18, 2023 | HallucinationRAG | CodeCode Available | 4 |
| Silkie: Preference Distillation for Large Visual Language Models | Dec 17, 2023 | HallucinationMME | —Unverified | 0 |
| Towards Verifiable Text Generation with Evolving Memory and Self-Reflection | Dec 14, 2023 | HallucinationRetrieval | —Unverified | 0 |
| Vista-LLaMA: Reliable Video Narrator via Equal Distance to Visual Tokens | Dec 12, 2023 | HallucinationPosition | —Unverified | 0 |
| Improving Factual Error Correction by Learning to Inject Factual Errors | Dec 12, 2023 | Hallucination | CodeCode Available | 0 |
| Hallucination Augmented Contrastive Learning for Multimodal Large Language Model | Dec 12, 2023 | Contrastive LearningHallucination | CodeCode Available | 1 |
| Evaluating ChatGPT as a Question Answering System: A Comprehensive Analysis and Comparison with Existing Models | Dec 11, 2023 | HallucinationLanguage Modelling | —Unverified | 0 |
| Context Tuning for Retrieval Augmented Generation | Dec 9, 2023 | HallucinationRAG | —Unverified | 0 |
| Towards Enhanced Image Inpainting: Mitigating Unwanted Object Insertion and Preserving Color Consistency | Dec 8, 2023 | DecoderHallucination | CodeCode Available | 1 |
| DelucionQA: Detecting Hallucinations in Domain-specific Question Answering | Dec 8, 2023 | HallucinationInformation Retrieval | —Unverified | 0 |
| HALO: An Ontology for Representing and Categorizing Hallucinations in Large Language Models | Dec 8, 2023 | Hallucination | —Unverified | 0 |
| Mitigating Open-Vocabulary Caption Hallucinations | Dec 6, 2023 | DiversityHallucination | CodeCode Available | 1 |
| Weakly Supervised Detection of Hallucinations in LLM Activations | Dec 5, 2023 | HallucinationLanguage Modeling | CodeCode Available | 5 |
| Mitigating Fine-Grained Hallucination by Fine-Tuning Large Vision-Language Models with Caption Rewrites | Dec 4, 2023 | HallucinationHallucination Evaluation | CodeCode Available | 1 |
| Behind the Magic, MERLIM: Multi-modal Evaluation Benchmark for Large Image-Language Models | Dec 3, 2023 | HallucinationVisual Grounding | CodeCode Available | 0 |
| RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback | Dec 1, 2023 | HallucinationImage Captioning | CodeCode Available | 6 |
| On Exploring the Reasoning Capability of Large Language Models with Knowledge Graphs | Dec 1, 2023 | HallucinationKnowledge Graphs | —Unverified | 0 |
| Understanding Your Agent: Leveraging Large Language Models for Behavior Explanation | Nov 29, 2023 | counterfactualHallucination | —Unverified | 0 |
| OPERA: Alleviating Hallucination in Multi-Modal Large Language Models via Over-Trust Penalty and Retrospection-Allocation | Nov 29, 2023 | Hallucination | CodeCode Available | 2 |
| How to Build an AI Tutor That Can Adapt to Any Course Using Knowledge Graph-Enhanced Retrieval-Augmented Generation (KG-RAG) | Nov 29, 2023 | HallucinationKnowledge Graphs | —Unverified | 0 |
| Combating the "Sameness" in AI Art: Reflections on the Interactive AI Installation Fencing Hallucination | Nov 28, 2023 | Hallucination | —Unverified | 0 |
| Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decoding | Nov 28, 2023 | HallucinationObject | CodeCode Available | 2 |
| Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimization | Nov 28, 2023 | HallucinationMME | CodeCode Available | 1 |
| Mitigating Hallucination in Visual Language Models with Visual Supervision | Nov 27, 2023 | Hallucination | —Unverified | 0 |
| Deficiency of Large Language Models in Finance: An Empirical Examination of Hallucination | Nov 27, 2023 | Few-Shot LearningHallucination | —Unverified | 0 |
| UHGEval: Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained Generation | Nov 26, 2023 | BenchmarkingHallucination | CodeCode Available | 1 |
| Calibrated Language Models Must Hallucinate | Nov 24, 2023 | ArticlesHallucination | —Unverified | 0 |
| Challenges of Large Language Models for Mental Health Counseling | Nov 23, 2023 | HallucinationNavigate | —Unverified | 0 |
| Controlling Large Language Model-based Agents for Large-Scale Decision-Making: An Actor-Critic Approach | Nov 23, 2023 | Decision MakingHallucination | —Unverified | 0 |
| Minimizing Factual Inconsistency and Hallucination in Large Language Models | Nov 23, 2023 | HallucinationRAG | —Unverified | 0 |
| Enhancing Uncertainty-Based Hallucination Detection with Stronger Focus | Nov 22, 2023 | HallucinationRetrieval | CodeCode Available | 1 |
| HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction Data | Nov 22, 2023 | Attributecounterfactual | CodeCode Available | 1 |
| Mitigating Large Language Model Hallucinations via Autonomous Knowledge Graph-based Retrofitting | Nov 22, 2023 | HallucinationLanguage Modeling | —Unverified | 0 |
| Adapting LLMs for Efficient, Personalized Information Retrieval: Methods and Implications | Nov 21, 2023 | ChatbotHallucination | —Unverified | 0 |
| KNVQA: A Benchmark for evaluation knowledge-based VQA | Nov 21, 2023 | HallucinationObject Hallucination | —Unverified | 0 |
| Control in Hybrid Chatbots | Nov 20, 2023 | ChatbotHallucination | —Unverified | 0 |
| GPT-4V(ision) for Robotics: Multimodal Task Planning from Human Demonstration | Nov 20, 2023 | HallucinationLanguage Modeling | —Unverified | 0 |
| Chain of Visual Perception: Harnessing Multimodal Large Language Models for Zero-shot Camouflaged Object Detection | Nov 19, 2023 | counterfactualHallucination | CodeCode Available | 0 |
| Journey of Hallucination-minimized Generative AI Solutions for Financial Decision Makers | Nov 18, 2023 | Answer GenerationDecision Making | —Unverified | 0 |
| R-Tuning: Instructing Large Language Models to Say `I Don't Know' | Nov 16, 2023 | HallucinationSentence | CodeCode Available | 1 |
| Deceptive Semantic Shortcuts on Reasoning Chains: How Far Can Models Go without Hallucination? | Nov 16, 2023 | HallucinationSentence | CodeCode Available | 0 |