| M2K-VDG: Model-Adaptive Multimodal Knowledge Anchor Enhanced Video-grounded Dialogue Generation | Feb 19, 2024 | counterfactualDialogue Generation | —Unverified | 0 |
| Enabling Weak LLMs to Judge Response Reliability via Meta Ranking | Feb 19, 2024 | HallucinationIn-Context Learning | —Unverified | 0 |
| Reformatted Alignment | Feb 19, 2024 | GSM8KHallucination | CodeCode Available | 2 |
| Vision-Flan: Scaling Human-Labeled Tasks in Visual Instruction Tuning | Feb 18, 2024 | HallucinationVisual Question Answering | —Unverified | 0 |
| EventRL: Enhancing Event Extraction with Outcome Supervision for Large Language Models | Feb 18, 2024 | Event ExtractionHallucination | CodeCode Available | 3 |
| Aligning Modalities in Vision Large Language Models via Preference Fine-tuning | Feb 18, 2024 | HallucinationInstruction Following | CodeCode Available | 2 |
| Logical Closed Loop: Uncovering Object Hallucinations in Large Vision-Language Models | Feb 18, 2024 | HallucinationObject | CodeCode Available | 1 |
| LLMs in the Heart of Differential Testing: A Case Study on a Medical Rule Engine | Feb 16, 2024 | Hallucination | —Unverified | 0 |
| Using Hallucinations to Bypass GPT4's Filter | Feb 16, 2024 | Hallucination | —Unverified | 0 |
| Comparing Hallucination Detection Metrics for Multilingual Generation | Feb 16, 2024 | HallucinationNatural Language Inference | —Unverified | 0 |
| LLMDFA: Analyzing Dataflow in Code with Large Language Models | Feb 16, 2024 | Hallucination | CodeCode Available | 3 |
| Measuring and Reducing LLM Hallucination without Gold-Standard Answers | Feb 16, 2024 | HallucinationIn-Context Learning | —Unverified | 0 |
| Retrieve Only When It Needs: Adaptive Retrieval Augmentation for Hallucination Mitigation in Large Language Models | Feb 16, 2024 | HallucinationRetrieval | —Unverified | 0 |
| Towards Uncovering How Large Language Model Works: An Explainability Perspective | Feb 16, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| Trading off Consistency and Dimensionality of Convex Surrogates for the Mode | Feb 16, 2024 | HallucinationInformation Retrieval | —Unverified | 0 |
| EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language Models | Feb 15, 2024 | HallucinationObject Hallucination | CodeCode Available | 1 |
| Uncertainty Quantification for In-Context Learning of Large Language Models | Feb 15, 2024 | HallucinationIn-Context Learning | CodeCode Available | 1 |
| Do LLMs Know about Hallucination? An Empirical Investigation of LLM's Hidden States | Feb 15, 2024 | Hallucination | —Unverified | 0 |
| Visually Dehallucinative Instruction Generation: Know What You Don't Know | Feb 15, 2024 | HallucinationLanguage Modeling | CodeCode Available | 0 |
| Into the Unknown: Self-Learning Large Language Models | Feb 14, 2024 | HallucinationSelf-Learning | CodeCode Available | 1 |
| Large Language Model with Graph Convolution for Recommendation | Feb 14, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| LLMAuditor: A Framework for Auditing Large Language Models Using Human-in-the-Loop | Feb 14, 2024 | HallucinationTruthfulQA | —Unverified | 0 |
| InstructGraph: Boosting Large Language Models via Graph-centric Instruction Tuning and Preference Alignment | Feb 13, 2024 | Hallucination | CodeCode Available | 2 |
| Visually Dehallucinative Instruction Generation | Feb 13, 2024 | HallucinationLanguage Modeling | CodeCode Available | 0 |
| Mitigating Object Hallucination in Large Vision-Language Models via Classifier-Free Guidance | Feb 13, 2024 | HallucinationObject Hallucination | —Unverified | 0 |
| A Systematic Review of Data-to-Text NLG | Feb 13, 2024 | Data-to-Text GenerationHallucination | —Unverified | 0 |
| Careless Whisper: Speech-to-Text Hallucination Harms | Feb 12, 2024 | HallucinationLanguage Modeling | CodeCode Available | 0 |
| Prismatic VLMs: Investigating the Design Space of Visually-Conditioned Language Models | Feb 12, 2024 | HallucinationObject Localization | CodeCode Available | 4 |
| PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models | Feb 12, 2024 | Answer GenerationHallucination | CodeCode Available | 3 |
| G-Retriever: Retrieval-Augmented Generation for Textual Graph Understanding and Question Answering | Feb 12, 2024 | Common Sense ReasoningGraph Classification | CodeCode Available | 4 |
| Gemini Goes to Med School: Exploring the Capabilities of Multimodal Large Language Models on Medical Challenge Problems & Hallucinations | Feb 10, 2024 | DiagnosticHallucination | CodeCode Available | 1 |
| GLaM: Fine-Tuning Large Language Models for Domain Knowledge Graph Alignment via Neighborhood Partitioning and Generative Subgraph Encoding | Feb 9, 2024 | HallucinationKnowledge Graphs | —Unverified | 0 |
| ViGoR: Improving Visual Grounding of Large Vision Language Models with Fine-Grained Reward Modeling | Feb 9, 2024 | HallucinationNatural Language Understanding | CodeCode Available | 0 |
| ResumeFlow: An LLM-facilitated Pipeline for Personalized Resume Generation and Refinement | Feb 9, 2024 | HallucinationLanguage Modelling | CodeCode Available | 3 |
| Introspective Planning: Aligning Robots' Uncertainty with Inherent Task Ambiguity | Feb 9, 2024 | Conformal PredictionHallucination | CodeCode Available | 1 |
| An Examination on the Effectiveness of Divide-and-Conquer Prompting in Large Language Models | Feb 8, 2024 | Fact VerificationFake News Detection | —Unverified | 0 |
| Enhancing Retrieval Processes for Language Generation with Augmented Queries | Feb 6, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| INSIDE: LLMs' Internal States Retain the Power of Hallucination Detection | Feb 6, 2024 | DiversityHallucination | CodeCode Available | 1 |
| Training Language Models to Generate Text with Citations via Fine-grained Rewards | Feb 6, 2024 | HallucinationQuestion Answering | CodeCode Available | 1 |
| The Instinctive Bias: Spurious Images lead to Illusion in MLLMs | Feb 6, 2024 | Hallucination | CodeCode Available | 0 |
| Unified Hallucination Detection for Multimodal Large Language Models | Feb 5, 2024 | Hallucination | CodeCode Available | 1 |
| Improving Assessment of Tutoring Practices using Retrieval-Augmented Generation | Feb 4, 2024 | HallucinationMath | —Unverified | 0 |
| Aligner: Efficient Alignment by Learning to Correct | Feb 4, 2024 | Hallucination | —Unverified | 0 |
| LLM-Enhanced Data Management | Feb 4, 2024 | HallucinationManagement | CodeCode Available | 4 |
| A Closer Look at the Limitations of Instruction Tuning | Feb 3, 2024 | Hallucination | —Unverified | 0 |
| A Survey on Large Language Model Hallucination via a Creativity Perspective | Feb 2, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| CorpusLM: Towards a Unified Language Model on Corpus for Knowledge-Intensive Tasks | Feb 2, 2024 | Answer GenerationHallucination | —Unverified | 0 |
| Skip : A Simple Method to Reduce Hallucination in Large Vision-Language Models | Feb 2, 2024 | Hallucination | CodeCode Available | 1 |
| PokeLLMon: A Human-Parity Agent for Pokemon Battles with Large Language Models | Feb 2, 2024 | Action GenerationDecision Making | CodeCode Available | 3 |
| Redefining "Hallucination" in LLMs: Towards a psychology-informed framework for mitigating misinformation | Feb 1, 2024 | HallucinationMisinformation | —Unverified | 0 |