| Learning to Generate and Evaluate Fact-checking Explanations with Transformers | Oct 21, 2024 | Fact CheckingHallucination | —Unverified | 0 |
| A Survey of Hallucination in Large Visual Language Models | Oct 20, 2024 | HallucinationHallucination Evaluation | —Unverified | 0 |
| Hallucination Detox: Sensitivity Dropout (SenD) for Large Language Model Training | Oct 20, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| Explaining Graph Neural Networks with Large Language Models: A Counterfactual Perspective for Molecular Property Prediction | Oct 19, 2024 | counterfactualCounterfactual Explanation | CodeCode Available | 0 |
| Coarse-to-Fine Highlighting: Reducing Knowledge Hallucination in Large Language Models | Oct 19, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| Good Parenting is all you need -- Multi-agentic LLM Hallucination Mitigation | Oct 18, 2024 | AllHallucination | —Unverified | 0 |
| ELOQ: Resources for Enhancing LLM Detection of Out-of-Scope Questions | Oct 18, 2024 | HallucinationNatural Questions | CodeCode Available | 0 |
| Paths-over-Graph: Knowledge Graph Empowered Large Language Model Reasoning | Oct 18, 2024 | HallucinationKnowledge Base Question Answering | CodeCode Available | 1 |
| ETF: An Entity Tracing Framework for Hallucination Detection in Code Summaries | Oct 17, 2024 | Code SummarizationHallucination | —Unverified | 0 |
| From Single to Multi: How LLMs Hallucinate in Multi-Document Summarization | Oct 17, 2024 | Document SummarizationHallucination | CodeCode Available | 0 |
| MCQG-SRefine: Multiple Choice Question Generation and Evaluation with Iterative Self-Critique, Correction, and Comparison Feedback | Oct 17, 2024 | Fact VerificationHallucination | CodeCode Available | 0 |
| Mitigating Hallucinations in Large Vision-Language Models via Summary-Guided Decoding | Oct 17, 2024 | HallucinationObject Hallucination | CodeCode Available | 1 |
| Utilizing Large Language Models in an iterative paradigm with domain feedback for zero-shot molecule optimization | Oct 17, 2024 | Drug DiscoveryHallucination | —Unverified | 0 |
| FaithBench: A Diverse Hallucination Benchmark for Summarization by Modern LLMs | Oct 17, 2024 | DiversityHallucination | CodeCode Available | 1 |
| RosePO: Aligning LLM-based Recommenders with Human Values | Oct 16, 2024 | HallucinationRecommendation Systems | —Unverified | 0 |
| MMed-RAG: Versatile Multimodal RAG System for Medical Vision Language Models | Oct 16, 2024 | DiagnosticHallucination | CodeCode Available | 3 |
| On A Scale From 1 to 5: Quantifying Hallucination in Faithfulness Evaluation | Oct 16, 2024 | HallucinationNatural Language Inference | —Unverified | 0 |
| When Not to Answer: Evaluating Prompts on GPT Models for Effective Abstention in Unanswerable Math Word Problems | Oct 16, 2024 | HallucinationMath | —Unverified | 0 |
| Iter-AHMCL: Alleviate Hallucination for Large Language Model via Iterative Model-level Contrastive Learning | Oct 16, 2024 | Contrastive Learninggraph construction | —Unverified | 0 |
| Graph-constrained Reasoning: Faithful Reasoning on Knowledge Graphs with Large Language Models | Oct 16, 2024 | HallucinationKnowledge Graphs | CodeCode Available | 3 |
| What Do LLMs Need to Understand Graphs: A Survey of Parametric Representation of Graphs | Oct 16, 2024 | Drug DiscoveryGraph Generation | —Unverified | 0 |
| Controlled Automatic Task-Specific Synthetic Data Generation for Hallucination Detection | Oct 16, 2024 | HallucinationIn-Context Learning | —Unverified | 0 |
| A Claim Decomposition Benchmark for Long-form Answer Verification | Oct 16, 2024 | FormHallucination | CodeCode Available | 0 |
| The Curse of Multi-Modalities: Evaluating Hallucinations of Large Multimodal Models across Language, Visual, and Audio | Oct 16, 2024 | Hallucination | CodeCode Available | 3 |
| Search Engines in an AI Era: The False Promise of Factual and Verifiable Source-Cited Responses | Oct 15, 2024 | HallucinationLanguage Modeling | CodeCode Available | 1 |