| Meaningless is better: hashing bias-inducing words in LLM prompts improves performance in logical reasoning and statistical learning | Nov 26, 2024 | HallucinationLogical Reasoning | —Unverified | 0 |
| AI2T: Building Trustable AI Tutors by Interactively Teaching a Self-Aware Learning Agent | Nov 26, 2024 | Hallucination | —Unverified | 0 |
| VidHal: Benchmarking Temporal Hallucinations in Vision LLMs | Nov 25, 2024 | BenchmarkingHallucination | CodeCode Available | 1 |
| AtomR: Atomic Operator-Empowered Large Language Models for Heterogeneous Knowledge Reasoning | Nov 25, 2024 | HallucinationQuestion Answering | CodeCode Available | 1 |
| Enhancing Multi-Agent Consensus through Third-Party LLM Integration: Analyzing Uncertainty and Mitigating Hallucinations in Large Language Models | Nov 25, 2024 | Hallucination | —Unverified | 0 |
| O1 Replication Journey -- Part 2: Surpassing O1-preview through Simple Distillation, Big Progress or Bitter Lesson? | Nov 25, 2024 | HallucinationKnowledge Distillation | CodeCode Available | 7 |
| VaLiD: Mitigating the Hallucination of Large Vision Language Models by Visual Layer Fusion Contrastive Decoding | Nov 24, 2024 | HallucinationLanguage Modeling | CodeCode Available | 1 |
| Devils in Middle Layers of Large Vision-Language Models: Interpreting, Detecting and Mitigating Object Hallucinations via Attention Lens | Nov 23, 2024 | Hallucination | CodeCode Available | 2 |
| Ontology-Constrained Generation of Domain-Specific Clinical Summaries | Nov 23, 2024 | HallucinationText Summarization | CodeCode Available | 0 |
| ICT: Image-Object Cross-Level Trusted Intervention for Mitigating Object Hallucination in Large Vision-Language Models | Nov 22, 2024 | HallucinationObject | —Unverified | 0 |