| M2K-VDG: Model-Adaptive Multimodal Knowledge Anchor Enhanced Video-grounded Dialogue Generation | Feb 19, 2024 | counterfactualDialogue Generation | —Unverified | 0 |
| Enabling Weak LLMs to Judge Response Reliability via Meta Ranking | Feb 19, 2024 | HallucinationIn-Context Learning | —Unverified | 0 |
| Reformatted Alignment | Feb 19, 2024 | GSM8KHallucination | CodeCode Available | 2 |
| Vision-Flan: Scaling Human-Labeled Tasks in Visual Instruction Tuning | Feb 18, 2024 | HallucinationVisual Question Answering | —Unverified | 0 |
| EventRL: Enhancing Event Extraction with Outcome Supervision for Large Language Models | Feb 18, 2024 | Event ExtractionHallucination | CodeCode Available | 3 |
| Aligning Modalities in Vision Large Language Models via Preference Fine-tuning | Feb 18, 2024 | HallucinationInstruction Following | CodeCode Available | 2 |
| Logical Closed Loop: Uncovering Object Hallucinations in Large Vision-Language Models | Feb 18, 2024 | HallucinationObject | CodeCode Available | 1 |
| LLMs in the Heart of Differential Testing: A Case Study on a Medical Rule Engine | Feb 16, 2024 | Hallucination | —Unverified | 0 |
| Using Hallucinations to Bypass GPT4's Filter | Feb 16, 2024 | Hallucination | —Unverified | 0 |
| Comparing Hallucination Detection Metrics for Multilingual Generation | Feb 16, 2024 | HallucinationNatural Language Inference | —Unverified | 0 |