| BordIRlines: A Dataset for Evaluating Cross-lingual Retrieval-Augmented Generation | Oct 2, 2024 | HallucinationRAG | CodeCode Available | 0 |
| Enhancing Training Data Attribution for Large Language Models with Fitting Error Consideration | Oct 2, 2024 | Hallucination | —Unverified | 0 |
| The Labyrinth of Links: Navigating the Associative Maze of Multi-modal LLMs | Oct 2, 2024 | BenchmarkingHallucination | —Unverified | 0 |
| FactAlign: Long-form Factuality Alignment of Large Language Models | Oct 2, 2024 | FormHallucination | CodeCode Available | 1 |
| LMOD: A Large Multimodal Ophthalmology Dataset and Benchmark for Large Vision-Language Models | Oct 2, 2024 | Hallucination | —Unverified | 0 |
| VideoCLIP-XL: Advancing Long Description Understanding for Video CLIP Models | Oct 1, 2024 | Hallucinationtext similarity | —Unverified | 0 |
| ScVLM: Enhancing Vision-Language Model for Safety-Critical Event Understanding | Oct 1, 2024 | Contrastive LearningHallucination | CodeCode Available | 0 |
| FaithEval: Can Your Language Model Stay Faithful to Context, Even If "The Moon is Made of Marshmallows" | Sep 30, 2024 | counterfactualHallucination | CodeCode Available | 2 |
| Ingest-And-Ground: Dispelling Hallucinations from Continually-Pretrained LLMs with RAG | Sep 30, 2024 | HallucinationRAG | —Unverified | 0 |
| HELPD: Mitigating Hallucination of LVLMs by Hierarchical Feedback Learning with Vision-enhanced Penalty Decoding | Sep 30, 2024 | HallucinationObject | CodeCode Available | 0 |
| Contrastive Token Learning with Similarity Decay for Repetition Suppression in Machine Translation | Sep 30, 2024 | HallucinationMachine Translation | —Unverified | 0 |
| LLM Hallucinations in Practical Code Generation: Phenomena, Mechanism, and Mitigation | Sep 30, 2024 | Code GenerationHallucination | CodeCode Available | 0 |
| MedHalu: Hallucinations in Responses to Healthcare Queries by Large Language Models | Sep 29, 2024 | Hallucination | —Unverified | 0 |
| DENEB: A Hallucination-Robust Automatic Evaluation Metric for Image Captioning | Sep 28, 2024 | HallucinationImage Captioning | —Unverified | 0 |
| HaloScope: Harnessing Unlabeled LLM Generations for Hallucination Detection | Sep 26, 2024 | Hallucination | CodeCode Available | 0 |
| Enhancing Guardrails for Safe and Secure Healthcare AI | Sep 25, 2024 | HallucinationMisinformation | —Unverified | 0 |
| Pre-trained Language Models Return Distinguishable Probability Distributions to Unfaithfully Hallucinated Texts | Sep 25, 2024 | Hallucination | CodeCode Available | 0 |
| RoleBreak: Character Hallucination as a Jailbreak Attack in Role-Playing Systems | Sep 25, 2024 | Hallucination | —Unverified | 0 |
| EventHallusion: Diagnosing Event Hallucinations in Video LLMs | Sep 25, 2024 | HallucinationInstruction Following | CodeCode Available | 1 |
| A Unified Hallucination Mitigation Framework for Large Vision-Language Models | Sep 24, 2024 | HallucinationQuestion Answering | CodeCode Available | 0 |
| Controlling Risk of Retrieval-augmented Generation: A Counterfactual Prompting Framework | Sep 24, 2024 | Benchmarkingcounterfactual | CodeCode Available | 0 |
| XTRUST: On the Multilingual Trustworthiness of Large Language Models | Sep 24, 2024 | EthicsFairness | CodeCode Available | 1 |
| Planning in the Dark: LLM-Symbolic Planning Pipeline without Experts | Sep 24, 2024 | Hallucination | —Unverified | 0 |
| AsthmaBot: Multi-modal, Multi-Lingual Retrieval Augmented Generation For Asthma Patient Support | Sep 24, 2024 | HallucinationQuestion Answering | —Unverified | 0 |
| Long-horizon Embodied Planning with Implicit Logical Inference and Hallucination Mitigation | Sep 24, 2024 | DiversityHallucination | —Unverified | 0 |