| BordIRlines: A Dataset for Evaluating Cross-lingual Retrieval-Augmented Generation | Oct 2, 2024 | HallucinationRAG | CodeCode Available | 0 |
| Enhancing Training Data Attribution for Large Language Models with Fitting Error Consideration | Oct 2, 2024 | Hallucination | —Unverified | 0 |
| The Labyrinth of Links: Navigating the Associative Maze of Multi-modal LLMs | Oct 2, 2024 | BenchmarkingHallucination | —Unverified | 0 |
| FactAlign: Long-form Factuality Alignment of Large Language Models | Oct 2, 2024 | FormHallucination | CodeCode Available | 1 |
| LMOD: A Large Multimodal Ophthalmology Dataset and Benchmark for Large Vision-Language Models | Oct 2, 2024 | Hallucination | —Unverified | 0 |
| ScVLM: Enhancing Vision-Language Model for Safety-Critical Event Understanding | Oct 1, 2024 | Contrastive LearningHallucination | CodeCode Available | 0 |
| VideoCLIP-XL: Advancing Long Description Understanding for Video CLIP Models | Oct 1, 2024 | Hallucinationtext similarity | —Unverified | 0 |
| Ingest-And-Ground: Dispelling Hallucinations from Continually-Pretrained LLMs with RAG | Sep 30, 2024 | HallucinationRAG | —Unverified | 0 |
| FaithEval: Can Your Language Model Stay Faithful to Context, Even If "The Moon is Made of Marshmallows" | Sep 30, 2024 | counterfactualHallucination | CodeCode Available | 2 |
| Contrastive Token Learning with Similarity Decay for Repetition Suppression in Machine Translation | Sep 30, 2024 | HallucinationMachine Translation | —Unverified | 0 |
| HELPD: Mitigating Hallucination of LVLMs by Hierarchical Feedback Learning with Vision-enhanced Penalty Decoding | Sep 30, 2024 | HallucinationObject | CodeCode Available | 0 |
| LLM Hallucinations in Practical Code Generation: Phenomena, Mechanism, and Mitigation | Sep 30, 2024 | Code GenerationHallucination | CodeCode Available | 0 |
| MedHalu: Hallucinations in Responses to Healthcare Queries by Large Language Models | Sep 29, 2024 | Hallucination | —Unverified | 0 |
| DENEB: A Hallucination-Robust Automatic Evaluation Metric for Image Captioning | Sep 28, 2024 | HallucinationImage Captioning | —Unverified | 0 |
| HaloScope: Harnessing Unlabeled LLM Generations for Hallucination Detection | Sep 26, 2024 | Hallucination | CodeCode Available | 0 |
| Enhancing Guardrails for Safe and Secure Healthcare AI | Sep 25, 2024 | HallucinationMisinformation | —Unverified | 0 |
| RoleBreak: Character Hallucination as a Jailbreak Attack in Role-Playing Systems | Sep 25, 2024 | Hallucination | —Unverified | 0 |
| Pre-trained Language Models Return Distinguishable Probability Distributions to Unfaithfully Hallucinated Texts | Sep 25, 2024 | Hallucination | CodeCode Available | 0 |
| EventHallusion: Diagnosing Event Hallucinations in Video LLMs | Sep 25, 2024 | HallucinationInstruction Following | CodeCode Available | 1 |
| XTRUST: On the Multilingual Trustworthiness of Large Language Models | Sep 24, 2024 | EthicsFairness | CodeCode Available | 1 |
| A Unified Hallucination Mitigation Framework for Large Vision-Language Models | Sep 24, 2024 | HallucinationQuestion Answering | CodeCode Available | 0 |
| Controlling Risk of Retrieval-augmented Generation: A Counterfactual Prompting Framework | Sep 24, 2024 | Benchmarkingcounterfactual | CodeCode Available | 0 |
| Long-horizon Embodied Planning with Implicit Logical Inference and Hallucination Mitigation | Sep 24, 2024 | DiversityHallucination | —Unverified | 0 |
| Planning in the Dark: LLM-Symbolic Planning Pipeline without Experts | Sep 24, 2024 | Hallucination | —Unverified | 0 |
| AsthmaBot: Multi-modal, Multi-Lingual Retrieval Augmented Generation For Asthma Patient Support | Sep 24, 2024 | HallucinationQuestion Answering | —Unverified | 0 |
| Enhancing Text-to-SQL Capabilities of Large Language Models via Domain Database Knowledge Injection | Sep 24, 2024 | HallucinationSemantic Parsing | —Unverified | 0 |
| Parse Trees Guided LLM Prompt Compression | Sep 23, 2024 | Hallucination | CodeCode Available | 0 |
| A Preliminary Study of o1 in Medicine: Are We Closer to an AI Doctor? | Sep 23, 2024 | HallucinationMedQA | —Unverified | 0 |
| Enhancing Scientific Reproducibility Through Automated BioCompute Object Creation Using Retrieval-Augmented Generation from Publications | Sep 23, 2024 | HallucinationLong-Context Understanding | —Unverified | 0 |
| Effectively Enhancing Vision Language Large Models by Prompt Augmentation and Caption Utilization | Sep 22, 2024 | HallucinationHallucination Evaluation | CodeCode Available | 0 |
| Contrastive Learning for Knowledge-Based Question Generation in Large Language Models | Sep 21, 2024 | Contrastive LearningHallucination | —Unverified | 0 |
| FAIR GPT: A virtual consultant for research data management in ChatGPT | Sep 20, 2024 | FairnessHallucination | CodeCode Available | 1 |
| A Multiple-Fill-in-the-Blank Exam Approach for Enhancing Zero-Resource Hallucination Detection in Large Language Models | Sep 20, 2024 | HallucinationSentence | —Unverified | 0 |
| FIHA: Autonomous Hallucination Evaluation in Vision-Language Models with Davidson Scene Graphs | Sep 20, 2024 | HallucinationHallucination Evaluation | —Unverified | 0 |
| JourneyBench: A Challenging One-Stop Vision-Language Understanding Benchmark of Generated Images | Sep 19, 2024 | HallucinationImage Captioning | CodeCode Available | 0 |
| Textualized Agent-Style Reasoning for Complex Tasks by Multiple Round LLM Generation | Sep 19, 2024 | Hallucination | —Unverified | 0 |
| LLMs Can Check Their Own Results to Mitigate Hallucinations in Traffic Understanding Tasks | Sep 19, 2024 | Autonomous DrivingHallucination | —Unverified | 0 |
| Evaluating Image Hallucination in Text-to-Image Generation with Question-Answering | Sep 19, 2024 | HallucinationHallucination Evaluation | CodeCode Available | 1 |
| Depth-based Privileged Information for Boosting 3D Human Pose Estimation on RGB | Sep 17, 2024 | 3D Human Pose EstimationHallucination | —Unverified | 0 |
| Zero-resource Hallucination Detection for Text Generation via Graph-based Contextual Knowledge Triples Modeling | Sep 17, 2024 | HallucinationText Generation | —Unverified | 0 |
| Exploring the Trade-Offs: Quantization Methods, Task Difficulty, and Model Size in Large Language Models From Edge to Giant | Sep 17, 2024 | HallucinationInstruction Following | CodeCode Available | 0 |
| THaMES: An End-to-End Tool for Hallucination Mitigation and Evaluation in Large Language Models | Sep 17, 2024 | BenchmarkingBinary Classification | CodeCode Available | 0 |
| Optimizing Resource Consumption in Diffusion Models through Hallucination Early Detection | Sep 16, 2024 | Hallucination | —Unverified | 0 |
| SFR-RAG: Towards Contextually Faithful LLMs | Sep 16, 2024 | counterfactualHallucination | —Unverified | 0 |
| Trustworthiness in Retrieval-Augmented Generation Systems: A Survey | Sep 16, 2024 | FairnessHallucination | CodeCode Available | 1 |
| HALO: Hallucination Analysis and Learning Optimization to Empower LLMs with Retrieval-Augmented Context for Guided Clinical Decision Making | Sep 16, 2024 | Answer GenerationDecision Making | CodeCode Available | 0 |
| Confidence Estimation for LLM-Based Dialogue State Tracking | Sep 15, 2024 | Dialogue State TrackingHallucination | CodeCode Available | 0 |
| Explore the Hallucination on Low-level Perception for MLLMs | Sep 15, 2024 | HallucinationQuestion Answering | —Unverified | 0 |
| ODE: Open-Set Evaluation of Hallucinations in Multimodal Large Language Models | Sep 14, 2024 | AttributeHallucination | —Unverified | 0 |
| Winning Solution For Meta KDD Cup' 24 | Sep 13, 2024 | HallucinationKnowledge Graphs | —Unverified | 0 |