| LLM-QE: Improving Query Expansion by Aligning Large Language Models with Ranking Preferences | Feb 24, 2025 | HallucinationInformation Retrieval | CodeCode Available | 1 |
| LettuceDetect: A Hallucination Detection Framework for RAG Applications | Feb 24, 2025 | 8kGPU | CodeCode Available | 4 |
| `Generalization is hallucination' through the lens of tensor completions | Feb 24, 2025 | HallucinationPosition | —Unverified | 0 |
| Uncertainty-Aware Fusion: An Ensemble Framework for Mitigating Hallucinations in Large Language Models | Feb 22, 2025 | HallucinationQuestion Answering | —Unverified | 0 |
| ZiGong 1.0: A Large Language Model for Financial Credit | Feb 22, 2025 | HallucinationLanguage Modeling | —Unverified | 0 |
| The Law of Knowledge Overshadowing: Towards Understanding, Predicting, and Preventing LLM Hallucination | Feb 22, 2025 | HallucinationText Generation | —Unverified | 0 |
| PIP-KAG: Mitigating Knowledge Conflicts in Knowledge-Augmented Generation via Parametric Pruning | Feb 21, 2025 | Hallucination | CodeCode Available | 2 |
| The Role of Background Information in Reducing Object Hallucination in Vision-Language Models: Insights from Cutoff API Prompting | Feb 21, 2025 | HallucinationObject | —Unverified | 0 |
| Hallucination Detection in Large Language Models with Metamorphic Relations | Feb 20, 2025 | Hallucination | —Unverified | 0 |
| Verify when Uncertain: Beyond Self-Consistency in Black Box Hallucination Detection | Feb 20, 2025 | Hallucination | —Unverified | 0 |
| Large Language Models Struggle to Describe the Haystack without Human Help: Human-in-the-loop Evaluation of LLMs | Feb 20, 2025 | HallucinationTopic Models | —Unverified | 0 |
| MedHallu: A Comprehensive Benchmark for Detecting Medical Hallucinations in Large Language Models | Feb 20, 2025 | Decision MakingHallucination | —Unverified | 0 |
| SegSub: Evaluating Robustness to Knowledge Conflicts and Hallucinations in Vision-Language Models | Feb 19, 2025 | counterfactualHallucination | CodeCode Available | 0 |
| OpenSearch-SQL: Enhancing Text-to-SQL with Dynamic Few-shot and Consistency Alignment | Feb 19, 2025 | HallucinationInstruction Following | —Unverified | 0 |
| Detecting LLM Fact-conflicting Hallucinations Enhanced by Temporal-logic-based Reasoning | Feb 19, 2025 | Hallucination | —Unverified | 0 |
| REFIND: Retrieval-Augmented Factuality Hallucination Detection in Large Language Models | Feb 19, 2025 | HallucinationLanguage Modeling | —Unverified | 0 |
| What are Models Thinking about? Understanding Large Language Model Hallucinations "Psychology" through Model Inner State Analysis | Feb 19, 2025 | HallucinationLanguage Modeling | —Unverified | 0 |
| TreeCut: A Synthetic Unanswerable Math Word Problem Dataset for LLM Hallucination Evaluation | Feb 19, 2025 | Dataset GenerationGSM8K | CodeCode Available | 0 |
| Lost in Transcription, Found in Distribution Shift: Demystifying Hallucination in Speech Foundation Models | Feb 18, 2025 | Automatic Speech RecognitionAutomatic Speech Recognition (ASR) | —Unverified | 0 |
| CutPaste&Find: Efficient Multimodal Hallucination Detector with Visual-aid Knowledge Base | Feb 18, 2025 | AttributeHallucination | —Unverified | 0 |
| R2-KG: General-Purpose Dual-Agent Framework for Reliable Reasoning on Knowledge Graphs | Feb 18, 2025 | HallucinationKnowledge Graphs | CodeCode Available | 1 |
| How Much Do LLMs Hallucinate across Languages? On Multilingual Estimation of LLM Hallucination in the Wild | Feb 18, 2025 | ArticlesHallucination | CodeCode Available | 0 |
| Unveiling the Magic of Code Reasoning through Hypothesis Decomposition and Amendment | Feb 17, 2025 | HallucinationLogical Reasoning | CodeCode Available | 2 |
| Can Your Uncertainty Scores Detect Hallucinated Entity? | Feb 17, 2025 | HallucinationSentence | —Unverified | 0 |
| Smoothing Out Hallucinations: Mitigating LLM Hallucination with Smoothed Knowledge Distillation | Feb 16, 2025 | HallucinationKnowledge Distillation | —Unverified | 0 |
| Valuable Hallucinations: Realizable Non-realistic Propositions | Feb 16, 2025 | Hallucination | —Unverified | 0 |
| A Survey of LLM-based Agents in Medicine: How far are we from Baymax? | Feb 16, 2025 | HallucinationSurvey | —Unverified | 0 |
| Automated Hypothesis Validation with Agentic Sequential Falsifications | Feb 14, 2025 | Decision MakingHallucination | CodeCode Available | 3 |
| Enhancing RAG with Active Learning on Conversation Records: Reject Incapables and Answer Capables | Feb 13, 2025 | Active LearningHallucination | —Unverified | 0 |
| DeepSeek on a Trip: Inducing Targeted Visual Hallucinations via Representation Vulnerabilities | Feb 11, 2025 | HallucinationSSIM | —Unverified | 0 |
| Elevating Legal LLM Responses: Harnessing Trainable Logical Structures and Semantic Knowledge with Legal Reasoning | Feb 11, 2025 | HallucinationIn-Context Learning | CodeCode Available | 0 |
| Hallucination, Monofacts, and Miscalibration: An Empirical Investigation | Feb 11, 2025 | DecoderHallucination | CodeCode Available | 0 |
| Refine Knowledge of Large Language Models via Adaptive Contrastive Learning | Feb 11, 2025 | Contrastive LearningHallucination | —Unverified | 0 |
| Hallucination Detection: A Probabilistic Framework Using Embeddings Distance Analysis | Feb 10, 2025 | Hallucination | —Unverified | 0 |
| Knowledge Graph-Guided Retrieval Augmented Generation | Feb 8, 2025 | DiversityHallucination | CodeCode Available | 2 |
| Learning Conformal Abstention Policies for Adaptive Risk Management in Large Language and Vision-Language Models | Feb 8, 2025 | Conformal PredictionDecision Making | CodeCode Available | 0 |
| Self-Rationalization in the Wild: A Large Scale Out-of-Distribution Evaluation on NLI-related tasks | Feb 7, 2025 | Abstractive Text SummarizationExplanation Generation | CodeCode Available | 0 |
| VideoRoPE: What Makes for Good Video Rotary Position Embedding? | Feb 7, 2025 | HallucinationPosition | CodeCode Available | 3 |
| ChallengeMe: An Adversarial Learning-enabled Text Summarization Framework | Feb 7, 2025 | HallucinationSpecificity | —Unverified | 0 |
| Linear Correlation in LM's Compositional Generalization and Hallucination | Feb 6, 2025 | Hallucination | CodeCode Available | 0 |
| TruthFlow: Truthful LLM Generation via Representation Flow Correction | Feb 6, 2025 | HallucinationTruthfulQA | —Unverified | 0 |
| Large Language Models for Multi-Robot Systems: A Survey | Feb 6, 2025 | Action GenerationBenchmarking | CodeCode Available | 1 |
| Enhancing Hallucination Detection through Noise Injection | Feb 6, 2025 | Hallucination | —Unverified | 0 |
| The Hidden Life of Tokens: Reducing Hallucination of Large Vision-Language Models via Visual Information Steering | Feb 5, 2025 | Hallucination | CodeCode Available | 2 |
| A Schema-Guided Reason-while-Retrieve framework for Reasoning on Scene Graphs with Large-Language-Models (LLMs) | Feb 5, 2025 | HallucinationSpatial Reasoning | —Unverified | 0 |
| DAMO: Data- and Model-aware Alignment of Multi-modal LLMs | Feb 4, 2025 | Hallucination | CodeCode Available | 1 |
| Mitigating Object Hallucinations in Large Vision-Language Models via Attention Calibration | Feb 4, 2025 | AttributeHallucination | —Unverified | 0 |
| Eliciting Language Model Behaviors with Investigator Agents | Feb 3, 2025 | Bayesian InferenceHallucination | —Unverified | 0 |
| SelfCheckAgent: Zero-Resource Hallucination Detection in Generative Large Language Models | Feb 3, 2025 | Hallucination | —Unverified | 0 |
| MJ-VIDEO: Fine-Grained Benchmarking and Rewarding Video Preferences in Video Generation | Feb 3, 2025 | BenchmarkingFairness | —Unverified | 0 |