SOTAVerified

Hallucination

Papers

Showing 651700 of 1816 papers

TitleStatusHype
ReDeEP: Detecting Hallucination in Retrieval-Augmented Generation via Mechanistic Interpretability0
LargePiG: Your Large Language Model is Secretly a Pointer Generator0
Automatically Generating Visual Hallucination Test Cases for Multimodal Large Language ModelsCode0
Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs0
On the Capacity of Citation Generation by Large Language Models0
Magnifier Prompt: Tackling Multimodal Hallucination via Extremely Simple Instructions0
MLLM can see? Dynamic Correction Decoding for Hallucination MitigationCode2
AGENTiGraph: An Interactive Knowledge Graph Platform for LLM-based Chatbots Utilizing Private Data0
Can Structured Data Reduce Epistemic Uncertainty?0
Parenting: Optimizing Knowledge Selection of Retrieval-Augmented Language Models with Parameter Decoupling and Tailored Tuning0
Medico: Towards Hallucination Detection and Correction with Multi-source Evidence Fusion0
SkillAggregation: Reference-free LLM-Dependent Aggregation0
VideoAgent: Self-Improving Video GenerationCode2
LongHalQA: Long-Context Hallucination Evaluation for MultiModal Large Language ModelsCode0
Honest AI: Fine-Tuning "Small" Language Models to Say "I Don't Know", and Reducing Hallucination in RAG0
Collu-Bench: A Benchmark for Predicting Language Model Hallucinations in Code0
VLFeedback: A Large-Scale AI Feedback Dataset for Large Vision-Language Models Alignment0
Measuring the Inconsistency of Large Language Models in Preferential Ranking0
A Methodology for Evaluating RAG Systems: A Case Study On Configuration Dependency ValidationCode0
VERIFIED: A Video Corpus Moment Retrieval Benchmark for Fine-Grained Video UnderstandingCode1
PublicHearingBR: A Brazilian Portuguese Dataset of Public Hearing Transcripts for Summarization of Long Documents0
Can Knowledge Graphs Make Large Language Models More Trustworthy? An Empirical Study over Open-ended Question Answering0
OneNet: A Fine-Tuning Free Framework for Few-Shot Entity Linking via Large Language Model PromptingCode1
Automatic Curriculum Expert Iteration for Reliable LLM ReasoningCode1
LatteCLIP: Unsupervised CLIP Fine-Tuning via LMM-Synthetic Texts0
IterGen: Iterative Semantic-aware Structured LLM Generation with BacktrackingCode1
From Pixels to Tokens: Revisiting Object Hallucinations in Large Vision-Language Models0
Utilize the Flow before Stepping into the Same River Twice: Certainty Represented Knowledge Flow for Refusal-Aware Instruction TuningCode0
Embodied Agent Interface: Benchmarking LLMs for Embodied Decision MakingCode3
EMMA: Empowering Multi-modal Mamba with Structural and Hierarchical Alignment0
ReFIR: Grounding Large Restoration Models with Retrieval AugmentationCode2
Listening to Patients: A Framework of Detecting and Mitigating Patient Misreport for Medical Dialogue Generation0
Gradual Learning: Optimizing Fine-Tuning with Partially Mastered Knowledge in Large Language Models0
FG-PRM: Fine-grained Hallucination Detection and Mitigation in Language Model Mathematical Reasoning0
Differential TransformerCode2
TLDR: Token-Level Detective Reward Model for Large Vision Language Models0
AI-Enhanced Ethical Hacking: A Linux-Focused Experiment0
Mitigating Modality Prior-Induced Hallucinations in Multimodal Large Language Models via Deciphering Attention CausalityCode2
Mitigating Hallucinations Using Ensemble of Knowledge Graph and Vector Store in Large Language Models to Enhance Mental Health Support0
DAMRO: Dive into the Attention Mechanism of LVLM to Reduce Object Hallucination0
DiDOTS: Knowledge Distillation from Large-Language-Models for Dementia Obfuscation in Transcribed Speech0
TUBench: Benchmarking Large Vision-Language Models on Trustworthiness with Unanswerable QuestionsCode0
SAG: Style-Aligned Article Generation via Model Collaboration0
Auto-GDA: Automatic Domain Adaptation for Efficient Grounding Verification in Retrieval Augmented Generation0
Look Twice Before You Answer: Memory-Space Visual Retracing for Hallucination Mitigation in Multimodal Large Language ModelsCode2
Investigating and Mitigating Object Hallucinations in Pretrained Vision-Language (CLIP) ModelsCode0
FactCheckmate: Preemptively Detecting and Mitigating Hallucinations in LMs0
Characterizing Context Influence and Hallucination in SummarizationCode0
CriSPO: Multi-Aspect Critique-Suggestion-guided Automatic Prompt Optimization for Text GenerationCode1
Salient Information Prompting to Steer Content in Prompt-based Abstractive SummarizationCode0
Show:102550
← PrevPage 14 of 37Next →

No leaderboard results yet.