SOTAVerified

Hallucination

Papers

Showing 101150 of 1816 papers

TitleStatusHype
Mitigating Modality Prior-Induced Hallucinations in Multimodal Large Language Models via Deciphering Attention CausalityCode2
Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local AttentionCode2
MindMap: Knowledge Graph Prompting Sparks Graph of Thoughts in Large Language ModelsCode2
MeMemo: On-device Retrieval Augmentation for Private and Personalized Text GenerationCode2
Mitigating Hallucinations in Large Vision-Language Models via DPO: On-Policy Data Hold the KeyCode2
Mitigating Object Hallucination via Concentric Causal AttentionCode2
MQAG: Multiple-choice Question Answering and Generation for Assessing Information Consistency in SummarizationCode2
Look Twice Before You Answer: Memory-Space Visual Retracing for Hallucination Mitigation in Multimodal Large Language ModelsCode2
Reducing Hallucinations in Vision-Language Models via Latent Space SteeringCode2
DyFo: A Training-Free Dynamic Focus Visual Search for Enhancing LMMs in Fine-Grained Visual UnderstandingCode2
LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language ModelsCode2
Aligning Modalities in Vision Large Language Models via Preference Fine-tuningCode2
Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction TuningCode2
Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention MapsCode2
mDPO: Conditional Preference Optimization for Multimodal Large Language ModelsCode2
Controllable and Reliable Knowledge-Intensive Task-Oriented Conversational Agents with Declarative Genie WorksheetsCode2
Calibrated Self-Rewarding Vision Language ModelsCode2
LLaMP: Large Language Model Made Powerful for High-fidelity Materials Knowledge Retrieval and DistillationCode2
Self-Introspective Decoding: Alleviating Hallucinations for Large Vision-Language ModelsCode2
Exploring Human-Like Translation Strategy with Large Language ModelsCode2
Think-on-Graph: Deep and Responsible Reasoning of Large Language Model on Knowledge GraphCode2
TimeSuite: Improving MLLMs for Long Video Understanding via Grounded TuningCode2
Medical Hallucinations in Foundation Models and Their Impact on HealthcareCode2
Knowledge Graph-Guided Retrieval Augmented GenerationCode2
TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful SpaceCode2
Differential TransformerCode2
Devils in Middle Layers of Large Vision-Language Models: Interpreting, Detecting and Mitigating Object Hallucinations via Attention LensCode2
KnowHalu: Hallucination Detection via Multi-Form Knowledge Based Factual CheckingCode2
Lawyer LLaMA Technical ReportCode2
In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination MitigationCode2
Image Textualization: An Automatic Framework for Creating Accurate and Detailed Image DescriptionsCode2
InstructGraph: Boosting Large Language Models via Graph-centric Instruction Tuning and Preference AlignmentCode2
HaluEval: A Large-Scale Hallucination Evaluation Benchmark for Large Language ModelsCode2
MLAgentBench: Evaluating Language Agents on Machine Learning ExperimentationCode2
DiscoSG: Towards Discourse-Level Text Scene Graph Parsing through Iterative Graph RefinementCode2
Interleaving Retrieval with Chain-of-Thought Reasoning for Knowledge-Intensive Multi-Step QuestionsCode2
VHM: Versatile and Honest Vision Language Model for Remote Sensing Image AnalysisCode2
Granite GuardianCode2
GPT-NER: Named Entity Recognition via Large Language ModelsCode2
HALC: Object Hallucination Reduction via Adaptive Focal-Contrast DecodingCode2
Generate, but Verify: Reducing Hallucination in Vision-Language Models with Retrospective ResamplingCode2
Generate-on-Graph: Treat LLM as both Agent and KG in Incomplete Knowledge Graph Question AnsweringCode2
Automatically Correcting Large Language Models: Surveying the landscape of diverse self-correction strategiesCode2
FinMME: Benchmark Dataset for Financial Multi-Modal Reasoning EvaluationCode2
HalOmi: A Manually Annotated Benchmark for Multilingual Hallucination and Omission Detection in Machine TranslationCode2
FaithEval: Can Your Language Model Stay Faithful to Context, Even If "The Moon is Made of Marshmallows"Code2
Dynamic Parametric Retrieval Augmented Generation for Test-time Knowledge EnhancementCode2
Benchmarking Large Language Models in Retrieval-Augmented GenerationCode2
Fast Think-on-Graph: Wider, Deeper and Faster Reasoning of Large Language Model on Knowledge GraphCode2
FreshLLMs: Refreshing Large Language Models with Search Engine AugmentationCode2
Show:102550
← PrevPage 3 of 37Next →

No leaderboard results yet.