SOTAVerified

Hallucination

Papers

Showing 201250 of 1816 papers

TitleStatusHype
Mitigating Hallucinations in Large Vision-Language Models by Adaptively Constraining Information FlowCode1
ProAPO: Progressively Automatic Prompt Optimization for Visual ClassificationCode1
LLM-QE: Improving Query Expansion by Aligning Large Language Models with Ranking PreferencesCode1
Hallucination Detection in LLMs Using Spectral Features of Attention MapsCode1
R2-KG: General-Purpose Dual-Agent Framework for Reliable Reasoning on Knowledge GraphsCode1
Large Language Models for Multi-Robot Systems: A SurveyCode1
DAMO: Data- and Model-aware Alignment of Multi-modal LLMsCode1
PAINT: Paying Attention to INformed Tokens to Mitigate Hallucination in Large Vision-Language ModelCode1
ChartInsighter: An Approach for Mitigating Hallucination in Time-series Chart Summary Generation with A Benchmark DatasetCode1
Knowledge Graph-based Retrieval-Augmented Generation for Schema MatchingCode1
Multimodal LLMs Can Reason about Aesthetics in Zero-ShotCode1
VASparse: Towards Efficient Visual Hallucination Mitigation for Large Vision-Language Model via Visual-Aware SparsificationCode1
ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition BenchmarkCode1
Mitigating Hallucination for Large Vision Language Model by Inter-Modality Correlation Calibration DecodingCode1
Octopus: Alleviating Hallucination via Dynamic Contrastive DecodingCode1
VASparse: Towards Efficient Visual Hallucination Mitigation via Visual-Aware Token SparsificationCode1
Harnessing Large Language Models for Knowledge Graph Question Answering via Adaptive Multi-Aspect Retrieval-AugmentationCode1
Extract Free Dense Misalignment from CLIPCode1
Filter-then-Generate: Large Language Models with Structure-Text Adapter for Knowledge Graph CompletionCode1
Can LLMs be Good Graph Judge for Knowledge Graph Construction?Code1
AtomR: Atomic Operator-Empowered Large Language Models for Heterogeneous Knowledge ReasoningCode1
VidHal: Benchmarking Temporal Hallucinations in Vision LLMsCode1
VaLiD: Mitigating the Hallucination of Large Vision Language Models by Visual Layer Fusion Contrastive DecodingCode1
Thinking Before Looking: Improving Multimodal LLM Reasoning via Mitigating Visual HallucinationCode1
AssistRAG: Boosting the Potential of Large Language Models with an Intelligent Information AssistantCode1
Distinguishing Ignorance from Error in LLM HallucinationsCode1
Can Knowledge Editing Really Correct Hallucinations?Code1
Paths-over-Graph: Knowledge Graph Empowered Large Language Model ReasoningCode1
Mitigating Hallucinations in Large Vision-Language Models via Summary-Guided DecodingCode1
FaithBench: A Diverse Hallucination Benchmark for Summarization by Modern LLMsCode1
Search Engines in an AI Era: The False Promise of Factual and Verifiable Source-Cited ResponsesCode1
VERIFIED: A Video Corpus Moment Retrieval Benchmark for Fine-Grained Video UnderstandingCode1
Automatic Curriculum Expert Iteration for Reliable LLM ReasoningCode1
OneNet: A Fine-Tuning Free Framework for Few-Shot Entity Linking via Large Language Model PromptingCode1
IterGen: Iterative Semantic-aware Structured LLM Generation with BacktrackingCode1
CriSPO: Multi-Aspect Critique-Suggestion-guided Automatic Prompt Optimization for Text GenerationCode1
FactAlign: Long-form Factuality Alignment of Large Language ModelsCode1
EventHallusion: Diagnosing Event Hallucinations in Video LLMsCode1
XTRUST: On the Multilingual Trustworthiness of Large Language ModelsCode1
FAIR GPT: A virtual consultant for research data management in ChatGPTCode1
Evaluating Image Hallucination in Text-to-Image Generation with Question-AnsweringCode1
Trustworthiness in Retrieval-Augmented Generation Systems: A SurveyCode1
Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path ReasoningCode1
Towards Empathetic Conversational Recommender SystemsCode1
ConVis: Contrastive Decoding with Hallucination Visualization for Mitigating Hallucinations in Multimodal Large Language ModelsCode1
SLM Meets LLM: Balancing Latency, Interpretability and Consistency in Hallucination DetectionCode1
Reefknot: A Comprehensive Benchmark for Relation Hallucination Evaluation, Analysis and Mitigation in Multimodal Large Language ModelsCode1
Hallu-PI: Evaluating Hallucination in Multi-modal Large Language Models within Perturbed InputsCode1
Mitigating Multilingual Hallucination in Large Vision-Language ModelsCode1
Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMsCode1
Show:102550
← PrevPage 5 of 37Next →

No leaderboard results yet.