SOTAVerified

Hallucination

Papers

Showing 351400 of 1816 papers

TitleStatusHype
Evaluation and Analysis of Hallucination in Large Vision-Language ModelsCode1
AssistRAG: Boosting the Potential of Large Language Models with an Intelligent Information AssistantCode1
Pensieve: Retrospect-then-Compare Mitigates Visual HallucinationCode1
Enhancing Text-based Knowledge Graph Completion with Zero-Shot Large Language Models: A Focus on Semantic EnhancementCode1
CHATREPORT: Democratizing Sustainability Disclosure Analysis through LLM-based ToolsCode1
Hallucinated Neural Radiance Fields in the WildCode1
Entity-Based Knowledge Conflicts in Question AnsweringCode1
Knowledge Verification to Nip Hallucination in the BudCode1
CriSPO: Multi-Aspect Critique-Suggestion-guided Automatic Prompt Optimization for Text GenerationCode1
Enhancing Semantics in Multimodal Chain of Thought via Soft Negative SamplingCode1
Enhancing Uncertainty-Based Hallucination Detection with Stronger FocusCode1
CR-LT-KGQA: A Knowledge Graph Question Answering Dataset Requiring Commonsense Reasoning and Long-Tail KnowledgeCode1
Entity-level Factual Consistency of Abstractive Text SummarizationCode1
ChartSumm: A Comprehensive Benchmark for Automatic Chart Summarization of Long and Short SummariesCode1
ChartInsighter: An Approach for Mitigating Hallucination in Time-series Chart Summary Generation with A Benchmark DatasetCode1
ProDehaze: Prompting Diffusion Models Toward Faithful Image DehazingCode1
EventHallusion: Diagnosing Event Hallucinations in Video LLMsCode1
MJ-Bench: Is Your Multimodal Reward Model Really a Good Judge for Text-to-Image Generation?Code1
Multimodal LLMs Can Reason about Aesthetics in Zero-ShotCode1
Mitigating Hallucination for Large Vision Language Model by Inter-Modality Correlation Calibration DecodingCode1
EmbodiedAgent: A Scalable Hierarchical Approach to Overcome Practical Challenge in Multi-Robot ControlCode1
CuriousLLM: Elevating Multi-Document QA with Reasoning-Infused Knowledge Graph PromptingCode1
Mitigating Fine-Grained Hallucination by Fine-Tuning Large Vision-Language Models with Caption RewritesCode1
Mitigating Hallucinations in Large Vision-Language Models via Summary-Guided DecodingCode1
EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language ModelsCode1
Efficient Dynamic Clustering-Based Document Compression for Retrieval-Augmented-GenerationCode1
Element-aware Summarization with Large Language Models: Expert-aligned Evaluation and Chain-of-Thought MethodCode1
Automatic Curriculum Expert Iteration for Reliable LLM ReasoningCode1
MedVH: Towards Systematic Evaluation of Hallucination for Large Vision Language Models in the Medical ContextCode1
MemLLM: Finetuning LLMs to Use An Explicit Read-Write MemoryCode1
EDFace-Celeb-1M: Benchmarking Face Hallucination with a Million-scale DatasetCode1
Dataset Distillation via FactorizationCode1
MedChat: A Multi-Agent Framework for Multimodal Diagnosis with Large Language ModelsCode1
ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition BenchmarkCode1
High-resolution Face Swapping via Latent Semantics DisentanglementCode1
Lyra: Orchestrating Dual Correction in Automated Theorem ProvingCode1
Holistic Analysis of Hallucination in GPT-4V(ision): Bias and Interference ChallengesCode1
How Language Model Hallucinations Can SnowballCode1
Med-HALT: Medical Domain Hallucination Test for Large Language ModelsCode1
Chain of Natural Language Inference for Reducing Large Language Model Ungrounded HallucinationsCode1
DEEM: Diffusion Models Serve as the Eyes of Large Language Models for Image PerceptionCode1
BachGAN: High-Resolution Image Synthesis from Salient Object LayoutCode1
Balanced Classification: A Unified Framework for Long-Tailed Object DetectionCode1
DomainRAG: A Chinese Benchmark for Evaluating Domain-specific Retrieval-Augmented GenerationCode1
Chain-of-Knowledge: Grounding Large Language Models via Dynamic Knowledge Adapting over Heterogeneous SourcesCode1
A Head to Predict and a Head to Question: Pre-trained Uncertainty Quantification Heads for Hallucination Detection in LLM OutputsCode1
A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and InteractivityCode1
Robust 3D Object Detection from LiDAR-Radar Point Clouds via Cross-Modal Feature AugmentationCode1
Enhancing LLM's Cognition via StructurizationCode1
Mitigating Hallucinations in Large Vision-Language Models by Adaptively Constraining Information FlowCode1
Show:102550
← PrevPage 8 of 37Next →

No leaderboard results yet.