SOTAVerified

Hallucination

Papers

Showing 201250 of 1816 papers

TitleStatusHype
Analyzing and Mitigating Object Hallucination in Large Vision-Language ModelsCode1
EmbodiedAgent: A Scalable Hierarchical Approach to Overcome Practical Challenge in Multi-Robot ControlCode1
Label Hallucination for Few-Shot ClassificationCode1
EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language ModelsCode1
Element-aware Summarization with Large Language Models: Expert-aligned Evaluation and Chain-of-Thought MethodCode1
Efficient Dynamic Clustering-Based Document Compression for Retrieval-Augmented-GenerationCode1
"Kelly is a Warm Person, Joseph is a Role Model": Gender Biases in LLM-Generated Reference LettersCode1
Joint Evaluation of Answer and Reasoning Consistency for Hallucination Detection in Large Reasoning ModelsCode1
Benchmarking LLM Faithfulness in RAG with Evolving LeaderboardsCode1
Phare: A Safety Probe for Large Language ModelsCode1
JDocQA: Japanese Document Question Answering Dataset for Generative Language ModelsCode1
KCTS: Knowledge-Constrained Tree Search Decoding with Token-Level Hallucination DetectionCode1
Knowledge Graph-based Retrieval-Augmented Generation for Schema MatchingCode1
Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference OptimizationCode1
A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and InteractivityCode1
Enhancing LLM's Cognition via StructurizationCode1
BAMBOO: A Comprehensive Benchmark for Evaluating Long Text Modeling Capacities of Large Language ModelsCode1
Enhancing Semantics in Multimodal Chain of Thought via Soft Negative SamplingCode1
An Empirical Study on Parameter-Efficient Fine-Tuning for MultiModal Large Language ModelsCode1
ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition BenchmarkCode1
BIGPrior: Towards Decoupling Learned Prior Hallucination and Data Fidelity in Image RestorationCode1
Entity-level Factual Consistency of Abstractive Text SummarizationCode1
Balanced Classification: A Unified Framework for Long-Tailed Object DetectionCode1
BachGAN: High-Resolution Image Synthesis from Salient Object LayoutCode1
EDFace-Celeb-1M: Benchmarking Face Hallucination with a Million-scale DatasetCode1
Is ChatGPT a Good Causal Reasoner? A Comprehensive EvaluationCode1
Evaluating and Analyzing Relationship Hallucinations in Large Vision-Language ModelsCode1
Exploring the Transferability of Visual Prompting for Multimodal Large Language ModelsCode1
Evaluation and Analysis of Hallucination in Large Vision-Language ModelsCode1
Accuracy and Political Bias of News Source Credibility Ratings by Large Language ModelsCode1
DomainRAG: A Chinese Benchmark for Evaluating Domain-specific Retrieval-Augmented GenerationCode1
Investigating and Mitigating the Multimodal Hallucination Snowballing in Large Vision-Language ModelsCode1
Bridging the Data Gap between Training and Inference for Unsupervised Neural Machine TranslationCode1
3D Sketch-aware Semantic Scene Completion via Semi-supervised Structure PriorCode1
Advancing TTP Analysis: Harnessing the Power of Large Language Models with Retrieval Augmented GenerationCode1
BTR: Binary Token Representations for Efficient Retrieval Augmented Language ModelsCode1
Invoke Interfaces Only When Needed: Adaptive Invocation for Large Language Models in Question AnsweringCode1
CAFe: Unifying Representation and Generation with Contrastive-Autoregressive FinetuningCode1
IterGen: Iterative Semantic-aware Structured LLM Generation with BacktrackingCode1
Knowledge Graph-Enhanced Large Language Models via Path SelectionCode1
FactAlign: Long-form Factuality Alignment of Large Language ModelsCode1
Face Hallucination via Split-Attention in Split-Attention NetworkCode1
LAN-HDR: Luminance-based Alignment Network for High Dynamic Range Video ReconstructionCode1
FAIR GPT: A virtual consultant for research data management in ChatGPTCode1
Antidote: A Unified Framework for Mitigating LVLM Hallucinations in Counterfactual Presupposition and Object PerceptionCode1
Can Knowledge Editing Really Correct Hallucinations?Code1
LLMs Know What They Need: Leveraging a Missing Information Guided Framework to Empower Retrieval-Augmented GenerationCode1
DiffFuSR: Super-Resolution of all Sentinel-2 Multispectral Bands using Diffusion ModelsCode1
InterrogateLLM: Zero-Resource Hallucination Detection in LLM-Generated AnswersCode1
Automatic Curriculum Expert Iteration for Reliable LLM ReasoningCode1
Show:102550
← PrevPage 5 of 37Next →

No leaderboard results yet.