SOTAVerified

Hallucination

Papers

Showing 151200 of 1816 papers

TitleStatusHype
In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination MitigationCode2
InstructGraph: Boosting Large Language Models via Graph-centric Instruction Tuning and Preference AlignmentCode2
DeliLaw: A Chinese Legal Counselling System Based on a Large Language ModelCode2
A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open QuestionsCode2
Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial ReasoningCode2
MQAG: Multiple-choice Question Answering and Generation for Assessing Information Consistency in SummarizationCode2
Multimodal Needle in a Haystack: Benchmarking Long-Context Capability of Multimodal Large Language ModelsCode2
Granite GuardianCode2
Image Textualization: An Automatic Framework for Creating Accurate and Detailed Image DescriptionsCode2
KnowHalu: Hallucination Detection via Multi-Form Knowledge Based Factual CheckingCode2
Look Twice Before You Answer: Memory-Space Visual Retracing for Hallucination Mitigation in Multimodal Large Language ModelsCode2
Unveiling the Magic of Code Reasoning through Hypothesis Decomposition and AmendmentCode2
FaithEval: Can Your Language Model Stay Faithful to Context, Even If "The Moon is Made of Marshmallows"Code2
High-resolution Face Swapping via Latent Semantics DisentanglementCode1
Holistic Analysis of Hallucination in GPT-4V(ision): Bias and Interference ChallengesCode1
EDFace-Celeb-1M: Benchmarking Face Hallucination with a Million-scale DatasetCode1
Harnessing GPT-4V(ision) for Insurance: A Preliminary ExplorationCode1
Harnessing Large Language Models for Knowledge Graph Question Answering via Adaptive Multi-Aspect Retrieval-AugmentationCode1
How Language Model Hallucinations Can SnowballCode1
Antidote: A Unified Framework for Mitigating LVLM Hallucinations in Counterfactual Presupposition and Object PerceptionCode1
Adversarial Feature Hallucination Networks for Few-Shot LearningCode1
ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition BenchmarkCode1
Efficient Dynamic Clustering-Based Document Compression for Retrieval-Augmented-GenerationCode1
How well can a large language model explain business processes as perceived by users?Code1
AMBER: An LLM-free Multi-dimensional Benchmark for MLLMs Hallucination EvaluationCode1
Advancing TTP Analysis: Harnessing the Power of Large Language Models with Retrieval Augmented GenerationCode1
3D Sketch-aware Semantic Scene Completion via Semi-supervised Structure PriorCode1
DomainRAG: A Chinese Benchmark for Evaluating Domain-specific Retrieval-Augmented GenerationCode1
Hallucination-Aware Multimodal Benchmark for Gastrointestinal Image Analysis with Large Vision-Language ModelsCode1
EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language ModelsCode1
Hallucination Detection in LLMs Using Spectral Features of Attention MapsCode1
Doc2Query--: When Less is MoreCode1
ADeLA: Automatic Dense Labeling with Attention for Viewpoint Adaptation in Semantic SegmentationCode1
HaloQuest: A Visual Hallucination Dataset for Advancing Multimodal ReasoningCode1
DiffFuSR: Super-Resolution of all Sentinel-2 Multispectral Bands using Diffusion ModelsCode1
An Empirical Study on Parameter-Efficient Fine-Tuning for MultiModal Large Language ModelsCode1
HallE-Control: Controlling Object Hallucination in Large Multimodal ModelsCode1
HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction DataCode1
Analyzing LLMs' Knowledge Boundary Cognition Across Languages Through the Lens of Internal RepresentationsCode1
DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language ModelsCode1
A Data-Centric Approach To Generate Faithful and High Quality Patient Summaries with Large Language ModelsCode1
Analyzing and Mitigating Object Hallucination in Large Vision-Language ModelsCode1
Detecting Machine-Generated Texts by Multi-Population Aware Optimization for Maximum Mean DiscrepancyCode1
Hallucinated Neural Radiance Fields in the WildCode1
Detecting and Preventing Hallucinations in Large Vision Language ModelsCode1
Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI FeedbackCode1
Grounded Chain-of-Thought for Multimodal Large Language ModelsCode1
Phare: A Safety Probe for Large Language ModelsCode1
A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and InteractivityCode1
BAMBOO: A Comprehensive Benchmark for Evaluating Long Text Modeling Capacities of Large Language ModelsCode1
Show:102550
← PrevPage 4 of 37Next →

No leaderboard results yet.