SOTAVerified

Hallucination

Papers

Showing 176200 of 1816 papers

TitleStatusHype
How well can a large language model explain business processes as perceived by users?Code1
AMBER: An LLM-free Multi-dimensional Benchmark for MLLMs Hallucination EvaluationCode1
Advancing TTP Analysis: Harnessing the Power of Large Language Models with Retrieval Augmented GenerationCode1
3D Sketch-aware Semantic Scene Completion via Semi-supervised Structure PriorCode1
HaloQuest: A Visual Hallucination Dataset for Advancing Multimodal ReasoningCode1
Hallucination-Aware Multimodal Benchmark for Gastrointestinal Image Analysis with Large Vision-Language ModelsCode1
DomainRAG: A Chinese Benchmark for Evaluating Domain-specific Retrieval-Augmented GenerationCode1
Hallucination Detection in LLMs Using Spectral Features of Attention MapsCode1
ADeLA: Automatic Dense Labeling with Attention for Viewpoint Adaptation in Semantic SegmentationCode1
An Empirical Study on Parameter-Efficient Fine-Tuning for MultiModal Large Language ModelsCode1
Distinguishing Ignorance from Error in LLM HallucinationsCode1
DiffFuSR: Super-Resolution of all Sentinel-2 Multispectral Bands using Diffusion ModelsCode1
HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction DataCode1
DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language ModelsCode1
Analyzing LLMs' Knowledge Boundary Cognition Across Languages Through the Lens of Internal RepresentationsCode1
A Data-Centric Approach To Generate Faithful and High Quality Patient Summaries with Large Language ModelsCode1
Analyzing and Mitigating Object Hallucination in Large Vision-Language ModelsCode1
HallE-Control: Controlling Object Hallucination in Large Multimodal ModelsCode1
Hallucinated Neural Radiance Fields in the WildCode1
Detecting Machine-Generated Texts by Multi-Population Aware Optimization for Maximum Mean DiscrepancyCode1
Detecting Hallucinated Content in Conditional Neural Sequence GenerationCode1
Phare: A Safety Probe for Large Language ModelsCode1
Detecting and Preventing Hallucinations in Large Vision Language ModelsCode1
Grounded Chain-of-Thought for Multimodal Large Language ModelsCode1
A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and InteractivityCode1
Show:102550
← PrevPage 8 of 73Next →

No leaderboard results yet.