SOTAVerified

Hallucination Evaluation

Evaluate the ability of LLM to generate non-hallucination text or assess the capability of LLM to recognize hallucinations.

Papers

Showing 4149 of 49 papers

TitleStatusHype
HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction DataCode1
Investigating Hallucinations in Pruned Large Language Models for Abstractive SummarizationCode1
AMBER: An LLM-free Multi-dimensional Benchmark for MLLMs Hallucination EvaluationCode1
HallusionBench: An Advanced Diagnostic Suite for Entangled Language Hallucination and Visual Illusion in Large Vision-Language ModelsCode2
ReEval: Automatic Hallucination Evaluation for Retrieval-Augmented Large Language Models via Transferable Adversarial Attacks0
Analyzing and Mitigating Object Hallucination in Large Vision-Language ModelsCode1
Evaluation and Analysis of Hallucination in Large Vision-Language ModelsCode1
MindMap: Knowledge Graph Prompting Sparks Graph of Thoughts in Large Language ModelsCode2
HaluEval: A Large-Scale Hallucination Evaluation Benchmark for Large Language ModelsCode2
Show:102550
← PrevPage 5 of 5Next →

No leaderboard results yet.