SOTAVerified

Hallucination Evaluation

Evaluate the ability of LLM to generate non-hallucination text or assess the capability of LLM to recognize hallucinations.

Papers

Showing 1120 of 49 papers

TitleStatusHype
Enhancing LLM's Cognition via StructurizationCode1
PhD: A ChatGPT-Prompted Visual hallucination Evaluation DatasetCode1
DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language ModelsCode1
Alleviating Hallucinations of Large Language Models through Induced HallucinationsCode1
Mitigating Fine-Grained Hallucination by Fine-Tuning Large Vision-Language Models with Caption RewritesCode1
UHGEval: Benchmarking the Hallucination of Chinese Large Language Models via Unconstrained GenerationCode1
HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction DataCode1
Investigating Hallucinations in Pruned Large Language Models for Abstractive SummarizationCode1
AMBER: An LLM-free Multi-dimensional Benchmark for MLLMs Hallucination EvaluationCode1
Analyzing and Mitigating Object Hallucination in Large Vision-Language ModelsCode1
Show:102550
← PrevPage 2 of 5Next →

No leaderboard results yet.