SOTAVerified

Hallucination Evaluation

Evaluate the ability of LLM to generate non-hallucination text or assess the capability of LLM to recognize hallucinations.

Papers

Showing 1120 of 49 papers

TitleStatusHype
Alleviating Hallucinations of Large Language Models through Induced HallucinationsCode1
Exploring Hallucination of Large Multimodal Models in Video Understanding: Benchmark, Analysis and MitigationCode1
Investigating Hallucinations in Pruned Large Language Models for Abstractive SummarizationCode1
DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language ModelsCode1
Enhancing LLM's Cognition via StructurizationCode1
Benchmarking LLM Faithfulness in RAG with Evolving LeaderboardsCode1
HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction DataCode1
Evaluating Image Hallucination in Text-to-Image Generation with Question-AnsweringCode1
Evaluation and Analysis of Hallucination in Large Vision-Language ModelsCode1
Analyzing and Mitigating Object Hallucination in Large Vision-Language ModelsCode1
Show:102550
← PrevPage 2 of 5Next →

No leaderboard results yet.