SOTAVerified

Hallucination Evaluation

Evaluate the ability of LLM to generate non-hallucination text or assess the capability of LLM to recognize hallucinations.

Papers

Showing 1120 of 49 papers

TitleStatusHype
Alleviating Hallucinations of Large Language Models through Induced HallucinationsCode1
Investigating Hallucinations in Pruned Large Language Models for Abstractive SummarizationCode1
Benchmarking LLM Faithfulness in RAG with Evolving LeaderboardsCode1
DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language ModelsCode1
Exploring Hallucination of Large Multimodal Models in Video Understanding: Benchmark, Analysis and MitigationCode1
Analyzing and Mitigating Object Hallucination in Large Vision-Language ModelsCode1
Enhancing LLM's Cognition via StructurizationCode1
Evaluating Image Hallucination in Text-to-Image Generation with Question-AnsweringCode1
Evaluation and Analysis of Hallucination in Large Vision-Language ModelsCode1
KnowRL: Exploring Knowledgeable Reinforcement Learning for FactualityCode1
Show:102550
← PrevPage 2 of 5Next →

No leaderboard results yet.