SOTAVerified

Hallucination Evaluation

Evaluate the ability of LLM to generate non-hallucination text or assess the capability of LLM to recognize hallucinations.

Papers

Showing 3140 of 49 papers

TitleStatusHype
Unified Triplet-Level Hallucination Evaluation for Large Vision-Language ModelsCode0
Lynx: An Open Source Hallucination Evaluation Model0
A Survey of Hallucination in Large Visual Language Models0
FIHA: Autonomous Hallucination Evaluation in Vision-Language Models with Davidson Scene Graphs0
Mitigating Hallucination in Multimodal Large Language Model via Hallucination-targeted Direct Preference Optimization0
CHARP: Conversation History AwaReness Probing for Knowledge-grounded Dialogue Systems0
Mitigating Image Captioning Hallucinations in Vision-Language Models0
Do Androids Know They're Only Dreaming of Electric Sheep?0
Real-Time Evaluation Models for RAG: Who Detects Hallucinations Best?0
Exploring and Evaluating Hallucinations in LLM-Powered Code Generation0
Show:102550
← PrevPage 4 of 5Next →

No leaderboard results yet.