SOTAVerified

Hallucination

Papers

Showing 901925 of 1816 papers

TitleStatusHype
VideoHallucer: Evaluating Intrinsic and Extrinsic Hallucinations in Large Video-Language Models0
Evaluating and Analyzing Relationship Hallucinations in Large Vision-Language ModelsCode1
Prompt-Consistency Image Generation (PCIG): A Unified Framework Integrating LLMs, Knowledge Graphs, and Controllable Diffusion ModelsCode0
Semantic Entropy Probes: Robust and Cheap Hallucination Detection in LLMsCode2
Evaluating RAG-Fusion with RAGElo: an Automated Elo-based FrameworkCode2
Does Object Grounding Really Reduce Hallucination of Large Vision-Language Models?0
From Descriptive Richness to Bias: Unveiling the Dark Side of Generative Image Caption Enrichment0
HIGHT: Hierarchical Graph Tokenization for Molecule-Language Alignment0
Large Language Models are Skeptics: False Negative Problem of Input-conflicting Hallucination0
Rethinking Abdominal Organ Segmentation (RAOS) in the clinical scenario: A robustness evaluation benchmark with challenging casesCode2
Knowledge Graph-Enhanced Large Language Models via Path SelectionCode1
StackRAG Agent: Improving Developer Answers with Retrieval-Augmented GenerationCode0
Detecting Errors through Ensembling Prompts (DEEP): An End-to-End LLM Framework for Detecting Factual ErrorsCode0
Fast and Slow Generating: An Empirical Study on Large and Small Language Models Collaborative DecodingCode1
RichRAG: Crafting Rich Responses for Multi-faceted Queries in Retrieval-Augmented Generation0
What Matters in Memorizing and Recalling Facts? Multifaceted Benchmarks for Knowledge Probing in Language Models0
On-Policy Fine-grained Knowledge Feedback for Hallucination MitigationCode0
Do More Details Always Introduce More Hallucinations in LVLM-based Image Captioning?0
Beyond Under-Alignment: Atomic Preference Enhanced Factuality Tuning for Large Language Models0
InternalInspector I^2: Robust Confidence Estimation in LLMs through Internal States0
Self-training Large Language Models through Knowledge DetectionCode0
Small Agent Can Also Rock! Empowering Small Language Models as Hallucination DetectorCode1
Mitigating Large Language Model Hallucination with Faithful Finetuning0
Counterfactual Debating with Preset Stances for Hallucination Elimination of LLMsCode0
Hallucination Mitigation Prompts Long-term Video UnderstandingCode0
Show:102550
← PrevPage 37 of 73Next →

No leaderboard results yet.