SOTAVerified

Hallucination

Papers

Showing 876900 of 1816 papers

TitleStatusHype
Reducing Tool Hallucination via Reliability Alignment0
GenMAC: Compositional Text-to-Video Generation with Multi-Agent Collaboration0
VidHalluc: Evaluating Temporal Hallucinations in Multimodal Large Language Models for Video Understanding0
Who Brings the Frisbee: Probing Hidden Hallucination Factors in Large Vision-Language Model via Causality Analysis0
An Evolutionary Large Language Model for Hallucination Mitigation0
CC-OCR: A Comprehensive and Challenging OCR Benchmark for Evaluating Large Multimodal Models in Literacy0
AI Benchmarks and Datasets for LLM Evaluation0
Automating Feedback Analysis in Surgical Training: Detection, Categorization, and AssessmentCode0
Beyond Logit Lens: Contextual Embeddings for Robust Hallucination Detection & Grounding in VLMs0
DHCP: Detecting Hallucinations by Cross-modal Attention Pattern in Large Vision-Language Models0
OPCap:Object-aware Prompting Captioning0
Efficient Self-Improvement in Multimodal Large Language Models: A Model-Level Judge-Free Approach0
Meaningless is better: hashing bias-inducing words in LLM prompts improves performance in logical reasoning and statistical learning0
A Topic-level Self-Correctional Approach to Mitigate Hallucinations in MLLMs0
VLRewardBench: A Challenging Benchmark for Vision-Language Generative Reward Models0
AI2T: Building Trustable AI Tutors by Interactively Teaching a Self-Aware Learning Agent0
Enhancing Multi-Agent Consensus through Third-Party LLM Integration: Analyzing Uncertainty and Mitigating Hallucinations in Large Language Models0
Ontology-Constrained Generation of Domain-Specific Clinical SummariesCode0
Leveraging LLMs for Legacy Code Modernization: Challenges and Opportunities for LLM-Generated Documentation0
Detecting Hallucinations in Virtual Histology with Neural Precursors0
ICT: Image-Object Cross-Level Trusted Intervention for Mitigating Object Hallucination in Large Vision-Language Models0
Sycophancy in Large Language Models: Causes and Mitigations0
CATCH: Complementary Adaptive Token-level Contrastive Decoding to Mitigate Hallucinations in LVLMs0
Can Open-source LLMs Enhance Data Synthesis for Toxic Detection?: An Experimental Study0
Mitigating Knowledge Conflicts in Language Model-Driven Question Answering0
Show:102550
← PrevPage 36 of 73Next →

No leaderboard results yet.