SOTAVerified

Hallucination

Papers

Showing 526550 of 1816 papers

TitleStatusHype
Delve into Visual Contrastive Decoding for Hallucination Mitigation of Large Vision-Language ModelsCode0
Mechanistic Understanding and Mitigation of Language Model Non-Factual HallucinationsCode0
MCiteBench: A Multimodal Benchmark for Generating Text with CitationsCode0
Benchmarking Hallucination in Large Language Models based on Unanswerable Math Word ProblemCode0
MCQG-SRefine: Multiple Choice Question Generation and Evaluation with Iterative Self-Critique, Correction, and Comparison FeedbackCode0
MAF: Multi-Aspect Feedback for Improving Reasoning in Large Language ModelsCode0
MAVEN-Fact: A Large-scale Event Factuality Detection DatasetCode0
DefAn: Definitive Answer Dataset for LLMs Hallucination EvaluationCode0
Benchmarking ChatGPT-4 on ACR Radiation Oncology In-Training (TXIT) Exam and Red Journal Gray Zone Cases: Potentials and Challenges for AI-Assisted Medical Education and Decision Making in Radiation OncologyCode0
LVLM-Compress-Bench: Benchmarking the Broader Impact of Large Vision-Language Model CompressionCode0
Low to High Dimensional Modality Hallucination using Aggregated Fields of ViewCode0
Behind the Magic, MERLIM: Multi-modal Evaluation Benchmark for Large Image-Language ModelsCode0
Machine Translation Hallucination Detection for Low and High Resource Languages using Large Language ModelsCode0
MedScore: Factuality Evaluation of Free-Form Medical AnswersCode0
Logic Query of Thoughts: Guiding Large Language Models to Answer Complex Logic Queries with Knowledge GraphsCode0
LLMs and Memorization: On Quality and Specificity of Copyright ComplianceCode0
LongHalQA: Long-Context Hallucination Evaluation for MultiModal Large Language ModelsCode0
Deep CNN Denoiser and Multi-layer Neighbor Component Embedding for Face HallucinationCode0
EmotionHallucer: Evaluating Emotion Hallucinations in Multimodal Large Language ModelsCode0
LLM Hallucinations in Practical Code Generation: Phenomena, Mechanism, and MitigationCode0
DecoPrompt : Decoding Prompts Reduces Hallucinations when Large Language Models Meet False PremisesCode0
LLM Inference Enhanced by External Knowledge: A SurveyCode0
LLM-based Query Expansion Fails for Unfamiliar and Ambiguous QueriesCode0
LLM Internal States Reveal Hallucination Risk Faced With a QueryCode0
Deceptive Semantic Shortcuts on Reasoning Chains: How Far Can Models Go without Hallucination?Code0
Show:102550
← PrevPage 22 of 73Next →

No leaderboard results yet.