SOTAVerified

Hallucination

Papers

Showing 9761000 of 1816 papers

TitleStatusHype
Good Parenting is all you need -- Multi-agentic LLM Hallucination Mitigation0
MCQG-SRefine: Multiple Choice Question Generation and Evaluation with Iterative Self-Critique, Correction, and Comparison FeedbackCode0
ETF: An Entity Tracing Framework for Hallucination Detection in Code Summaries0
From Single to Multi: How LLMs Hallucinate in Multi-Document SummarizationCode0
Utilizing Large Language Models in an iterative paradigm with domain feedback for zero-shot molecule optimization0
On A Scale From 1 to 5: Quantifying Hallucination in Faithfulness Evaluation0
What Do LLMs Need to Understand Graphs: A Survey of Parametric Representation of Graphs0
Iter-AHMCL: Alleviate Hallucination for Large Language Model via Iterative Model-level Contrastive Learning0
A Claim Decomposition Benchmark for Long-form Answer VerificationCode0
RosePO: Aligning LLM-based Recommenders with Human Values0
When Not to Answer: Evaluating Prompts on GPT Models for Effective Abstention in Unanswerable Math Word Problems0
Controlled Automatic Task-Specific Synthetic Data Generation for Hallucination Detection0
AGENTiGraph: An Interactive Knowledge Graph Platform for LLM-based Chatbots Utilizing Private Data0
On the Capacity of Citation Generation by Large Language Models0
ReDeEP: Detecting Hallucination in Retrieval-Augmented Generation via Mechanistic Interpretability0
LargePiG: Your Large Language Model is Secretly a Pointer Generator0
Automatically Generating Visual Hallucination Test Cases for Multimodal Large Language ModelsCode0
Magnifier Prompt: Tackling Multimodal Hallucination via Extremely Simple Instructions0
Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs0
Can Structured Data Reduce Epistemic Uncertainty?0
Parenting: Optimizing Knowledge Selection of Retrieval-Augmented Language Models with Parameter Decoupling and Tailored Tuning0
SkillAggregation: Reference-free LLM-Dependent Aggregation0
Medico: Towards Hallucination Detection and Correction with Multi-source Evidence Fusion0
Honest AI: Fine-Tuning "Small" Language Models to Say "I Don't Know", and Reducing Hallucination in RAG0
Collu-Bench: A Benchmark for Predicting Language Model Hallucinations in Code0
Show:102550
← PrevPage 40 of 73Next →

No leaderboard results yet.