SOTAVerified

Hallucination

Papers

Showing 701725 of 1816 papers

TitleStatusHype
Improving Factuality in Large Language Models via Decoding-Time Hallucinatory and Truthful ComparatorsCode0
How Trustworthy are Open-Source LLMs? An Assessment under Malicious Demonstrations Shows their VulnerabilitiesCode0
Ever: Mitigating Hallucination in Large Language Models through Real-Time Verification and RectificationCode0
Chain-of-Action: Faithful and Multimodal Question Answering through Large Language ModelsCode0
How Much Do LLMs Hallucinate across Languages? On Multilingual Estimation of LLM Hallucination in the WildCode0
HypoTermQA: Hypothetical Terms Dataset for Benchmarking Hallucination Tendency of LLMsCode0
HELPD: Mitigating Hallucination of LVLMs by Hierarchical Feedback Learning with Vision-enhanced Penalty DecodingCode0
HICD: Hallucination-Inducing via Attention Dispersion for Contrastive Decoding to Mitigate Hallucinations in Large Language ModelsCode0
HaRiM^+: Evaluating Summary Quality with Hallucination RiskCode0
Handling Ontology Gaps in Semantic ParsingCode0
Abstract Meaning Representation for Hospital Discharge SummarizationCode0
Handwritten Code Recognition for Pen-and-Paper CS EducationCode0
HaloScope: Harnessing Unlabeled LLM Generations for Hallucination DetectionCode0
CCHall: A Novel Benchmark for Joint Cross-Lingual and Cross-Modal Hallucinations Detection in Large Language ModelsCode0
HALOS: Hallucination-free Organ Segmentation after Organ Resection SurgeryCode0
HaluEval-Wild: Evaluating Hallucinations of Language Models in the WildCode0
Are Large Language Models Good at Utility Judgments?Code0
Projected Distribution Loss for Image EnhancementCode0
Causal-LLaVA: Causal Disentanglement for Mitigating Hallucination in Multimodal Large Language ModelsCode0
HALO: Hallucination Analysis and Learning Optimization to Empower LLMs with Retrieval-Augmented Context for Guided Clinical Decision MakingCode0
Error Typing for Smarter Rewards: Improving Process Reward Models with Error-Aware Hierarchical SupervisionCode0
ERBench: An Entity-Relationship based Automatically Verifiable Hallucination Benchmark for Large Language ModelsCode0
How Helpful is Inverse Reinforcement Learning for Table-to-Text Generation?Code0
HalluciNet-ing Spatiotemporal Representations Using a 2D-CNNCode0
Catch Me if You Search: When Contextual Web Search Results Affect the Detection of HallucinationsCode0
Show:102550
← PrevPage 29 of 73Next →

No leaderboard results yet.