SOTAVerified

Hallucination

Papers

Showing 11011150 of 1816 papers

TitleStatusHype
VHM: Versatile and Honest Vision Language Model for Remote Sensing Image AnalysisCode2
Learning From Correctness Without Prompting Makes LLM Efficient ReasonerCode1
Are Large Language Models Good at Utility Judgments?Code0
JDocQA: Japanese Document Question Answering Dataset for Generative Language ModelsCode1
FACTOID: FACtual enTailment fOr hallucInation Detection0
Retrieval-enhanced Knowledge Editing in Language Models for Multi-Hop Question AnsweringCode1
"Sorry, Come Again?" Prompting -- Enhancing Comprehension and Diminishing Hallucination with [PAUSE]-injected Optimal Paraphrasing0
Mechanistic Understanding and Mitigation of Language Model Non-Factual HallucinationsCode0
A Diffusion-Based Generative Equalizer for Music RestorationCode2
Rejection Improves Reliability: Training LLMs to Refuse Unknown Questions Using RL from Knowledge Feedback0
Chain-of-Action: Faithful and Multimodal Question Answering through Large Language ModelsCode0
Visual Hallucination: Definition, Quantification, and Prescriptive Remediations0
DGoT: Dynamic Graph of Thoughts for Scientific Abstract GenerationCode0
UrbanVLP: Multi-Granularity Vision-Language Pretraining for Urban Socioeconomic Indicator PredictionCode1
Dyna-LfLH: Learning Agile Navigation in Dynamic Environments from Learned Hallucination0
Hallucination Detection in Foundation Models for Decision-Making: A Flexible Definition and Review of the State of the Art0
ESREAL: Exploiting Semantic Reconstruction to Mitigate Hallucinations in Vision-Language Models0
Make VLM Recognize Visual Hallucination on Cartoon Character Image with Pose Information0
Sphere Neural-Networks for Rational Reasoning0
Pensieve: Retrospect-then-Compare Mitigates Visual HallucinationCode1
Multi-Modal Hallucination Control by Visual Information Grounding0
What if...?: Thinking Counterfactual Keywords Helps to Mitigate Hallucination in Large Multi-modal ModelsCode1
Agent-FLAN: Designing Data and Methods of Effective Agent Tuning for Large Language ModelsCode3
SpatialPIN: Enhancing Spatial Reasoning Capabilities of Vision-Language Models through Prompting and Interacting 3D Priors0
Zero-Shot Multi-task Hallucination Detection0
DEE: Dual-stage Explainable Evaluation Method for Text Generation0
Logic Query of Thoughts: Guiding Large Language Models to Answer Complex Logic Queries with Knowledge GraphsCode0
PhD: A ChatGPT-Prompted Visual hallucination Evaluation DatasetCode1
Think Twice Before Trusting: Self-Detection for Large Language Models through Comprehensive Answer Reflection0
DiffMAC: Diffusion Manifold Hallucination Correction for High Generalization Blind Face Restoration0
Mitigating Dialogue Hallucination for Large Vision Language Models via Adversarial Instruction Tuning0
Circuit Transformer: A Transformer That Preserves Logical EquivalenceCode1
XReal: Realistic Anatomy and Pathology-Aware X-ray Generation via Controllable Diffusion ModelCode1
The First to Know: How Token Distributions Reveal Hidden Knowledge in Large Vision-Language Models?Code1
Detecting Hallucination and Coverage Errors in Retrieval Augmented Generation for Controversial Topics0
AIGCs Confuse AI Too: Investigating and Explaining Synthetic Image-induced Hallucinations in Large Vision-Language ModelsCode0
Investigating the performance of Retrieval-Augmented Generation and fine-tuning for the development of AI-driven knowledge-based systemsCode0
Put Myself in Your Shoes: Lifting the Egocentric Perspective from Exocentric Videos0
TRAWL: External Knowledge-Enhanced Recommendation with LLM Assistance0
Guiding Clinical Reasoning with Large Language Models via Knowledge Seeds0
Unsupervised Real-Time Hallucination Detection based on the Internal States of Large Language ModelsCode2
On the Benefits of Fine-Grained Loss Truncation: A Case Study on Factuality in SummarizationCode0
Tuning-Free Accountable Intervention for LLM Deployment -- A Metacognitive Approach0
Can Large Language Models Play Games? A Case Study of A Self-Play Approach0
ERBench: An Entity-Relationship based Automatically Verifiable Hallucination Benchmark for Large Language ModelsCode0
ChatASU: Evoking LLM's Reflexion to Truly Understand Aspect Sentiment in Dialogues0
Sora as an AGI World Model? A Complete Survey on Text-to-Video Generation0
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Horizon GenerationCode3
HaluEval-Wild: Evaluating Hallucinations of Language Models in the WildCode0
Federated Recommendation via Hybrid Retrieval Augmented GenerationCode1
Show:102550
← PrevPage 23 of 37Next →

No leaderboard results yet.