SOTAVerified

Hallucination

Papers

Showing 651700 of 1816 papers

TitleStatusHype
Few-shot learning via tensor hallucinationCode0
AILS-NTUA at SemEval-2024 Task 6: Efficient model tuning for hallucination detection and analysisCode0
Leveraging Pretrained Models for Automatic Summarization of Doctor-Patient ConversationsCode0
LightHouse: A Survey of AGI HallucinationCode0
MedHallTune: An Instruction-Tuning Benchmark for Mitigating Medical Hallucination in Vision-Language ModelsCode0
Learning to Localize Objects Improves Spatial Reasoning in Visual-LLMsCode0
Learning with privileged information via adversarial discriminative modality distillationCode0
GAPO: Learning Preferential Prompt through Generative Adversarial Policy OptimizationCode0
CiteBART: Learning to Generate Citations for Local Citation RecommendationCode0
Fakes of Varying Shades: How Warning Affects Human Perception and Engagement Regarding LLM HallucinationsCode0
Learning on LLM Output Signatures for gray-box LLM Behavior AnalysisCode0
Chain of Visual Perception: Harnessing Multimodal Large Language Models for Zero-shot Camouflaged Object DetectionCode0
Learning Conformal Abstention Policies for Adaptive Risk Management in Large Language and Vision-Language ModelsCode0
AIGCs Confuse AI Too: Investigating and Explaining Synthetic Image-induced Hallucinations in Large Vision-Language ModelsCode0
Large Language Models on Wikipedia-Style Survey Generation: an Evaluation in NLP ConceptsCode0
Learning Fine-grained Domain Generalization via Hyperbolic State Space HallucinationCode0
Language Models Hallucinate, but May Excel at Fact VerificationCode0
AI-Enhanced Cognitive Behavioral Therapy: Deep Learning and Large Language Models for Extracting Cognitive Pathways from Social Media TextsCode0
Large Language Models Are Involuntary Truth-Tellers: Exploiting Fallacy Failure for Jailbreak AttacksCode0
KG-FPQ: Evaluating Factuality Hallucination in LLMs with Knowledge Graph-based False Premise QuestionsCode0
Assessing the Reliability of Large Language Model KnowledgeCode0
JourneyBench: A Challenging One-Stop Vision-Language Understanding Benchmark of Generated ImagesCode0
Joint stereo 3D object detection and implicit surface reconstructionCode0
keepitsimple at SemEval-2025 Task 3: LLM-Uncertainty based Approach for Multilingual Hallucination Span DetectionCode0
Investigating the performance of Retrieval-Augmented Generation and fine-tuning for the development of AI-driven knowledge-based systemsCode0
Investigating Multi-Pivot Ensembling with Massively Multilingual Machine Translation ModelsCode0
Characterizing Multimodal Long-form Summarization: A Case Study on Financial ReportsCode0
Characterizing Context Influence and Hallucination in SummarizationCode0
Investigating and Mitigating Object Hallucinations in Pretrained Vision-Language (CLIP) ModelsCode0
Integrating Chemistry Knowledge in Large Language Models via Prompt EngineeringCode0
Iterative Teaching by Data HallucinationCode0
Incorporating Task-specific Concept Knowledge into Script LearningCode0
Improving Factual Error Correction by Learning to Inject Factual ErrorsCode0
A Claim Decomposition Benchmark for Long-form Answer VerificationCode0
Improving Factuality in Large Language Models via Decoding-Time Hallucinatory and Truthful ComparatorsCode0
Image Denoising with Control over Deep Network HallucinationCode0
CHAIR -- Classifier of Hallucination as ImproverCode0
Instruction Makes a DifferenceCode0
Chainpoll: A high efficacy method for LLM hallucination detectionCode0
Explaining Graph Neural Networks with Large Language Models: A Counterfactual Perspective for Molecular Property PredictionCode0
Chain-of-Verification Reduces Hallucination in Large Language ModelsCode0
HypoTermQA: Hypothetical Terms Dataset for Benchmarking Hallucination Tendency of LLMsCode0
Controlling Risk of Retrieval-augmented Generation: A Counterfactual Prompting FrameworkCode0
On the Benefits of Fine-Grained Loss Truncation: A Case Study on Factuality in SummarizationCode0
How Helpful is Inverse Reinforcement Learning for Table-to-Text Generation?Code0
How Much Do LLMs Hallucinate across Languages? On Multilingual Estimation of LLM Hallucination in the WildCode0
Evolutionary thoughts: integration of large language models and evolutionary algorithmsCode0
How Trustworthy are Open-Source LLMs? An Assessment under Malicious Demonstrations Shows their VulnerabilitiesCode0
Im2Avatar: Colorful 3D Reconstruction from a Single ImageCode0
Ever: Mitigating Hallucination in Large Language Models through Real-Time Verification and RectificationCode0
Show:102550
← PrevPage 14 of 37Next →

No leaderboard results yet.