SOTAVerified

Hallucination

Papers

Showing 501550 of 1816 papers

TitleStatusHype
Diving Deep into Modes of Fact Hallucinations in Dialogue SystemsCode0
Enhancing Hallucination Detection through Perturbation-Based Synthetic Data Generation in System ResponsesCode0
MultiHal: Multilingual Dataset for Knowledge-Graph Grounded Evaluation of LLM HallucinationsCode0
NCL-UoR at SemEval-2025 Task 3: Detecting Multilingual Hallucination and Related Observable Overgeneration Text Spans with Modified RefChecker and Modified SeflCheckGPTCode0
NormSAGE: Multi-Lingual Multi-Cultural Norm Discovery from Conversations On-the-FlyCode0
On the Hallucination in Simultaneous Machine TranslationCode0
Pushing the Limits of Low-Resource Morphological InflectionCode0
Addressing Topic Granularity and Hallucination in Large Language Models for Topic ModellingCode0
MMBoundary: Advancing MLLM Knowledge Boundary Awareness through Reasoning Step Confidence CalibrationCode0
Projected Distribution Loss for Image EnhancementCode0
Mixture of Decoding: An Attention-Inspired Adaptive Decoding Strategy to Mitigate Hallucinations in Large Vision-Language ModelsCode0
Modality Distillation with Multiple Stream Networks for Action RecognitionCode0
Differentially Private Steering for Large Language Model AlignmentCode0
Beyond Ontology in Dialogue State Tracking for Goal-Oriented ChatbotCode0
DGoT: Dynamic Graph of Thoughts for Scientific Abstract GenerationCode0
The Knowledge Alignment Problem: Bridging Human and External Knowledge for Large Language ModelsCode0
Mitigating Hallucination of Large Vision-Language Models via Dynamic Logits CalibrationCode0
Mitigating Hallucination in Fictional Character Role-PlayCode0
Mitigating Entity-Level Hallucination in Large Language ModelsCode0
Mitigating Hallucination in Abstractive Summarization with Domain-Conditional Mutual InformationCode0
Detecting Errors through Ensembling Prompts (DEEP): An End-to-End LLM Framework for Detecting Factual ErrorsCode0
MELO: Enhancing Model Editing with Neuron-Indexed Dynamic LoRACode0
"Merge Conflicts!" Exploring the Impacts of External Distractors to Parametric Knowledge GraphsCode0
Analogy Models for Neural Word InflectionCode0
MedHallTune: An Instruction-Tuning Benchmark for Mitigating Medical Hallucination in Vision-Language ModelsCode0
Delve into Visual Contrastive Decoding for Hallucination Mitigation of Large Vision-Language ModelsCode0
Mechanistic Understanding and Mitigation of Language Model Non-Factual HallucinationsCode0
MCiteBench: A Multimodal Benchmark for Generating Text with CitationsCode0
Benchmarking Hallucination in Large Language Models based on Unanswerable Math Word ProblemCode0
MCQG-SRefine: Multiple Choice Question Generation and Evaluation with Iterative Self-Critique, Correction, and Comparison FeedbackCode0
MAF: Multi-Aspect Feedback for Improving Reasoning in Large Language ModelsCode0
MAVEN-Fact: A Large-scale Event Factuality Detection DatasetCode0
DefAn: Definitive Answer Dataset for LLMs Hallucination EvaluationCode0
Benchmarking ChatGPT-4 on ACR Radiation Oncology In-Training (TXIT) Exam and Red Journal Gray Zone Cases: Potentials and Challenges for AI-Assisted Medical Education and Decision Making in Radiation OncologyCode0
LVLM-Compress-Bench: Benchmarking the Broader Impact of Large Vision-Language Model CompressionCode0
Low to High Dimensional Modality Hallucination using Aggregated Fields of ViewCode0
Behind the Magic, MERLIM: Multi-modal Evaluation Benchmark for Large Image-Language ModelsCode0
Machine Translation Hallucination Detection for Low and High Resource Languages using Large Language ModelsCode0
MedScore: Factuality Evaluation of Free-Form Medical AnswersCode0
Logic Query of Thoughts: Guiding Large Language Models to Answer Complex Logic Queries with Knowledge GraphsCode0
LLMs and Memorization: On Quality and Specificity of Copyright ComplianceCode0
LongHalQA: Long-Context Hallucination Evaluation for MultiModal Large Language ModelsCode0
Deep CNN Denoiser and Multi-layer Neighbor Component Embedding for Face HallucinationCode0
EmotionHallucer: Evaluating Emotion Hallucinations in Multimodal Large Language ModelsCode0
LLM Hallucinations in Practical Code Generation: Phenomena, Mechanism, and MitigationCode0
DecoPrompt : Decoding Prompts Reduces Hallucinations when Large Language Models Meet False PremisesCode0
LLM Inference Enhanced by External Knowledge: A SurveyCode0
LLM-based Query Expansion Fails for Unfamiliar and Ambiguous QueriesCode0
LLM Internal States Reveal Hallucination Risk Faced With a QueryCode0
Deceptive Semantic Shortcuts on Reasoning Chains: How Far Can Models Go without Hallucination?Code0
Show:102550
← PrevPage 11 of 37Next →

No leaderboard results yet.