SOTAVerified

Hallucination

Papers

Showing 551600 of 1816 papers

TitleStatusHype
Data-Centric Human Preference Optimization with RationalesCode0
Mitigating Entity-Level Hallucination in Large Language ModelsCode0
"Merge Conflicts!" Exploring the Impacts of External Distractors to Parametric Knowledge GraphsCode0
DAHL: Domain-specific Automated Hallucination Evaluation of Long-Form Text through a Benchmark Dataset in BiomedicineCode0
Automatically Generating Visual Hallucination Test Cases for Multimodal Large Language ModelsCode0
Mitigating Hallucination in Abstractive Summarization with Domain-Conditional Mutual InformationCode0
DAFNet: Dynamic Auxiliary Fusion for Sequential Model Editing in Large Language ModelsCode0
MELO: Enhancing Model Editing with Neuron-Indexed Dynamic LoRACode0
Evaluating LLMs' Assessment of Mixed-Context Hallucination Through the Lens of SummarizationCode0
MedScore: Factuality Evaluation of Free-Form Medical AnswersCode0
MedTSS: transforming abstractive summarization of scientific articles with linguistic analysis and concept reinforcementCode0
Mitigating Hallucination in Fictional Character Role-PlayCode0
Cross-modal Learning by Hallucinating Missing Modalities in RGB-D VisionCode0
MCQG-SRefine: Multiple Choice Question Generation and Evaluation with Iterative Self-Critique, Correction, and Comparison FeedbackCode0
MAVEN-Fact: A Large-scale Event Factuality Detection DatasetCode0
MCiteBench: A Multimodal Benchmark for Generating Text with CitationsCode0
Mechanistic Understanding and Mitigation of Language Model Non-Factual HallucinationsCode0
Machine Translation Hallucination Detection for Low and High Resource Languages using Large Language ModelsCode0
Critic-Driven Decoding for Mitigating Hallucinations in Data-to-text GenerationCode0
MAF: Multi-Aspect Feedback for Improving Reasoning in Large Language ModelsCode0
Low to High Dimensional Modality Hallucination using Aggregated Fields of ViewCode0
Crafting In-context Examples according to LMs' Parametric KnowledgeCode0
LVLM-Compress-Bench: Benchmarking the Broader Impact of Large Vision-Language Model CompressionCode0
Counterfactual Debating with Preset Stances for Hallucination Elimination of LLMsCode0
A Unified Hallucination Mitigation Framework for Large Vision-Language ModelsCode0
Correction with Backtracking Reduces Hallucination in SummarizationCode0
MedHallTune: An Instruction-Tuning Benchmark for Mitigating Medical Hallucination in Vision-Language ModelsCode0
Mitigating Hallucination of Large Vision-Language Models via Dynamic Logits CalibrationCode0
Conversational Gold: Evaluating Personalized Conversational Search System using Gold NuggetsCode0
LLMs and Memorization: On Quality and Specificity of Copyright ComplianceCode0
Controlling Risk of Retrieval-augmented Generation: A Counterfactual Prompting FrameworkCode0
Alleviating Hallucinations in Large Vision-Language Models through Hallucination-Induced OptimizationCode0
LLM Inference Enhanced by External Knowledge: A SurveyCode0
LLM Hallucinations in Practical Code Generation: Phenomena, Mechanism, and MitigationCode0
LLM Internal States Reveal Hallucination Risk Faced With a QueryCode0
LLM-based Query Expansion Fails for Unfamiliar and Ambiguous QueriesCode0
A Comprehensive Survey of Hallucination Mitigation Techniques in Large Language ModelsCode0
Linear Correlation in LM's Compositional Generalization and HallucinationCode0
Logic Query of Thoughts: Guiding Large Language Models to Answer Complex Logic Queries with Knowledge GraphsCode0
Leveraging Pretrained Models for Automatic Summarization of Doctor-Patient ConversationsCode0
Learning with privileged information via adversarial discriminative modality distillationCode0
Confidence Estimation for LLM-Based Dialogue State TrackingCode0
Confidence-aware Denoised Fine-tuning of Off-the-shelf Models for Certified RobustnessCode0
Exploring the Trade-Offs: Quantization Methods, Task Difficulty, and Model Size in Large Language Models From Edge to GiantCode0
Learning on LLM Output Signatures for gray-box LLM Behavior AnalysisCode0
Learning to Localize Objects Improves Spatial Reasoning in Visual-LLMsCode0
Large Language Models on Wikipedia-Style Survey Generation: an Evaluation in NLP ConceptsCode0
Learning Conformal Abstention Policies for Adaptive Risk Management in Large Language and Vision-Language ModelsCode0
Large Language Models Are Involuntary Truth-Tellers: Exploiting Fallacy Failure for Jailbreak AttacksCode0
Language Models Hallucinate, but May Excel at Fact VerificationCode0
Show:102550
← PrevPage 12 of 37Next →

No leaderboard results yet.