SOTAVerified

Hallucination

Papers

Showing 51100 of 1816 papers

TitleStatusHype
LLMDFA: Analyzing Dataflow in Code with Large Language ModelsCode3
Siren's Song in the AI Ocean: A Survey on Hallucination in Large Language ModelsCode3
Florence-VL: Enhancing Vision-Language Models with Generative Vision Encoder and Depth-Breadth FusionCode3
Benchmarking Multimodal Retrieval Augmented Generation with Dynamic VQA Dataset and Self-adaptive Planning AgentCode3
Retrieval Head Mechanistically Explains Long-Context FactualityCode3
Evaluating Hallucinations in Chinese Large Language ModelsCode3
RAGEval: Scenario Specific RAG Evaluation Dataset Generation FrameworkCode3
ResumeFlow: An LLM-facilitated Pipeline for Personalized Resume Generation and RefinementCode3
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Horizon GenerationCode3
PULSE: Self-Supervised Photo Upsampling via Latent Space Exploration of Generative ModelsCode3
CRAG -- Comprehensive RAG BenchmarkCode3
PokeLLMon: A Human-Parity Agent for Pokemon Battles with Large Language ModelsCode3
Agent-FLAN: Designing Data and Methods of Effective Agent Tuning for Large Language ModelsCode3
RefChecker: Reference-based Fine-grained Hallucination Checker and Benchmark for Large Language ModelsCode3
The Curse of Multi-Modalities: Evaluating Hallucinations of Large Multimodal Models across Language, Visual, and AudioCode3
PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language ModelsCode3
Mitigating Object Hallucination via Concentric Causal AttentionCode2
Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local AttentionCode2
MLLM can see? Dynamic Correction Decoding for Hallucination MitigationCode2
Self-Introspective Decoding: Alleviating Hallucinations for Large Vision-Language ModelsCode2
Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive DecodingCode2
MindMap: Knowledge Graph Prompting Sparks Graph of Thoughts in Large Language ModelsCode2
A Diffusion-Based Generative Equalizer for Music RestorationCode2
Mitigating Hallucinations in Large Vision-Language Models via DPO: On-Policy Data Hold the KeyCode2
Medical Hallucinations in Foundation Models and Their Impact on HealthcareCode2
mDPO: Conditional Preference Optimization for Multimodal Large Language ModelsCode2
MeMemo: On-device Retrieval Augmentation for Private and Personalized Text GenerationCode2
Mitigating Modality Prior-Induced Hallucinations in Multimodal Large Language Models via Deciphering Attention CausalityCode2
MMICL: Empowering Vision-language Model with Multi-Modal In-Context LearningCode2
3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less HallucinationCode2
Dynamic Parametric Retrieval Augmented Generation for Test-time Knowledge EnhancementCode2
Controllable and Reliable Knowledge-Intensive Task-Oriented Conversational Agents with Declarative Genie WorksheetsCode2
Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention MapsCode2
Look Twice Before You Answer: Memory-Space Visual Retracing for Hallucination Mitigation in Multimodal Large Language ModelsCode2
MLAgentBench: Evaluating Language Agents on Machine Learning ExperimentationCode2
ANAH: Analytical Annotation of Hallucinations in Large Language ModelsCode2
LLaMP: Large Language Model Made Powerful for High-fidelity Materials Knowledge Retrieval and DistillationCode2
Lawyer LLaMA Technical ReportCode2
Benchmarking Large Language Models in Retrieval-Augmented GenerationCode2
Less is More: Mitigating Multimodal Hallucination from an EOS Decision PerspectiveCode2
LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language ModelsCode2
In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination MitigationCode2
Image Textualization: An Automatic Framework for Creating Accurate and Detailed Image DescriptionsCode2
InstructGraph: Boosting Large Language Models via Graph-centric Instruction Tuning and Preference AlignmentCode2
HaluEval: A Large-Scale Hallucination Evaluation Benchmark for Large Language ModelsCode2
Automatically Correcting Large Language Models: Surveying the landscape of diverse self-correction strategiesCode2
Interleaving Retrieval with Chain-of-Thought Reasoning for Knowledge-Intensive Multi-Step QuestionsCode2
HalOmi: A Manually Annotated Benchmark for Multilingual Hallucination and Omission Detection in Machine TranslationCode2
HallusionBench: An Advanced Diagnostic Suite for Entangled Language Hallucination and Visual Illusion in Large Vision-Language ModelsCode2
Aligning Modalities in Vision Large Language Models via Preference Fine-tuningCode2
Show:102550
← PrevPage 2 of 37Next →

No leaderboard results yet.