SOTAVerified

Hallucination

Papers

Showing 276300 of 1816 papers

TitleStatusHype
DEEM: Diffusion Models Serve as the Eyes of Large Language Models for Image PerceptionCode1
Visual Description Grounding Reduces Hallucinations and Boosts Reasoning in LVLMsCode1
The 2nd FutureDial Challenge: Dialog Systems with Retrieval Augmented Generation (FutureDial-RAG)Code1
Automated Multi-level Preference for MLLMsCode1
Enhancing Semantics in Multimodal Chain of Thought via Soft Negative SamplingCode1
THRONE: An Object-based Hallucination Benchmark for the Free-form Generations of Large Vision-Language ModelsCode1
CodeHalu: Investigating Code Hallucinations in LLMs via Execution-based VerificationCode1
LLMs Know What They Need: Leveraging a Missing Information Guided Framework to Empower Retrieval-Augmented GenerationCode1
VALOR-EVAL: Holistic Coverage and Faithfulness Evaluation of Large Vision-Language ModelsCode1
Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI FeedbackCode1
Exploring the Transferability of Visual Prompting for Multimodal Large Language ModelsCode1
MemLLM: Finetuning LLMs to Use An Explicit Read-Write MemoryCode1
Benchmarking Llama2, Mistral, Gemma and GPT for Factuality, Toxicity, Bias and Propensity for HallucinationsCode1
Harnessing GPT-4V(ision) for Insurance: A Preliminary ExplorationCode1
Constructing Benchmarks and Interventions for Combating Hallucinations in LLMsCode1
CuriousLLM: Elevating Multi-Document QA with Reasoning-Infused Knowledge Graph PromptingCode1
Tackling Structural Hallucination in Image Translation with Local DiffusionCode1
Learning From Correctness Without Prompting Makes LLM Efficient ReasonerCode1
Retrieval-enhanced Knowledge Editing in Language Models for Multi-Hop Question AnsweringCode1
JDocQA: Japanese Document Question Answering Dataset for Generative Language ModelsCode1
UrbanVLP: Multi-Granularity Vision-Language Pretraining for Urban Socioeconomic Indicator PredictionCode1
Pensieve: Retrospect-then-Compare Mitigates Visual HallucinationCode1
What if...?: Thinking Counterfactual Keywords Helps to Mitigate Hallucination in Large Multi-modal ModelsCode1
PhD: A ChatGPT-Prompted Visual hallucination Evaluation DatasetCode1
Circuit Transformer: A Transformer That Preserves Logical EquivalenceCode1
Show:102550
← PrevPage 12 of 73Next →

No leaderboard results yet.