SOTAVerified

Hallucination

Papers

Showing 501550 of 1816 papers

TitleStatusHype
HalluCana: Fixing LLM Hallucination with A Canary Lookahead0
Hallucination Elimination and Semantic Enhancement Framework for Vision-Language Models in Traffic ScenariosCode0
Granite GuardianCode2
Delve into Visual Contrastive Decoding for Hallucination Mitigation of Large Vision-Language ModelsCode0
Methods for Legal Citation Prediction in the Age of LLMs: An Australian Law Case Study0
Evaluating Hallucination in Text-to-Image Diffusion Models with Scene-Graph based Question-Answering Agent0
Multi-Objective Alignment of Large Language Models Through Hypervolume Maximization0
Steps are all you need: Rethinking STEM Education with Prompt Engineering0
TOBUGraph: Knowledge Graph-Based Retrieval for Enhanced LLM Performance Beyond RAG0
100% Elimination of Hallucinations on RAGTruth for GPT-4 and GPT-3.5 Turbo0
Verb Mirage: Unveiling and Assessing Verb Concept Hallucinations in Multimodal Large Language Models0
LLM-Align: Utilizing Large Language Models for Entity Alignment in Knowledge Graphs0
Expanding Performance Boundaries of Open-Source Multimodal Models with Model, Data, and Test-Time Scaling0
Florence-VL: Enhancing Vision-Language Models with Generative Vision Encoder and Depth-Breadth FusionCode3
GenMAC: Compositional Text-to-Video Generation with Multi-Agent Collaboration0
Reducing Tool Hallucination via Reliability Alignment0
Deep priors for satellite image restoration with accurate uncertainties0
Who Brings the Frisbee: Probing Hidden Hallucination Factors in Large Vision-Language Model via Causality Analysis0
VidHalluc: Evaluating Temporal Hallucinations in Multimodal Large Language Models for Video Understanding0
An Evolutionary Large Language Model for Hallucination Mitigation0
CC-OCR: A Comprehensive and Challenging OCR Benchmark for Evaluating Large Multimodal Models in Literacy0
AI Benchmarks and Datasets for LLM Evaluation0
Automating Feedback Analysis in Surgical Training: Detection, Categorization, and AssessmentCode0
Beyond Logit Lens: Contextual Embeddings for Robust Hallucination Detection & Grounding in VLMs0
OPCap:Object-aware Prompting Captioning0
DHCP: Detecting Hallucinations by Cross-modal Attention Pattern in Large Vision-Language Models0
Can LLMs be Good Graph Judge for Knowledge Graph Construction?Code1
Efficient Self-Improvement in Multimodal Large Language Models: A Model-Level Judge-Free Approach0
Meaningless is better: hashing bias-inducing words in LLM prompts improves performance in logical reasoning and statistical learning0
VLRewardBench: A Challenging Benchmark for Vision-Language Generative Reward Models0
A Topic-level Self-Correctional Approach to Mitigate Hallucinations in MLLMs0
AI2T: Building Trustable AI Tutors by Interactively Teaching a Self-Aware Learning Agent0
VidHal: Benchmarking Temporal Hallucinations in Vision LLMsCode1
AtomR: Atomic Operator-Empowered Large Language Models for Heterogeneous Knowledge ReasoningCode1
Enhancing Multi-Agent Consensus through Third-Party LLM Integration: Analyzing Uncertainty and Mitigating Hallucinations in Large Language Models0
O1 Replication Journey -- Part 2: Surpassing O1-preview through Simple Distillation, Big Progress or Bitter Lesson?Code7
VaLiD: Mitigating the Hallucination of Large Vision Language Models by Visual Layer Fusion Contrastive DecodingCode1
Devils in Middle Layers of Large Vision-Language Models: Interpreting, Detecting and Mitigating Object Hallucinations via Attention LensCode2
Ontology-Constrained Generation of Domain-Specific Clinical SummariesCode0
ICT: Image-Object Cross-Level Trusted Intervention for Mitigating Object Hallucination in Large Vision-Language Models0
Detecting Hallucinations in Virtual Histology with Neural Precursors0
Leveraging LLMs for Legacy Code Modernization: Challenges and Opportunities for LLM-Generated Documentation0
Sycophancy in Large Language Models: Causes and Mitigations0
CATCH: Complementary Adaptive Token-level Contrastive Decoding to Mitigate Hallucinations in LVLMs0
Can Open-source LLMs Enhance Data Synthesis for Toxic Detection?: An Experimental Study0
VL-Uncertainty: Detecting Hallucination in Large Vision-Language Model via Uncertainty EstimationCode0
Mitigating Knowledge Conflicts in Language Model-Driven Question Answering0
Enabling Explainable Recommendation in E-commerce with LLM-powered Product Knowledge Graph0
INVARLLM: LLM-assisted Physical Invariant Extraction for Cyber-Physical Systems Anomaly Detection0
Understanding Multimodal LLMs: the Mechanistic Interpretability of Llava in Visual Question AnsweringCode0
Show:102550
← PrevPage 11 of 37Next →

No leaderboard results yet.