SOTAVerified

Hallucination

Papers

Showing 201250 of 1816 papers

TitleStatusHype
Benchmarking Llama2, Mistral, Gemma and GPT for Factuality, Toxicity, Bias and Propensity for HallucinationsCode1
Analyzing and Mitigating Object Hallucination in Large Vision-Language ModelsCode1
High-resolution Face Swapping via Latent Semantics DisentanglementCode1
Into the Unknown: Self-Learning Large Language ModelsCode1
Hallucination Detection in LLMs Using Spectral Features of Attention MapsCode1
Hallucination Augmented Contrastive Learning for Multimodal Large Language ModelCode1
Hallucinated Neural Radiance Fields in the WildCode1
Hallucination-Aware Multimodal Benchmark for Gastrointestinal Image Analysis with Large Vision-Language ModelsCode1
Benchmarking LLM Faithfulness in RAG with Evolving LeaderboardsCode1
HallE-Control: Controlling Object Hallucination in Large Multimodal ModelsCode1
Phare: A Safety Probe for Large Language ModelsCode1
Chain of Natural Language Inference for Reducing Large Language Model Ungrounded HallucinationsCode1
Beyond Generic: Enhancing Image Captioning with Real-World Knowledge using Vision-Language Pre-Training ModelCode1
Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference OptimizationCode1
HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction DataCode1
A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and InteractivityCode1
How well can a large language model explain business processes as perceived by users?Code1
BAMBOO: A Comprehensive Benchmark for Evaluating Long Text Modeling Capacities of Large Language ModelsCode1
Grounded Chain-of-Thought for Multimodal Large Language ModelsCode1
Balanced Classification: A Unified Framework for Long-Tailed Object DetectionCode1
BIGPrior: Towards Decoupling Learned Prior Hallucination and Data Fidelity in Image RestorationCode1
InterrogateLLM: Zero-Resource Hallucination Detection in LLM-Generated AnswersCode1
BachGAN: High-Resolution Image Synthesis from Salient Object LayoutCode1
GraphArena: Benchmarking Large Language Models on Graph Computational ProblemsCode1
HaloQuest: A Visual Hallucination Dataset for Advancing Multimodal ReasoningCode1
Investigating and Mitigating the Multimodal Hallucination Snowballing in Large Vision-Language ModelsCode1
Introspective Planning: Aligning Robots' Uncertainty with Inherent Task AmbiguityCode1
IterGen: Iterative Semantic-aware Structured LLM Generation with BacktrackingCode1
Joint Evaluation of Answer and Reasoning Consistency for Hallucination Detection in Large Reasoning ModelsCode1
KCTS: Knowledge-Constrained Tree Search Decoding with Token-Level Hallucination DetectionCode1
LLMs Know What They Need: Leveraging a Missing Information Guided Framework to Empower Retrieval-Augmented GenerationCode1
FlySearch: Exploring how vision-language models exploreCode1
Automatic Curriculum Expert Iteration for Reliable LLM ReasoningCode1
3D Sketch-aware Semantic Scene Completion via Semi-supervised Structure PriorCode1
Advancing TTP Analysis: Harnessing the Power of Large Language Models with Retrieval Augmented GenerationCode1
BTR: Binary Token Representations for Efficient Retrieval Augmented Language ModelsCode1
AMBER: An LLM-free Multi-dimensional Benchmark for MLLMs Hallucination EvaluationCode1
CAFe: Unifying Representation and Generation with Contrastive-Autoregressive FinetuningCode1
AdaPlanner: Adaptive Planning from Feedback with Language ModelsCode1
Gemini Goes to Med School: Exploring the Capabilities of Multimodal Large Language Models on Medical Challenge Problems & HallucinationsCode1
LAN-HDR: Luminance-based Alignment Network for High Dynamic Range Video ReconstructionCode1
Large Language Models are Versatile Decomposers: Decompose Evidence and Questions for Table-based ReasoningCode1
Automated Review Generation Method Based on Large Language ModelsCode1
FineSurE: Fine-grained Summarization Evaluation using LLMsCode1
Antidote: A Unified Framework for Mitigating LVLM Hallucinations in Counterfactual Presupposition and Object PerceptionCode1
Can Knowledge Editing Really Correct Hallucinations?Code1
Automated Multi-level Preference for MLLMsCode1
Finetune-RAG: Fine-Tuning Language Models to Resist Hallucination in Retrieval-Augmented GenerationCode1
LiDAR-based 4D Occupancy Completion and ForecastingCode1
PAINT: Paying Attention to INformed Tokens to Mitigate Hallucination in Large Vision-Language ModelCode1
Show:102550
← PrevPage 5 of 37Next →

No leaderboard results yet.