SOTAVerified

Hallucination

Papers

Showing 151200 of 1816 papers

TitleStatusHype
Ornithologist: Towards Trustworthy "Reasoning" about Central Bank Communications0
Prioritizing Image-Related Tokens Enhances Vision-Language Pre-TrainingCode0
Improving the Reliability of LLMs: Combining CoT, RAG, Self-Consistency, and Self-Verification0
Adaptive Schema-aware Event Extraction with Retrieval-Augmented Generation0
A Head to Predict and a Head to Question: Pre-trained Uncertainty Quantification Heads for Hallucination Detection in LLM OutputsCode1
SEReDeEP: Hallucination Detection in Retrieval-Augmented Models via Semantic Entropy and Context-Parameter Fusion0
On the Cost and Benefits of Training Context with Utterance or Full Conversation Training: A Comparative Stud0
Multimodal Survival Modeling in the Age of Foundation ModelsCode0
Critique Before Thinking: Mitigating Hallucination through Rationale-Augmented Instruction Tuning0
TrumorGPT: Graph-Based Retrieval-Augmented Large Language Model for Fact-Checking0
Hallucination-Aware Multimodal Benchmark for Gastrointestinal Image Analysis with Large Vision-Language ModelsCode1
Evolutionary thoughts: integration of large language models and evolutionary algorithmsCode0
Osiris: A Lightweight Open-Source Hallucination Detection System0
Benchmarking LLM Faithfulness in RAG with Evolving LeaderboardsCode1
Interpretable Zero-shot Learning with Infinite Class Concepts0
Mitigating Image Captioning Hallucinations in Vision-Language Models0
Invoke Interfaces Only When Needed: Adaptive Invocation for Large Language Models in Question AnsweringCode1
UCSC at SemEval-2025 Task 3: Context, Models and Prompt Optimization for Automated Hallucination Detection in LLM OutputCode0
Knowledge Graphs for Enhancing Large Language Models in Entity Disambiguation0
A Comprehensive Analysis for Visual Object Hallucination in Large Vision-Language Models0
SEval-Ex: A Statement-Level Framework for Explainable Summarization Evaluation0
Regression is all you need for medical image translationCode0
Multi-agents based User Values Mining for Recommendation0
VideoHallu: Evaluating and Mitigating Multi-modal Hallucinations on Synthetic Video UnderstandingCode1
Automated Parsing of Engineering Drawings for Structured Information Extraction Using a Fine-tuned Document Understanding Transformer0
HalluMix: A Task-Agnostic, Multi-Domain Benchmark for Real-World Hallucination Detection0
Triggering Hallucinations in LLMs: A Quantitative Study of Prompt-Induced Hallucination in Large Language Models0
SmallPlan: Leverage Small Language Models for Sequential Path Planning with Simulation-Powered, LLM-Guided DistillationCode0
Efficient and robust 3D blind harmonization for large domain gaps0
MAC-Tuning: LLM Multi-Compositional Problem Reasoning with Enhanced Knowledge Boundary Awareness0
Black-Box Visual Prompt Engineering for Mitigating Object Hallucination in Large Vision Language Models0
Localizing Before Answering: A Hallucination Evaluation Benchmark for Grounded Medical Multimodal LLMs0
Antidote: A Unified Framework for Mitigating LVLM Hallucinations in Counterfactual Presupposition and Object PerceptionCode1
Can LLMs Detect Intrinsic Hallucinations in Paraphrasing and Machine Translation?0
Hallucination by Code Generation LLMs: Taxonomy, Benchmarks, Mitigation, and Challenges0
An Automated Reinforcement Learning Reward Design Framework with Large Language Model for Cooperative Platoon Coordination0
Explanatory Summarization with Discourse-Driven Planning0
Uncertainty Quantification for Language Models: A Suite of Black-Box, White-Box, LLM Judge, and Ensemble ScorersCode5
Validating Network Protocol Parsers with Traceable RFC Document Interpretation0
Data-Driven Calibration of Prediction Sets in Large Vision-Language Models Based on Inductive Conformal Prediction0
Toward Personalizing Quantum Computing Education: An Evolutionary LLM-Powered Approach0
The Dance of Atoms-De Novo Protein Design with Diffusion Model0
(Im)possibility of Automated Hallucination Detection in Large Language Models0
Grounded in Context: Retrieval-Based Method for Hallucination Detection0
Insights from Verification: Training a Verilog Generation LLM with Reinforcement Learning with Testbench Feedback0
DyFo: A Training-Free Dynamic Focus Visual Search for Enhancing LMMs in Fine-Grained Visual UnderstandingCode2
POLYRAG: Integrating Polyviews into Retrieval-Augmented Generation for Medical Applications0
aiXamine: Simplified LLM Safety and Security0
ResNetVLLM-2: Addressing ResNetVLLM's Multi-Modal Hallucinations0
Hydra: An Agentic Reasoning Approach for Enhancing Adversarial Robustness and Mitigating Hallucinations in Vision-Language Models0
Show:102550
← PrevPage 4 of 37Next →

No leaderboard results yet.