SOTAVerified

Hallucination

Papers

Showing 351400 of 1816 papers

TitleStatusHype
LLM-QE: Improving Query Expansion by Aligning Large Language Models with Ranking PreferencesCode1
LettuceDetect: A Hallucination Detection Framework for RAG ApplicationsCode4
`Generalization is hallucination' through the lens of tensor completions0
Uncertainty-Aware Fusion: An Ensemble Framework for Mitigating Hallucinations in Large Language Models0
ZiGong 1.0: A Large Language Model for Financial Credit0
The Law of Knowledge Overshadowing: Towards Understanding, Predicting, and Preventing LLM Hallucination0
PIP-KAG: Mitigating Knowledge Conflicts in Knowledge-Augmented Generation via Parametric PruningCode2
The Role of Background Information in Reducing Object Hallucination in Vision-Language Models: Insights from Cutoff API Prompting0
Hallucination Detection in Large Language Models with Metamorphic Relations0
Verify when Uncertain: Beyond Self-Consistency in Black Box Hallucination Detection0
Large Language Models Struggle to Describe the Haystack without Human Help: Human-in-the-loop Evaluation of LLMs0
MedHallu: A Comprehensive Benchmark for Detecting Medical Hallucinations in Large Language Models0
SegSub: Evaluating Robustness to Knowledge Conflicts and Hallucinations in Vision-Language ModelsCode0
OpenSearch-SQL: Enhancing Text-to-SQL with Dynamic Few-shot and Consistency Alignment0
Detecting LLM Fact-conflicting Hallucinations Enhanced by Temporal-logic-based Reasoning0
REFIND: Retrieval-Augmented Factuality Hallucination Detection in Large Language Models0
What are Models Thinking about? Understanding Large Language Model Hallucinations "Psychology" through Model Inner State Analysis0
TreeCut: A Synthetic Unanswerable Math Word Problem Dataset for LLM Hallucination EvaluationCode0
Lost in Transcription, Found in Distribution Shift: Demystifying Hallucination in Speech Foundation Models0
CutPaste&Find: Efficient Multimodal Hallucination Detector with Visual-aid Knowledge Base0
R2-KG: General-Purpose Dual-Agent Framework for Reliable Reasoning on Knowledge GraphsCode1
How Much Do LLMs Hallucinate across Languages? On Multilingual Estimation of LLM Hallucination in the WildCode0
Unveiling the Magic of Code Reasoning through Hypothesis Decomposition and AmendmentCode2
Can Your Uncertainty Scores Detect Hallucinated Entity?0
Smoothing Out Hallucinations: Mitigating LLM Hallucination with Smoothed Knowledge Distillation0
Valuable Hallucinations: Realizable Non-realistic Propositions0
A Survey of LLM-based Agents in Medicine: How far are we from Baymax?0
Automated Hypothesis Validation with Agentic Sequential FalsificationsCode3
Enhancing RAG with Active Learning on Conversation Records: Reject Incapables and Answer Capables0
DeepSeek on a Trip: Inducing Targeted Visual Hallucinations via Representation Vulnerabilities0
Elevating Legal LLM Responses: Harnessing Trainable Logical Structures and Semantic Knowledge with Legal ReasoningCode0
Hallucination, Monofacts, and Miscalibration: An Empirical InvestigationCode0
Refine Knowledge of Large Language Models via Adaptive Contrastive Learning0
Hallucination Detection: A Probabilistic Framework Using Embeddings Distance Analysis0
Knowledge Graph-Guided Retrieval Augmented GenerationCode2
Learning Conformal Abstention Policies for Adaptive Risk Management in Large Language and Vision-Language ModelsCode0
Self-Rationalization in the Wild: A Large Scale Out-of-Distribution Evaluation on NLI-related tasksCode0
VideoRoPE: What Makes for Good Video Rotary Position Embedding?Code3
ChallengeMe: An Adversarial Learning-enabled Text Summarization Framework0
Linear Correlation in LM's Compositional Generalization and HallucinationCode0
TruthFlow: Truthful LLM Generation via Representation Flow Correction0
Large Language Models for Multi-Robot Systems: A SurveyCode1
Enhancing Hallucination Detection through Noise Injection0
The Hidden Life of Tokens: Reducing Hallucination of Large Vision-Language Models via Visual Information SteeringCode2
A Schema-Guided Reason-while-Retrieve framework for Reasoning on Scene Graphs with Large-Language-Models (LLMs)0
DAMO: Data- and Model-aware Alignment of Multi-modal LLMsCode1
Mitigating Object Hallucinations in Large Vision-Language Models via Attention Calibration0
Eliciting Language Model Behaviors with Investigator Agents0
SelfCheckAgent: Zero-Resource Hallucination Detection in Generative Large Language Models0
MJ-VIDEO: Fine-Grained Benchmarking and Rewarding Video Preferences in Video Generation0
Show:102550
← PrevPage 8 of 37Next →

No leaderboard results yet.