SOTAVerified

Hallucination

Papers

Showing 301350 of 1816 papers

TitleStatusHype
Citation-Enhanced Generation for LLM-based ChatbotsCode1
Circuit Transformer: A Transformer That Preserves Logical EquivalenceCode1
ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition BenchmarkCode1
EDFace-Celeb-1M: Benchmarking Face Hallucination with a Million-scale DatasetCode1
EmbodiedAgent: A Scalable Hierarchical Approach to Overcome Practical Challenge in Multi-Robot ControlCode1
Evaluating and Analyzing Relationship Hallucinations in Large Vision-Language ModelsCode1
EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language ModelsCode1
LightLM: A Lightweight Deep and Narrow Language Model for Generative RecommendationCode1
LLM Lies: Hallucinations are not Bugs, but Features as Adversarial ExamplesCode1
AssistRAG: Boosting the Potential of Large Language Models with an Intelligent Information AssistantCode1
CHATREPORT: Democratizing Sustainability Disclosure Analysis through LLM-based ToolsCode1
DomainRAG: A Chinese Benchmark for Evaluating Domain-specific Retrieval-Augmented GenerationCode1
CogniBench: A Legal-inspired Framework and Dataset for Assessing Cognitive Faithfulness of Large Language ModelsCode1
Enhancing Uncertainty-Based Hallucination Detection with Stronger FocusCode1
LiDAR-based 4D Occupancy Completion and ForecastingCode1
Aladdin: Zero-Shot Hallucination of Stylized 3D Assets from Abstract Scene DescriptionsCode1
Entity-level Factual Consistency of Abstractive Text SummarizationCode1
Collaborative Large Language Model for Recommender SystemsCode1
Distinguishing Ignorance from Error in LLM HallucinationsCode1
Doc2Query--: When Less is MoreCode1
Element-aware Summarization with Large Language Models: Expert-aligned Evaluation and Chain-of-Thought MethodCode1
Evaluating the Quality of Hallucination Benchmarks for Large Vision-Language ModelsCode1
ChartSumm: A Comprehensive Benchmark for Automatic Chart Summarization of Long and Short SummariesCode1
ChartInsighter: An Approach for Mitigating Hallucination in Time-series Chart Summary Generation with A Benchmark DatasetCode1
Investigating Hallucinations in Pruned Large Language Models for Abstractive SummarizationCode1
Exploring the Transferability of Visual Prompting for Multimodal Large Language ModelsCode1
LLM-QE: Improving Query Expansion by Aligning Large Language Models with Ranking PreferencesCode1
Extract Free Dense Misalignment from CLIPCode1
FactAlign: Long-form Factuality Alignment of Large Language ModelsCode1
Federated Recommendation via Hybrid Retrieval Augmented GenerationCode1
FAIR GPT: A virtual consultant for research data management in ChatGPTCode1
FaithBench: A Diverse Hallucination Benchmark for Summarization by Modern LLMsCode1
DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language ModelsCode1
Learning From Correctness Without Prompting Makes LLM Efficient ReasonerCode1
Learning to Automate Follow-up Question Generation using Process Knowledge for Depression Triage on Reddit PostsCode1
Detecting Hallucinated Content in Conditional Neural Sequence GenerationCode1
Detecting and Preventing Hallucinations in Large Vision Language ModelsCode1
Context-Patch Face Hallucination Based on Thresholding Locality-constrained Representation and Reproducing LearningCode1
Accuracy and Political Bias of News Source Credibility Ratings by Large Language ModelsCode1
Contrastive Learning Reduces Hallucination in ConversationsCode1
Detecting Machine-Generated Texts by Multi-Population Aware Optimization for Maximum Mean DiscrepancyCode1
Finding and Editing Multi-Modal Neurons in Pre-Trained TransformersCode1
Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI FeedbackCode1
DiffFuSR: Super-Resolution of all Sentinel-2 Multispectral Bands using Diffusion ModelsCode1
Large Language Models for Multi-Robot Systems: A SurveyCode1
Improving Large Language Models in Event Relation Logical PredictionCode1
K-QA: A Real-World Medical Q&A BenchmarkCode1
PAINT: Paying Attention to INformed Tokens to Mitigate Hallucination in Large Vision-Language ModelCode1
KoLA: Carefully Benchmarking World Knowledge of Large Language ModelsCode1
Label Hallucination for Few-Shot ClassificationCode1
Show:102550
← PrevPage 7 of 37Next →

No leaderboard results yet.