SOTAVerified

Hallucination

Papers

Showing 9511000 of 1816 papers

TitleStatusHype
Conditional Hallucinations for Image Compression0
Investigating the Role of Prompting and External Tools in Hallucination Rates of Large Language Models0
MaCTG: Multi-Agent Collaborative Thought Graph for Automatic Programming0
AVHBench: A Cross-Modal Hallucination Benchmark for Audio-Visual Large Language Models0
Multilingual Hallucination Gaps in Large Language Models0
Leveraging the Domain Adaptation of Retrieval Augmented Generation Models for Question Answering and Reducing Hallucination0
GeoCode-GPT: A Large Language Model for Geospatial Code Generation Tasks0
IPL: Leveraging Multimodal Large Language Models for Intelligent Product Listing0
SG-FSM: A Self-Guiding Zero-Shot Prompting Paradigm for Multi-Hop Question Answering Based on Finite State Machine0
Fine-Tuning Large Language Models to Appropriately Abstain with Semantic Entropy0
ProveRAG: Provenance-Driven Vulnerability Analysis with Automated Retrieval-Augmented LLMsCode0
Privacy-hardened and hallucination-resistant synthetic data generation with logic-solvers0
Navigating Noisy Feedback: Enhancing Reinforcement Learning with Error-Prone Language ModelsCode0
Do Robot Snakes Dream like Electric Sheep? Investigating the Effects of Architectural Inductive Biases on Hallucination0
Mitigating Hallucinations of Large Language Models in Medical Information Extraction via Contrastive Decoding0
Towards a Reliable Offline Personal AI Assistant for Long Duration Spaceflight0
NetSafe: Exploring the Topological Safety of Multi-agent Networks0
Large language models enabled multiagent ensemble method for efficient EHR data labeling0
Learning to Generate and Evaluate Fact-checking Explanations with Transformers0
ToW: Thoughts of Words Improve Reasoning in Large Language ModelsCode0
A Survey of Hallucination in Large Visual Language Models0
Hallucination Detox: Sensitivity Dropout (SenD) for Large Language Model Training0
Coarse-to-Fine Highlighting: Reducing Knowledge Hallucination in Large Language Models0
Explaining Graph Neural Networks with Large Language Models: A Counterfactual Perspective for Molecular Property PredictionCode0
ELOQ: Resources for Enhancing LLM Detection of Out-of-Scope QuestionsCode0
Good Parenting is all you need -- Multi-agentic LLM Hallucination Mitigation0
MCQG-SRefine: Multiple Choice Question Generation and Evaluation with Iterative Self-Critique, Correction, and Comparison FeedbackCode0
ETF: An Entity Tracing Framework for Hallucination Detection in Code Summaries0
From Single to Multi: How LLMs Hallucinate in Multi-Document SummarizationCode0
Utilizing Large Language Models in an iterative paradigm with domain feedback for zero-shot molecule optimization0
On A Scale From 1 to 5: Quantifying Hallucination in Faithfulness Evaluation0
What Do LLMs Need to Understand Graphs: A Survey of Parametric Representation of Graphs0
Iter-AHMCL: Alleviate Hallucination for Large Language Model via Iterative Model-level Contrastive Learning0
A Claim Decomposition Benchmark for Long-form Answer VerificationCode0
RosePO: Aligning LLM-based Recommenders with Human Values0
When Not to Answer: Evaluating Prompts on GPT Models for Effective Abstention in Unanswerable Math Word Problems0
Controlled Automatic Task-Specific Synthetic Data Generation for Hallucination Detection0
AGENTiGraph: An Interactive Knowledge Graph Platform for LLM-based Chatbots Utilizing Private Data0
On the Capacity of Citation Generation by Large Language Models0
ReDeEP: Detecting Hallucination in Retrieval-Augmented Generation via Mechanistic Interpretability0
LargePiG: Your Large Language Model is Secretly a Pointer Generator0
Automatically Generating Visual Hallucination Test Cases for Multimodal Large Language ModelsCode0
Magnifier Prompt: Tackling Multimodal Hallucination via Extremely Simple Instructions0
Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs0
Can Structured Data Reduce Epistemic Uncertainty?0
Parenting: Optimizing Knowledge Selection of Retrieval-Augmented Language Models with Parameter Decoupling and Tailored Tuning0
SkillAggregation: Reference-free LLM-Dependent Aggregation0
Medico: Towards Hallucination Detection and Correction with Multi-source Evidence Fusion0
Honest AI: Fine-Tuning "Small" Language Models to Say "I Don't Know", and Reducing Hallucination in RAG0
Collu-Bench: A Benchmark for Predicting Language Model Hallucinations in Code0
Show:102550
← PrevPage 20 of 37Next →

No leaderboard results yet.