SOTAVerified

Hallucination

Papers

Showing 601650 of 1816 papers

TitleStatusHype
A Perspective for Adapting Generalist AI to Specialized Medical AI Applications and Their Challenges0
A Debate-Driven Experiment on LLM Hallucinations and Accuracy0
Conditional Hallucinations for Image Compression0
TimeSuite: Improving MLLMs for Long Video Understanding via Grounded TuningCode2
Investigating the Role of Prompting and External Tools in Hallucination Rates of Large Language Models0
MaCTG: Multi-Agent Collaborative Thought Graph for Automatic Programming0
Multilingual Hallucination Gaps in Large Language Models0
AVHBench: A Cross-Modal Hallucination Benchmark for Audio-Visual Large Language Models0
Leveraging the Domain Adaptation of Retrieval Augmented Generation Models for Question Answering and Reducing Hallucination0
ProveRAG: Provenance-Driven Vulnerability Analysis with Automated Retrieval-Augmented LLMsCode0
Do Robot Snakes Dream like Electric Sheep? Investigating the Effects of Architectural Inductive Biases on Hallucination0
GeoCode-GPT: A Large Language Model for Geospatial Code Generation Tasks0
Privacy-hardened and hallucination-resistant synthetic data generation with logic-solvers0
IPL: Leveraging Multimodal Large Language Models for Intelligent Product Listing0
Fine-Tuning Large Language Models to Appropriately Abstain with Semantic Entropy0
Navigating Noisy Feedback: Enhancing Reinforcement Learning with Error-Prone Language ModelsCode0
SG-FSM: A Self-Guiding Zero-Shot Prompting Paradigm for Multi-Hop Question Answering Based on Finite State Machine0
Large language models enabled multiagent ensemble method for efficient EHR data labeling0
Towards a Reliable Offline Personal AI Assistant for Long Duration Spaceflight0
Mitigating Hallucinations of Large Language Models in Medical Information Extraction via Contrastive Decoding0
Mitigating Object Hallucination via Concentric Causal AttentionCode2
ToW: Thoughts of Words Improve Reasoning in Large Language ModelsCode0
Can Knowledge Editing Really Correct Hallucinations?Code1
Reducing Hallucinations in Vision-Language Models via Latent Space SteeringCode2
NetSafe: Exploring the Topological Safety of Multi-agent Networks0
Learning to Generate and Evaluate Fact-checking Explanations with Transformers0
A Survey of Hallucination in Large Visual Language Models0
Hallucination Detox: Sensitivity Dropout (SenD) for Large Language Model Training0
Explaining Graph Neural Networks with Large Language Models: A Counterfactual Perspective for Molecular Property PredictionCode0
Coarse-to-Fine Highlighting: Reducing Knowledge Hallucination in Large Language Models0
Good Parenting is all you need -- Multi-agentic LLM Hallucination Mitigation0
ELOQ: Resources for Enhancing LLM Detection of Out-of-Scope QuestionsCode0
Paths-over-Graph: Knowledge Graph Empowered Large Language Model ReasoningCode1
ETF: An Entity Tracing Framework for Hallucination Detection in Code Summaries0
From Single to Multi: How LLMs Hallucinate in Multi-Document SummarizationCode0
MCQG-SRefine: Multiple Choice Question Generation and Evaluation with Iterative Self-Critique, Correction, and Comparison FeedbackCode0
Mitigating Hallucinations in Large Vision-Language Models via Summary-Guided DecodingCode1
Utilizing Large Language Models in an iterative paradigm with domain feedback for zero-shot molecule optimization0
FaithBench: A Diverse Hallucination Benchmark for Summarization by Modern LLMsCode1
RosePO: Aligning LLM-based Recommenders with Human Values0
MMed-RAG: Versatile Multimodal RAG System for Medical Vision Language ModelsCode3
On A Scale From 1 to 5: Quantifying Hallucination in Faithfulness Evaluation0
When Not to Answer: Evaluating Prompts on GPT Models for Effective Abstention in Unanswerable Math Word Problems0
Iter-AHMCL: Alleviate Hallucination for Large Language Model via Iterative Model-level Contrastive Learning0
Graph-constrained Reasoning: Faithful Reasoning on Knowledge Graphs with Large Language ModelsCode3
What Do LLMs Need to Understand Graphs: A Survey of Parametric Representation of Graphs0
Controlled Automatic Task-Specific Synthetic Data Generation for Hallucination Detection0
A Claim Decomposition Benchmark for Long-form Answer VerificationCode0
The Curse of Multi-Modalities: Evaluating Hallucinations of Large Multimodal Models across Language, Visual, and AudioCode3
Search Engines in an AI Era: The False Promise of Factual and Verifiable Source-Cited ResponsesCode1
Show:102550
← PrevPage 13 of 37Next →

No leaderboard results yet.