SOTAVerified

Hallucination

Papers

Showing 601625 of 1816 papers

TitleStatusHype
A Perspective for Adapting Generalist AI to Specialized Medical AI Applications and Their Challenges0
A Debate-Driven Experiment on LLM Hallucinations and Accuracy0
MaCTG: Multi-Agent Collaborative Thought Graph for Automatic Programming0
Conditional Hallucinations for Image Compression0
Investigating the Role of Prompting and External Tools in Hallucination Rates of Large Language Models0
TimeSuite: Improving MLLMs for Long Video Understanding via Grounded TuningCode2
Multilingual Hallucination Gaps in Large Language Models0
Leveraging the Domain Adaptation of Retrieval Augmented Generation Models for Question Answering and Reducing Hallucination0
AVHBench: A Cross-Modal Hallucination Benchmark for Audio-Visual Large Language Models0
ProveRAG: Provenance-Driven Vulnerability Analysis with Automated Retrieval-Augmented LLMsCode0
Privacy-hardened and hallucination-resistant synthetic data generation with logic-solvers0
Do Robot Snakes Dream like Electric Sheep? Investigating the Effects of Architectural Inductive Biases on Hallucination0
GeoCode-GPT: A Large Language Model for Geospatial Code Generation Tasks0
IPL: Leveraging Multimodal Large Language Models for Intelligent Product Listing0
Navigating Noisy Feedback: Enhancing Reinforcement Learning with Error-Prone Language ModelsCode0
SG-FSM: A Self-Guiding Zero-Shot Prompting Paradigm for Multi-Hop Question Answering Based on Finite State Machine0
Fine-Tuning Large Language Models to Appropriately Abstain with Semantic Entropy0
Towards a Reliable Offline Personal AI Assistant for Long Duration Spaceflight0
Large language models enabled multiagent ensemble method for efficient EHR data labeling0
ToW: Thoughts of Words Improve Reasoning in Large Language ModelsCode0
Mitigating Object Hallucination via Concentric Causal AttentionCode2
Mitigating Hallucinations of Large Language Models in Medical Information Extraction via Contrastive Decoding0
Learning to Generate and Evaluate Fact-checking Explanations with Transformers0
Can Knowledge Editing Really Correct Hallucinations?Code1
NetSafe: Exploring the Topological Safety of Multi-agent Networks0
Show:102550
← PrevPage 25 of 73Next →

No leaderboard results yet.