SOTAVerified

Hallucination

Papers

Showing 14761500 of 1816 papers

TitleStatusHype
Evaluating ChatGPT as a Question Answering System: A Comprehensive Analysis and Comparison with Existing Models0
Context Tuning for Retrieval Augmented Generation0
DelucionQA: Detecting Hallucinations in Domain-specific Question Answering0
HALO: An Ontology for Representing and Categorizing Hallucinations in Large Language Models0
Behind the Magic, MERLIM: Multi-modal Evaluation Benchmark for Large Image-Language ModelsCode0
On Exploring the Reasoning Capability of Large Language Models with Knowledge Graphs0
How to Build an AI Tutor That Can Adapt to Any Course Using Knowledge Graph-Enhanced Retrieval-Augmented Generation (KG-RAG)0
Understanding Your Agent: Leveraging Large Language Models for Behavior Explanation0
Combating the "Sameness" in AI Art: Reflections on the Interactive AI Installation Fencing Hallucination0
Mitigating Hallucination in Visual Language Models with Visual Supervision0
Deficiency of Large Language Models in Finance: An Empirical Examination of Hallucination0
Calibrated Language Models Must Hallucinate0
Controlling Large Language Model-based Agents for Large-Scale Decision-Making: An Actor-Critic Approach0
Challenges of Large Language Models for Mental Health Counseling0
Minimizing Factual Inconsistency and Hallucination in Large Language Models0
Mitigating Large Language Model Hallucinations via Autonomous Knowledge Graph-based Retrofitting0
KNVQA: A Benchmark for evaluation knowledge-based VQA0
Adapting LLMs for Efficient, Personalized Information Retrieval: Methods and Implications0
Control in Hybrid Chatbots0
GPT-4V(ision) for Robotics: Multimodal Task Planning from Human Demonstration0
Chain of Visual Perception: Harnessing Multimodal Large Language Models for Zero-shot Camouflaged Object DetectionCode0
Journey of Hallucination-minimized Generative AI Solutions for Financial Decision Makers0
Crafting In-context Examples according to LMs' Parametric KnowledgeCode0
Deceptive Semantic Shortcuts on Reasoning Chains: How Far Can Models Go without Hallucination?Code0
How Trustworthy are Open-Source LLMs? An Assessment under Malicious Demonstrations Shows their VulnerabilitiesCode0
Show:102550
← PrevPage 60 of 73Next →

No leaderboard results yet.