SOTAVerified

Hallucination

Papers

Showing 11511175 of 1816 papers

TitleStatusHype
On the Universal Truthfulness Hyperplane Inside LLMsCode0
Lynx: An Open Source Hallucination Evaluation Model0
Knowledge Overshadowing Causes Amalgamated Hallucination in Large Language Models0
Learning with Instance-Dependent Noisy Labels by Anchor Hallucination and Hard Sample Label Correction0
Fuse, Reason and Verify: Geometry Problem Solving with Parsed Clauses from Diagram0
GTP-4o: Modality-prompted Heterogeneous Graph Learning for Omni-modal Biomedical Representation0
KG-FPQ: Evaluating Factuality Hallucination in LLMs with Knowledge Graph-based False Premise QuestionsCode0
Vision-Language Models under Cultural and Inclusive Considerations0
Enhancing Hallucination Detection through Perturbation-Based Synthetic Data Generation in System ResponsesCode0
VideoCoT: A Video Chain-of-Thought Dataset with Active Annotation Tool0
Code Hallucination0
Query-Guided Self-Supervised Summarization of Nursing Notes0
Zero-shot Persuasive Chatbots with LLM-Generated Strategies and Information Retrieval0
Hallucination Detection: Robustly Discerning Reliable Answers in Large Language Models0
STOC-TOT: Stochastic Tree-of-Thought with Constrained Decoding for Complex Reasoning in Multi-Hop Question Answering0
Classification-Based Automatic HDL Code Generation Using LLMs0
FSM: A Finite State Machine Based Zero-Shot Prompting Paradigm for Multi-Hop Question Answering0
A Comparative Study of DSL Code Generation: Fine-Tuning vs. Optimized Retrieval Augmentation0
LLM Internal States Reveal Hallucination Risk Faced With a QueryCode0
Pelican: Correcting Hallucination in Vision-LLMs via Claim Decomposition and Program of Thought Verification0
Understanding Alignment in Multimodal LLMs: A Comprehensive Study0
Large Language Models Are Involuntary Truth-Tellers: Exploiting Fallacy Failure for Jailbreak AttacksCode0
The Need for Guardrails with Large Language Models in Medical Safety-Critical Settings: An Artificial Intelligence Application in the Pharmacovigilance Ecosystem0
LLM Uncertainty Quantification through Directional Entailment Graph and Claim Level Response Augmentation0
Free-text Rationale Generation under Readability Level Control0
Show:102550
← PrevPage 47 of 73Next →

No leaderboard results yet.