SOTAVerified

Hallucination

Papers

Showing 11511200 of 1816 papers

TitleStatusHype
On the Universal Truthfulness Hyperplane Inside LLMsCode0
Lynx: An Open Source Hallucination Evaluation Model0
Knowledge Overshadowing Causes Amalgamated Hallucination in Large Language Models0
Learning with Instance-Dependent Noisy Labels by Anchor Hallucination and Hard Sample Label Correction0
Fuse, Reason and Verify: Geometry Problem Solving with Parsed Clauses from Diagram0
GTP-4o: Modality-prompted Heterogeneous Graph Learning for Omni-modal Biomedical Representation0
KG-FPQ: Evaluating Factuality Hallucination in LLMs with Knowledge Graph-based False Premise QuestionsCode0
Vision-Language Models under Cultural and Inclusive Considerations0
Enhancing Hallucination Detection through Perturbation-Based Synthetic Data Generation in System ResponsesCode0
VideoCoT: A Video Chain-of-Thought Dataset with Active Annotation Tool0
Code Hallucination0
Query-Guided Self-Supervised Summarization of Nursing Notes0
Zero-shot Persuasive Chatbots with LLM-Generated Strategies and Information Retrieval0
Hallucination Detection: Robustly Discerning Reliable Answers in Large Language Models0
STOC-TOT: Stochastic Tree-of-Thought with Constrained Decoding for Complex Reasoning in Multi-Hop Question Answering0
Classification-Based Automatic HDL Code Generation Using LLMs0
FSM: A Finite State Machine Based Zero-Shot Prompting Paradigm for Multi-Hop Question Answering0
A Comparative Study of DSL Code Generation: Fine-Tuning vs. Optimized Retrieval Augmentation0
LLM Internal States Reveal Hallucination Risk Faced With a QueryCode0
Pelican: Correcting Hallucination in Vision-LLMs via Claim Decomposition and Program of Thought Verification0
Understanding Alignment in Multimodal LLMs: A Comprehensive Study0
Large Language Models Are Involuntary Truth-Tellers: Exploiting Fallacy Failure for Jailbreak AttacksCode0
The Need for Guardrails with Large Language Models in Medical Safety-Critical Settings: An Artificial Intelligence Application in the Pharmacovigilance Ecosystem0
LLM Uncertainty Quantification through Directional Entailment Graph and Claim Level Response Augmentation0
Free-text Rationale Generation under Readability Level Control0
Unveiling Glitches: A Deep Dive into Image Encoding Bugs within CLIP0
A Study on Effect of Reference Knowledge Choice in Generating Technical Content Relevant to SAPPhIRE Model Using Large Language Model0
BioKGBench: A Knowledge Graph Checking Benchmark of AI Agent for Biomedical ScienceCode0
PFME: A Modular Approach for Fine-grained Hallucination Detection and Editing of Large Language Models0
Applying RLAIF for Code Generation with API-usage in Lightweight LLMs0
Handling Ontology Gaps in Semantic ParsingCode0
From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic DataCode0
Mitigating Hallucination in Fictional Character Role-PlayCode0
VideoHallucer: Evaluating Intrinsic and Extrinsic Hallucinations in Large Video-Language Models0
Prompt-Consistency Image Generation (PCIG): A Unified Framework Integrating LLMs, Knowledge Graphs, and Controllable Diffusion ModelsCode0
Large Language Models are Skeptics: False Negative Problem of Input-conflicting Hallucination0
HIGHT: Hierarchical Graph Tokenization for Molecule-Language Alignment0
Does Object Grounding Really Reduce Hallucination of Large Vision-Language Models?0
From Descriptive Richness to Bias: Unveiling the Dark Side of Generative Image Caption Enrichment0
StackRAG Agent: Improving Developer Answers with Retrieval-Augmented GenerationCode0
What Matters in Memorizing and Recalling Facts? Multifaceted Benchmarks for Knowledge Probing in Language Models0
Detecting Errors through Ensembling Prompts (DEEP): An End-to-End LLM Framework for Detecting Factual ErrorsCode0
RichRAG: Crafting Rich Responses for Multi-faceted Queries in Retrieval-Augmented Generation0
On-Policy Fine-grained Knowledge Feedback for Hallucination MitigationCode0
Beyond Under-Alignment: Atomic Preference Enhanced Factuality Tuning for Large Language Models0
Do More Details Always Introduce More Hallucinations in LVLM-based Image Captioning?0
Counterfactual Debating with Preset Stances for Hallucination Elimination of LLMsCode0
CoMT: Chain-of-Medical-Thought Reduces Hallucination in Medical Report Generation0
Mitigating Large Language Model Hallucination with Faithful Finetuning0
InternalInspector I^2: Robust Confidence Estimation in LLMs through Internal States0
Show:102550
← PrevPage 24 of 37Next →

No leaderboard results yet.