SOTAVerified

Hallucination

Papers

Showing 551575 of 1816 papers

TitleStatusHype
Beyond Words: On Large Language Models Actionability in Mission-Critical Risk Analysis0
A Debate-Driven Experiment on LLM Hallucinations and Accuracy0
Beyond Under-Alignment: Atomic Preference Enhanced Factuality Tuning for Large Language Models0
Beyond the Black Box: Interpretability of LLMs in Finance0
DiDOTS: Knowledge Distillation from Large-Language-Models for Dementia Obfuscation in Transcribed Speech0
An Automated Reinforcement Learning Reward Design Framework with Large Language Model for Cooperative Platoon Coordination0
Cost-Effective Hallucination Detection for LLMs0
DHCP: Detecting Hallucinations by Cross-modal Attention Pattern in Large Vision-Language Models0
Beyond Logit Lens: Contextual Embeddings for Robust Hallucination Detection & Grounding in VLMs0
Anatomy of Industrial Scale Multilingual ASR0
Improving Zero-Shot ObjectNav with Generative Communication0
FRAME: Evaluating Rationale-Label Consistency Metrics for Free-Text Rationales0
Free-text Rationale Generation under Readability Level Control0
Developing a Reliable, Fast, General-Purpose Hallucination Detection and Mitigation Service0
LLMAuditor: A Framework for Auditing Large Language Models Using Human-in-the-Loop0
Detection and Mitigation of Hallucination in Large Reasoning Models: A Mechanistic Perspective0
An Analysis of Decoding Methods for LLM-based Agents for Faithful Multi-Hop Question Answering0
Detecting LLM Hallucination Through Layer-wise Information Deficiency: Analysis of Unanswerable Questions and Ambiguous Prompts0
Detecting LLM Fact-conflicting Hallucinations Enhanced by Temporal-logic-based Reasoning0
Detecting Hallucinations in Virtual Histology with Neural Precursors0
Detecting Hallucination and Coverage Errors in Retrieval Augmented Generation for Controversial Topics0
Benchmarking Retrieval-Augmented Large Language Models in Biomedical NLP: Application, Robustness, and Self-Awareness0
Addressing Image Hallucination in Text-to-Image Generation through Factual Image Retrieval0
Detecting Buggy Contracts via Smart Testing0
Analyzing LLM Behavior in Dialogue Summarization: Unveiling Circumstantial Hallucination Trends0
Show:102550
← PrevPage 23 of 73Next →

No leaderboard results yet.