SOTAVerified

Hallucination

Papers

Showing 11511200 of 1816 papers

TitleStatusHype
Fact-Checking the Output of Large Language Models via Token-Level Uncertainty Quantification0
Effectiveness Assessment of Recent Large Vision-Language Models0
Benchmarking Hallucination in Large Language Models based on Unanswerable Math Word ProblemCode0
German also Hallucinates! Inconsistency Detection in News Summaries with the Absinth DatasetCode0
KnowAgent: Knowledge-Augmented Planning for LLM-Based AgentsCode3
InterrogateLLM: Zero-Resource Hallucination Detection in LLM-Generated AnswersCode1
The Claude 3 Model Family: Opus, Sonnet, Haiku0
Right for Right Reasons: Large Language Models for Verifiable Commonsense Knowledge Graph Question Answering0
Quantity Matters: Towards Assessing and Mitigating Number Hallucination in Large Vision-Language Models0
CR-LT-KGQA: A Knowledge Graph Question Answering Dataset Requiring Commonsense Reasoning and Long-Tail KnowledgeCode1
In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination MitigationCode2
MALTO at SemEval-2024 Task 6: Leveraging Synthetic Data for LLM Hallucination Detection0
DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language ModelsCode1
HALC: Object Hallucination Reduction via Adaptive Focal-Contrast DecodingCode2
Crimson: Empowering Strategic Reasoning in Cybersecurity through Large Language Models0
Self-Consistent Decoding for More Factual Open ResponsesCode0
Whispers that Shake Foundations: Analyzing and Mitigating False Premise Hallucinations in Large Language Models0
The All-Seeing Project V2: Towards General Relation Comprehension of the Open WorldCode4
Navigating Hallucinations for Reasoning of Unintentional Activities0
Multi-FAct: Assessing Factuality of Multilingual LLMs using FActScoreCode0
Collaborative decoding of critical tokens for boosting factuality of large language models0
All in an Aggregated Image for In-Image LearningCode1
Editing Factual Knowledge and Explanatory Ability of Medical Large Language ModelsCode0
Securing Reliability: A Brief Overview on Enhancing In-Context Learning for Foundation Models0
TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful SpaceCode2
Re-Ex: Revising after Explanation Reduces the Factual Errors in LLM ResponsesCode0
Look Before You Leap: Towards Decision-Aware and Generalizable Tool-Usage for Large Language Models0
GROUNDHOG: Grounding Large Language Models to Holistic Segmentation0
HypoTermQA: Hypothetical Terms Dataset for Benchmarking Hallucination Tendency of LLMsCode0
Rethinking Software Engineering in the Foundation Model Era: A Curated Catalogue of Challenges in the Development of Trustworthy FMware0
AVI-Talking: Learning Audio-Visual Instructions for Expressive 3D Talking Face Generation0
Detecting Machine-Generated Texts by Multi-Population Aware Optimization for Maximum Mean DiscrepancyCode1
Citation-Enhanced Generation for LLM-based ChatbotsCode1
Hal-Eval: A Universal and Fine-grained Hallucination Evaluation Framework for Large Vision Language Models0
A Data-Centric Approach To Generate Faithful and High Quality Patient Summaries with Large Language ModelsCode1
CARBD-Ko: A Contextually Annotated Review Benchmark Dataset for Aspect-Level Sentiment Classification in Korean0
Seeing is Believing: Mitigating Hallucination in Large Vision-Language Models via CLIP-Guided DecodingCode1
UFO: a Unified and Flexible Framework for Evaluating Factuality of Large Language ModelsCode0
DualFocus: Integrating Macro and Micro Perspectives in Multi-modal Large Language ModelsCode0
Visual Hallucinations of Multi-modal Large Language ModelsCode1
Less is More: Mitigating Multimodal Hallucination from an EOS Decision PerspectiveCode2
Does the Generator Mind its Contexts? An Analysis of Generative Model Faithfulness under Context Transfer0
Science Checker Reloaded: A Bidirectional Paradigm for Transparency and Logical ReasoningCode0
Emergence and dynamics of delusions and hallucinations across stages in early psychosis0
Enhanced Hallucination Detection in Neural Machine Translation through Simple Detector Aggregation0
OPDAI at SemEval-2024 Task 6: Small LLMs can Accelerate Hallucination Detection with Weakly Supervised Data0
OWSM-CTC: An Open Encoder-Only Speech Foundation Model for Speech Recognition, Translation, and Language Identification0
GOOD: Towards Domain Generalized Orientated Object Detection0
TofuEval: Evaluating Hallucinations of LLMs on Topic-Focused Dialogue SummarizationCode1
Structured Chain-of-Thought Prompting for Few-Shot Generation of Content-Grounded QA Conversations0
Show:102550
← PrevPage 24 of 37Next →

No leaderboard results yet.