SOTAVerified

Hallucination

Papers

Showing 9511000 of 1816 papers

TitleStatusHype
Robustness Assessment of Mathematical Reasoning in the Presence of Missing and Contradictory Conditions0
3D-GRAND: A Million-Scale Dataset for 3D-LLMs with Better Grounding and Less HallucinationCode2
Chaos with Keywords: Exposing Large Language Models Sycophantic Hallucination to Misleading Keywords and Evaluating Defense Strategies0
ActionReasoningBench: Reasoning about Actions with and without Ramification Constraints0
Confabulation: The Surprising Value of Large Language Model Hallucinations0
Towards Detecting LLMs Hallucination via Markov Chain-based Multi-agent Debate Framework0
Analyzing LLM Behavior in Dialogue Summarization: Unveiling Circumstantial Hallucination Trends0
OTTAWA: Optimal TransporT Adaptive Word Aligner for Hallucination and Omission Translation Errors DetectionCode0
Enhancing Trust in LLMs: Algorithms for Comparing and Interpreting LLMs0
CODE: Contrasting Self-generated Description to Combat Hallucination in Large Multi-modal Models0
How to Explore with Belief: State Entropy Maximization in POMDPs0
Ask-EDA: A Design Assistant Empowered by LLM, Hybrid RAG and Abbreviation De-hallucination0
Large Language Model Assisted Optimal Bidding of BESS in FCAS Market: An AI-agent based Approach0
Decompose, Enrich, and Extract! Schema-aware Event Extraction using LLMs0
Luna: An Evaluation Foundation Model to Catch Language Model Hallucinations with High Accuracy and Low Cost0
Comprehensive Evaluation of Large Language Models for Topic Modeling0
DAFNet: Dynamic Auxiliary Fusion for Sequential Model Editing in Large Language ModelsCode0
Enhancing Noise Robustness of Retrieval-Augmented Language Models with Adaptive Adversarial TrainingCode1
Similarity is Not All You Need: Endowing Retrieval Augmented Generation with Multi Layered Thoughts0
Hallucination-Free? Assessing the Reliability of Leading AI Legal Research Tools0
ANAH: Analytical Annotation of Hallucinations in Large Language ModelsCode2
NoiseBoost: Alleviating Hallucination with Noise Perturbation for Multimodal Large Language ModelsCode0
MetaToken: Detecting Hallucination in Image Descriptions by Meta Classification0
MASSIVE Multilingual Abstract Meaning Representation: A Dataset and Baselines for Hallucination Detection0
Two-Layer Retrieval-Augmented Generation Framework for Low-Resource Medical Question Answering Using Reddit Data: Proof-of-Concept Study0
Personalized Steering of Large Language Models: Versatile Steering Vectors Through Bi-directional Preference OptimizationCode1
LLMs and Memorization: On Quality and Specificity of Copyright ComplianceCode0
Data-augmented phrase-level alignment for mitigating object hallucination0
RITUAL: Random Image Transformations as a Universal Anti-hallucination Lever in Large Vision Language Models0
Conv-CoA: Improving Open-domain Question Answering in Large Language Models via Conversational Chain-of-Action0
TimeChara: Evaluating Point-in-Time Character Hallucination of Role-Playing Large Language ModelsCode1
RLAIF-V: Open-Source AI Feedback Leads to Super GPT-4V TrustworthinessCode11
Laboratory-Scale AI: Open-Weight Models are Competitive with ChatGPT Even in Low-Resource Settings0
Think Before You Act: A Two-Stage Framework for Mitigating Gender Bias Towards Vision-Language TasksCode0
GeneAgent: Self-verification Language Agent for Gene Set Knowledge Discovery using Domain Databases0
Large Language Model Pruning0
Enhancing Visual-Language Modality Alignment in Large Vision Language Models via Self-ImprovementCode2
CHARP: Conversation History AwaReness Probing for Knowledge-grounded Dialogue Systems0
Scaling Laws for Discriminative Classification in Large Language Models0
DEEM: Diffusion Models Serve as the Eyes of Large Language Models for Image PerceptionCode1
Alleviating Hallucinations in Large Vision-Language Models through Hallucination-Induced OptimizationCode0
Visual Description Grounding Reduces Hallucinations and Boosts Reasoning in LVLMsCode1
Calibrated Self-Rewarding Vision Language ModelsCode2
RefChecker: Reference-based Fine-grained Hallucination Checker and Benchmark for Large Language ModelsCode3
WISE: Rethinking the Knowledge Memory for Lifelong Model Editing of Large Language Models0
Less for More: Enhanced Feedback-aligned Mixed LLMs for Molecule Caption Generation and Fine-Grained NLI Evaluation0
Gradient Projection For Continual Parameter-Efficient Tuning0
CrossCheckGPT: Universal Hallucination Ranking for Multimodal Foundation Models0
GameVLM: A Decision-making Framework for Robotic Task Planning Based on Visual Language Models and Zero-sum Games0
Presentations are not always linear! GNN meets LLM for Document-to-Presentation Transformation with Attribution0
Show:102550
← PrevPage 20 of 37Next →

No leaderboard results yet.