SOTAVerified

Hallucination

Papers

Showing 851900 of 1816 papers

TitleStatusHype
A Benchmark and Robustness Study of In-Context-Learning with Large Language Models in Music Entity DetectionCode0
CG-Bench: Clue-grounded Question Answering Benchmark for Long Video Understanding0
Task-Oriented Dialog Systems for the Senegalese Wolof Language0
Combating Multimodal LLM Hallucination via Bottom-Up Holistic Reasoning0
RAC3: Retrieval-Augmented Corner Case Comprehension for Autonomous Driving with Vision-Language Models0
Thinking with Knowledge Graphs: Enhancing LLM Reasoning Through Structured Data0
NoisyEQA: Benchmarking Embodied Question Answering Against Noisy Queries0
Accelerating Retrieval-Augmented Generation0
Detecting LLM Hallucination Through Layer-wise Information Deficiency: Analysis of Unanswerable Questions and Ambiguous Prompts0
Benchmarking large language models for materials synthesis: the case of atomic layer deposition0
TACOMORE: Leveraging the Potential of LLMs in Corpus-based Discourse Analysis with Prompt Engineering0
Multi-Task Learning with LLMs for Implicit Sentiment Analysis: Data-level and Task-level Automatic Weight Learning0
Hallucination Elimination and Semantic Enhancement Framework for Vision-Language Models in Traffic ScenariosCode0
HalluCana: Fixing LLM Hallucination with A Canary Lookahead0
Methods for Legal Citation Prediction in the Age of LLMs: An Australian Law Case Study0
Delve into Visual Contrastive Decoding for Hallucination Mitigation of Large Vision-Language ModelsCode0
Evaluating Hallucination in Text-to-Image Diffusion Models with Scene-Graph based Question-Answering Agent0
Multi-Objective Alignment of Large Language Models Through Hypervolume Maximization0
Steps are all you need: Rethinking STEM Education with Prompt Engineering0
LLM-Align: Utilizing Large Language Models for Entity Alignment in Knowledge Graphs0
100% Elimination of Hallucinations on RAGTruth for GPT-4 and GPT-3.5 Turbo0
Expanding Performance Boundaries of Open-Source Multimodal Models with Model, Data, and Test-Time Scaling0
TOBUGraph: Knowledge Graph-Based Retrieval for Enhanced LLM Performance Beyond RAG0
Verb Mirage: Unveiling and Assessing Verb Concept Hallucinations in Multimodal Large Language Models0
Deep priors for satellite image restoration with accurate uncertainties0
Reducing Tool Hallucination via Reliability Alignment0
GenMAC: Compositional Text-to-Video Generation with Multi-Agent Collaboration0
VidHalluc: Evaluating Temporal Hallucinations in Multimodal Large Language Models for Video Understanding0
Who Brings the Frisbee: Probing Hidden Hallucination Factors in Large Vision-Language Model via Causality Analysis0
An Evolutionary Large Language Model for Hallucination Mitigation0
CC-OCR: A Comprehensive and Challenging OCR Benchmark for Evaluating Large Multimodal Models in Literacy0
AI Benchmarks and Datasets for LLM Evaluation0
Automating Feedback Analysis in Surgical Training: Detection, Categorization, and AssessmentCode0
Beyond Logit Lens: Contextual Embeddings for Robust Hallucination Detection & Grounding in VLMs0
DHCP: Detecting Hallucinations by Cross-modal Attention Pattern in Large Vision-Language Models0
OPCap:Object-aware Prompting Captioning0
Efficient Self-Improvement in Multimodal Large Language Models: A Model-Level Judge-Free Approach0
Meaningless is better: hashing bias-inducing words in LLM prompts improves performance in logical reasoning and statistical learning0
A Topic-level Self-Correctional Approach to Mitigate Hallucinations in MLLMs0
VLRewardBench: A Challenging Benchmark for Vision-Language Generative Reward Models0
AI2T: Building Trustable AI Tutors by Interactively Teaching a Self-Aware Learning Agent0
Enhancing Multi-Agent Consensus through Third-Party LLM Integration: Analyzing Uncertainty and Mitigating Hallucinations in Large Language Models0
Ontology-Constrained Generation of Domain-Specific Clinical SummariesCode0
Leveraging LLMs for Legacy Code Modernization: Challenges and Opportunities for LLM-Generated Documentation0
Detecting Hallucinations in Virtual Histology with Neural Precursors0
ICT: Image-Object Cross-Level Trusted Intervention for Mitigating Object Hallucination in Large Vision-Language Models0
Sycophancy in Large Language Models: Causes and Mitigations0
CATCH: Complementary Adaptive Token-level Contrastive Decoding to Mitigate Hallucinations in LVLMs0
Can Open-source LLMs Enhance Data Synthesis for Toxic Detection?: An Experimental Study0
Mitigating Knowledge Conflicts in Language Model-Driven Question Answering0
Show:102550
← PrevPage 18 of 37Next →

No leaderboard results yet.