SOTAVerified

Hallucination

Papers

Showing 301325 of 1816 papers

TitleStatusHype
Distinguishing Ignorance from Error in LLM HallucinationsCode1
Large Language Models for Multi-Robot Systems: A SurveyCode1
Learning to Automate Follow-up Question Generation using Process Knowledge for Depression Triage on Reddit PostsCode1
AssistRAG: Boosting the Potential of Large Language Models with an Intelligent Information AssistantCode1
CHATREPORT: Democratizing Sustainability Disclosure Analysis through LLM-based ToolsCode1
ECBench: Can Multi-modal Foundation Models Understand the Egocentric World? A Holistic Embodied Cognition BenchmarkCode1
Element-aware Summarization with Large Language Models: Expert-aligned Evaluation and Chain-of-Thought MethodCode1
LAN-HDR: Luminance-based Alignment Network for High Dynamic Range Video ReconstructionCode1
DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language ModelsCode1
Efficient Dynamic Clustering-Based Document Compression for Retrieval-Augmented-GenerationCode1
Large Language Models are Versatile Decomposers: Decompose Evidence and Questions for Table-based ReasoningCode1
Detecting Machine-Generated Texts by Multi-Population Aware Optimization for Maximum Mean DiscrepancyCode1
DiffFuSR: Super-Resolution of all Sentinel-2 Multispectral Bands using Diffusion ModelsCode1
EmbodiedAgent: A Scalable Hierarchical Approach to Overcome Practical Challenge in Multi-Robot ControlCode1
ChartSumm: A Comprehensive Benchmark for Automatic Chart Summarization of Long and Short SummariesCode1
ChartInsighter: An Approach for Mitigating Hallucination in Time-series Chart Summary Generation with A Benchmark DatasetCode1
Label Hallucination for Few-Shot ClassificationCode1
Collaborative Large Language Model for Recommender SystemsCode1
Accuracy and Political Bias of News Source Credibility Ratings by Large Language ModelsCode1
Improving Large Language Models in Event Relation Logical PredictionCode1
LLM Lies: Hallucinations are not Bugs, but Features as Adversarial ExamplesCode1
Entity-level Factual Consistency of Abstractive Text SummarizationCode1
Deficiency-Aware Masked Transformer for Video InpaintingCode1
Evaluating and Analyzing Relationship Hallucinations in Large Vision-Language ModelsCode1
Detecting and Mitigating Hallucination in Large Vision Language Models via Fine-Grained AI FeedbackCode1
Show:102550
← PrevPage 13 of 73Next →

No leaderboard results yet.