SOTAVerified

Hallucination

Papers

Showing 16511700 of 1816 papers

TitleStatusHype
Lifelong Neural Topic Learning in Contextualized Autoregressive Topic Models of Language via Informative Transfers0
Listening to Patients: A Framework of Detecting and Mitigating Patient Misreport for Medical Dialogue Generation0
LLLMs: A Data-Driven Survey of Evolving Research on Limitations of Large Language Models0
LLM Agents for Education: Advances and Applications0
LLM-Align: Utilizing Large Language Models for Entity Alignment in Knowledge Graphs0
INVARLLM: LLM-assisted Physical Invariant Extraction for Cyber-Physical Systems Anomaly Detection0
LLM Hallucination Reasoning with Zero-shot Knowledge Test0
LLM-Powered Agents for Navigating Venice's Historical Cadastre0
LLM-R: A Framework for Domain-Adaptive Maintenance Scheme Generation Combining Hierarchical Agents and RAG0
LLMs Can Check Their Own Results to Mitigate Hallucinations in Traffic Understanding Tasks0
LLMs can Find Mathematical Reasoning Mistakes by Pedagogical Chain-of-Thought0
LLMSeR: Enhancing Sequential Recommendation via LLM-based Data Augmentation0
LLMs Prompted for Graphs: Hallucinations and Generative Capabilities0
LLMs in the Heart of Differential Testing: A Case Study on a Medical Rule Engine0
LLMs & Legal Aid: Understanding Legal Needs Exhibited Through User Queries0
LLMs Will Always Hallucinate, and We Need to Live With This0
LLM Uncertainty Quantification through Directional Entailment Graph and Claim Level Response Augmentation0
LMOD: A Large Multimodal Ophthalmology Dataset and Benchmark for Large Vision-Language Models0
Localizing Before Answering: A Hallucination Evaluation Benchmark for Grounded Medical Multimodal LLMs0
Locate-then-Merge: Neuron-Level Parameter Fusion for Mitigating Catastrophic Forgetting in Multimodal LLMs0
Logical Consistency of Large Language Models in Fact-checking0
Look Before You Leap: An Exploratory Study of Uncertainty Measurement for Large Language Models0
Look Before You Leap: Towards Decision-Aware and Generalizable Tool-Usage for Large Language Models0
Look Within, Why LLMs Hallucinate: A Causal Perspective0
Lost in Transcription, Found in Distribution Shift: Demystifying Hallucination in Speech Foundation Models0
Lower Layer Matters: Alleviating Hallucination via Multi-Layer Fusion Contrastive Decoding with Truthfulness Refocused0
Low-hallucination Synthetic Captions for Large-Scale Vision-Language Model Pre-training0
LR-to-HR Face Hallucination with an Adversarial Progressive Attribute-Induced Network0
Luna: An Evaluation Foundation Model to Catch Language Model Hallucinations with High Accuracy and Low Cost0
Lynx: An Open Source Hallucination Evaluation Model0
M2K-VDG: Model-Adaptive Multimodal Knowledge Anchor Enhanced Video-grounded Dialogue Generation0
Machine learning techniques for the Schizophrenia diagnosis: A comprehensive review and future research directions0
Machine Mirages: Defining the Undefined0
MAC-Tuning: LLM Multi-Compositional Problem Reasoning with Enhanced Knowledge Boundary Awareness0
Magic Mushroom: A Customizable Benchmark for Fine-grained Analysis of Retrieval Noise Erosion in RAG Systems0
Magnifier Prompt: Tackling Multimodal Hallucination via Extremely Simple Instructions0
MALTO at SemEval-2024 Task 6: Leveraging Synthetic Data for LLM Hallucination Detection0
Manipulating Attributes of Natural Scenes via Hallucination0
MAPLE: Enhancing Review Generation with Multi-Aspect Prompt LEarning in Explainable Recommendation0
Map&Make: Schema Guided Text to Table Generation0
MARCO: Multi-Agent Real-time Chat Orchestration0
MASH-VLM: Mitigating Action-Scene Hallucination in Video-LLMs through Disentangled Spatial-Temporal Representations0
MASSIVE Multilingual Abstract Meaning Representation: A Dataset and Baselines for Hallucination Detection0
Maximum Hallucination Standards for Domain-Specific Large Language Models0
Meaningless is better: hashing bias-inducing words in LLM prompts improves performance in logical reasoning and statistical learning0
Measuring and Mitigating Hallucinations in Vision-Language Dataset Generation for Remote Sensing0
Measuring and Reducing LLM Hallucination without Gold-Standard Answers0
Measuring Faithfulness and Abstention: An Automated Pipeline for Evaluating LLM-Generated 3-ply Case-Based Legal Arguments0
Measuring text summarization factuality using atomic facts entailment metrics in the context of retrieval augmented generation0
Measuring the Inconsistency of Large Language Models in Preferential Ranking0
Show:102550
← PrevPage 34 of 37Next →

No leaderboard results yet.