SOTAVerified

Hallucination

Papers

Showing 17511800 of 1816 papers

TitleStatusHype
Editing Factual Knowledge and Explanatory Ability of Medical Large Language ModelsCode0
Visually Dehallucinative Instruction Generation: Know What You Don't KnowCode0
HALOS: Hallucination-free Organ Segmentation after Organ Resection SurgeryCode0
An Investigation of Evaluation Metrics for Automated Medical Note GenerationCode0
Teacher-Student Adversarial Depth Hallucination to Improve Face RecognitionCode0
HaloScope: Harnessing Unlabeled LLM Generations for Hallucination DetectionCode0
Rate, Explain and Cite (REC): Enhanced Explanation and Attribution in Automatic Evaluation by Large Language ModelsCode0
HALO: Hallucination Analysis and Learning Optimization to Empower LLMs with Retrieval-Augmented Context for Guided Clinical Decision MakingCode0
An Inflectional Database for GitksanCode0
HalluShift: Measuring Distribution Shifts towards Hallucination Detection in LLMsCode0
HalluDial: A Large-Scale Benchmark for Automatic Dialogue-Level Hallucination EvaluationCode0
A Comparative Study on Language Models for Task-Oriented Dialogue SystemsCode0
Characterizing Context Influence and Hallucination in SummarizationCode0
DoG-Instruct: Towards Premium Instruction-Tuning Data via Text-Grounded Instruction WrappingCode0
Zero-Resource Hallucination Prevention for Large Language ModelsCode0
Tensor feature hallucination for few-shot learningCode0
CHAIR -- Classifier of Hallucination as ImproverCode0
Chainpoll: A high efficacy method for LLM hallucination detectionCode0
TUBench: Benchmarking Large Vision-Language Models on Trustworthiness with Unanswerable QuestionsCode0
HALLUCINOGEN: A Benchmark for Evaluating Object Hallucination in Large Visual-Language ModelsCode0
VLM4Bio: A Benchmark Dataset to Evaluate Pretrained Vision-Language Models for Trait Discovery from Biological ImagesCode0
Reducing Quantity Hallucinations in Abstractive SummarizationCode0
ANHALTEN: Cross-Lingual Transfer for German Token-Level Reference-Free Hallucination DetectionCode0
HalluciNet-ing Spatiotemporal Representations Using a 2D-CNNCode0
Re-Ex: Revising after Explanation Reduces the Factual Errors in LLM ResponsesCode0
Hallucination Reduction in Long Input Text SummarizationCode0
DualFocus: Integrating Macro and Micro Perspectives in Multi-modal Large Language ModelsCode0
Hallucination, Monofacts, and Miscalibration: An Empirical InvestigationCode0
UCSC at SemEval-2025 Task 3: Context, Models and Prompt Optimization for Automated Hallucination Detection in LLM OutputCode0
UFO: a Unified and Flexible Framework for Evaluating Factuality of Large Language ModelsCode0
Hallucination Mitigation using Agentic AI Natural Language-Based FrameworksCode0
Hallucination Mitigation Prompts Long-term Video UnderstandingCode0
Hallucination In Object Detection -- A Study In Visual Part VerificationCode0
Regression is all you need for medical image translationCode0
THaMES: An End-to-End Tool for Hallucination Mitigation and Evaluation in Large Language ModelsCode0
DO-RAG: A Domain-Specific QA Framework Using Knowledge Graph-Enhanced Retrieval-Augmented GenerationCode0
Do Language Models Know When They're Hallucinating References?Code0
Chain-of-Verification Reduces Hallucination in Large Language ModelsCode0
A New Benchmark and Reverse Validation Method for Passage-level Hallucination DetectionCode0
The Curious Case of Hallucinations in Neural Machine TranslationCode0
Reliable Academic Conference Question Answering: A Study Based on Large Language ModelCode0
Diving Deep into Modes of Fact Hallucinations in Dialogue SystemsCode0
Hallucination Elimination and Semantic Enhancement Framework for Vision-Language Models in Traffic ScenariosCode0
Analogy Models for Neural Word InflectionCode0
A Methodology for Evaluating RAG Systems: A Case Study On Configuration Dependency ValidationCode0
Differentially Private Steering for Large Language Model AlignmentCode0
Reproducing sensory induced hallucinations via neural fieldsCode0
DGoT: Dynamic Graph of Thoughts for Scientific Abstract GenerationCode0
Hallucination Detection in LLMs: Fast and Memory-Efficient Fine-Tuned ModelsCode0
Hallucination Benchmark in Medical Visual Question AnsweringCode0
Show:102550
← PrevPage 36 of 37Next →

No leaderboard results yet.