SOTAVerified

Hallucination

Papers

Showing 501525 of 1816 papers

TitleStatusHype
Diving Deep into Modes of Fact Hallucinations in Dialogue SystemsCode0
Mitigating Hallucination in Fictional Character Role-PlayCode0
Multi-party Goal Tracking with LLMs: Comparing Pre-training, Fine-tuning, and Prompt EngineeringCode0
Addressing Topic Granularity and Hallucination in Large Language Models for Topic ModellingCode0
"Merge Conflicts!" Exploring the Impacts of External Distractors to Parametric Knowledge GraphsCode0
Optimal Transport for Unsupervised Hallucination Detection in Neural Machine TranslationCode0
MedTSS: transforming abstractive summarization of scientific articles with linguistic analysis and concept reinforcementCode0
MELO: Enhancing Model Editing with Neuron-Indexed Dynamic LoRACode0
MedScore: Factuality Evaluation of Free-Form Medical AnswersCode0
Mitigating Entity-Level Hallucination in Large Language ModelsCode0
MedHallTune: An Instruction-Tuning Benchmark for Mitigating Medical Hallucination in Vision-Language ModelsCode0
Differentially Private Steering for Large Language Model AlignmentCode0
Mechanistic Understanding and Mitigation of Language Model Non-Factual HallucinationsCode0
Beyond Ontology in Dialogue State Tracking for Goal-Oriented ChatbotCode0
MCQG-SRefine: Multiple Choice Question Generation and Evaluation with Iterative Self-Critique, Correction, and Comparison FeedbackCode0
DGoT: Dynamic Graph of Thoughts for Scientific Abstract GenerationCode0
Machine Translation Hallucination Detection for Low and High Resource Languages using Large Language ModelsCode0
MAF: Multi-Aspect Feedback for Improving Reasoning in Large Language ModelsCode0
Low to High Dimensional Modality Hallucination using Aggregated Fields of ViewCode0
LVLM-Compress-Bench: Benchmarking the Broader Impact of Large Vision-Language Model CompressionCode0
MAVEN-Fact: A Large-scale Event Factuality Detection DatasetCode0
Detecting Errors through Ensembling Prompts (DEEP): An End-to-End LLM Framework for Detecting Factual ErrorsCode0
LongHalQA: Long-Context Hallucination Evaluation for MultiModal Large Language ModelsCode0
Logic Query of Thoughts: Guiding Large Language Models to Answer Complex Logic Queries with Knowledge GraphsCode0
MCiteBench: A Multimodal Benchmark for Generating Text with CitationsCode0
Show:102550
← PrevPage 21 of 73Next →

No leaderboard results yet.