| VideoHallucer: Evaluating Intrinsic and Extrinsic Hallucinations in Large Video-Language Models | Jun 24, 2024 | HallucinationVideo Understanding | —Unverified | 0 |
| Evaluating and Analyzing Relationship Hallucinations in Large Vision-Language Models | Jun 24, 2024 | Common Sense ReasoningHallucination | CodeCode Available | 1 |
| Prompt-Consistency Image Generation (PCIG): A Unified Framework Integrating LLMs, Knowledge Graphs, and Controllable Diffusion Models | Jun 24, 2024 | HallucinationImage Generation | CodeCode Available | 0 |
| Semantic Entropy Probes: Robust and Cheap Hallucination Detection in LLMs | Jun 22, 2024 | HallucinationUncertainty Quantification | CodeCode Available | 2 |
| Evaluating RAG-Fusion with RAGElo: an Automated Elo-based Framework | Jun 20, 2024 | HallucinationQuestion Answering | CodeCode Available | 2 |
| Does Object Grounding Really Reduce Hallucination of Large Vision-Language Models? | Jun 20, 2024 | Caption GenerationHallucination | —Unverified | 0 |
| From Descriptive Richness to Bias: Unveiling the Dark Side of Generative Image Caption Enrichment | Jun 20, 2024 | DescriptiveHallucination | —Unverified | 0 |
| HIGHT: Hierarchical Graph Tokenization for Molecule-Language Alignment | Jun 20, 2024 | Graph Neural NetworkHallucination | —Unverified | 0 |
| Large Language Models are Skeptics: False Negative Problem of Input-conflicting Hallucination | Jun 20, 2024 | Hallucination | —Unverified | 0 |
| Rethinking Abdominal Organ Segmentation (RAOS) in the clinical scenario: A robustness evaluation benchmark with challenging cases | Jun 19, 2024 | 8kHallucination | CodeCode Available | 2 |