| Learning on LLM Output Signatures for gray-box LLM Behavior Analysis | Mar 18, 2025 | Hallucination | CodeCode Available | 0 | 5 |
| Learning Fine-grained Domain Generalization via Hyperbolic State Space Hallucination | Apr 10, 2025 | Domain GeneralizationHallucination | CodeCode Available | 0 | 5 |
| Leveraging Pretrained Models for Automatic Summarization of Doctor-Patient Conversations | Sep 24, 2021 | Hallucination | CodeCode Available | 0 | 5 |
| LLM Internal States Reveal Hallucination Risk Faced With a Query | Jul 3, 2024 | HallucinationResponse Generation | CodeCode Available | 0 | 5 |
| MAF: Multi-Aspect Feedback for Improving Reasoning in Large Language Models | Oct 19, 2023 | HallucinationMathematical Reasoning | CodeCode Available | 0 | 5 |
| Large Language Models Are Involuntary Truth-Tellers: Exploiting Fallacy Failure for Jailbreak Attacks | Jul 1, 2024 | HallucinationLanguage Modeling | CodeCode Available | 0 | 5 |
| A Comparative Study on Language Models for Task-Oriented Dialogue Systems | Jan 21, 2022 | Dialogue State TrackingHallucination | CodeCode Available | 0 | 5 |
| Language Models Hallucinate, but May Excel at Fact Verification | Oct 23, 2023 | Fact VerificationHallucination | CodeCode Available | 0 | 5 |
| AIstorian lets AI be a historian: A KG-powered multi-agent system for accurate biography generation | Mar 14, 2025 | Abstractive Text SummarizationChunking | CodeCode Available | 0 | 5 |
| Multi-Source Knowledge Pruning for Retrieval-Augmented Generation: A Benchmark and Empirical Study | Sep 3, 2024 | BenchmarkingHallucination | CodeCode Available | 0 | 5 |