| LLM Hallucination Reasoning with Zero-shot Knowledge Test | Nov 14, 2024 | Hallucination | —Unverified | 0 |
| DAHL: Domain-specific Automated Hallucination Evaluation of Long-Form Text through a Benchmark Dataset in Biomedicine | Nov 14, 2024 | FormHallucination | CodeCode Available | 0 |
| On the Limits of Language Generation: Trade-Offs Between Hallucination and Mode Collapse | Nov 14, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| Bridging the Visual Gap: Fine-Tuning Multimodal Models with Knowledge-Adapted Captions | Nov 13, 2024 | DescriptiveHallucination | CodeCode Available | 0 |
| Confidence-aware Denoised Fine-tuning of Off-the-shelf Models for Certified Robustness | Nov 13, 2024 | Adversarial RobustnessDenoising | CodeCode Available | 0 |
| Verbosity Veracity: Demystify Verbosity Compensation Behavior of Large Language Models | Nov 12, 2024 | Hallucination | CodeCode Available | 0 |
| Trustful LLMs: Customizing and Grounding Text Generation with Knowledge Bases and Dual Decoders | Nov 12, 2024 | DecoderHallucination | —Unverified | 0 |
| DecoPrompt : Decoding Prompts Reduces Hallucinations when Large Language Models Meet False Premises | Nov 12, 2024 | Hallucination | CodeCode Available | 0 |
| SHARP: Unlocking Interactive Hallucination via Stance Transfer in Role-Playing Agents | Nov 12, 2024 | General KnowledgeHallucination | —Unverified | 0 |
| Evaluating the Accuracy of Chatbots in Financial Literature | Nov 11, 2024 | ChatbotHallucination | —Unverified | 0 |