| Unsupervised Real-Time Hallucination Detection based on the Internal States of Large Language Models | Mar 11, 2024 | Hallucination | CodeCode Available | 2 |
| On the Benefits of Fine-Grained Loss Truncation: A Case Study on Factuality in Summarization | Mar 9, 2024 | HallucinationText Summarization | CodeCode Available | 0 |
| Tuning-Free Accountable Intervention for LLM Deployment -- A Metacognitive Approach | Mar 8, 2024 | Decision MakingHallucination | —Unverified | 0 |
| Can Large Language Models Play Games? A Case Study of A Self-Play Approach | Mar 8, 2024 | Decision MakingHallucination | —Unverified | 0 |
| ERBench: An Entity-Relationship based Automatically Verifiable Hallucination Benchmark for Large Language Models | Mar 8, 2024 | AttributeHallucination | CodeCode Available | 0 |
| ChatASU: Evoking LLM's Reflexion to Truly Understand Aspect Sentiment in Dialogues | Mar 8, 2024 | HallucinationQuestion Answering | —Unverified | 0 |
| Sora as an AGI World Model? A Complete Survey on Text-to-Video Generation | Mar 8, 2024 | ArticlesHallucination | —Unverified | 0 |
| RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Horizon Generation | Mar 8, 2024 | Code GenerationHallucination | CodeCode Available | 3 |
| HaluEval-Wild: Evaluating Hallucinations of Language Models in the Wild | Mar 7, 2024 | HallucinationQuestion Answering | CodeCode Available | 0 |
| Federated Recommendation via Hybrid Retrieval Augmented Generation | Mar 7, 2024 | HallucinationPrivacy Preserving | CodeCode Available | 1 |