| Chain of Natural Language Inference for Reducing Large Language Model Ungrounded Hallucinations | Oct 6, 2023 | HallucinationLanguage Modeling | CodeCode Available | 1 |
| Evaluating Hallucinations in Chinese Large Language Models | Oct 5, 2023 | HallucinationQuestion Answering | CodeCode Available | 3 |
| FreshLLMs: Refreshing Large Language Models with Search Engine Augmentation | Oct 5, 2023 | HallucinationWorld Knowledge | CodeCode Available | 2 |
| MLAgentBench: Evaluating Language Agents on Machine Learning Experimentation | Oct 5, 2023 | BenchmarkingDecision Making | CodeCode Available | 2 |
| AGIR: Automating Cyber Threat Intelligence Reporting with Natural Language Generation | Oct 4, 2023 | HallucinationText Generation | CodeCode Available | 1 |
| HallE-Control: Controlling Object Hallucination in Large Multimodal Models | Oct 3, 2023 | AttributeDecoder | CodeCode Available | 1 |
| LLM Lies: Hallucinations are not Bugs, but Features as Adversarial Examples | Oct 2, 2023 | Hallucination | CodeCode Available | 1 |
| BTR: Binary Token Representations for Efficient Retrieval Augmented Language Models | Oct 2, 2023 | HallucinationRetrieval | CodeCode Available | 1 |
| Analyzing and Mitigating Object Hallucination in Large Vision-Language Models | Oct 1, 2023 | HallucinationHallucination Evaluation | CodeCode Available | 1 |
| AutoHall: Automated Hallucination Dataset Generation for Large Language Models | Sep 30, 2023 | Dataset GenerationFact Checking | —Unverified | 0 |