| A Perspective for Adapting Generalist AI to Specialized Medical AI Applications and Their Challenges | Oct 28, 2024 | Drug DiscoveryHallucination | —Unverified | 0 |
| A Debate-Driven Experiment on LLM Hallucinations and Accuracy | Oct 25, 2024 | Fact CheckingHallucination | —Unverified | 0 |
| MaCTG: Multi-Agent Collaborative Thought Graph for Automatic Programming | Oct 25, 2024 | Code GenerationHallucination | —Unverified | 0 |
| Conditional Hallucinations for Image Compression | Oct 25, 2024 | HallucinationImage Compression | —Unverified | 0 |
| Investigating the Role of Prompting and External Tools in Hallucination Rates of Large Language Models | Oct 25, 2024 | HallucinationPrompt Engineering | —Unverified | 0 |
| TimeSuite: Improving MLLMs for Long Video Understanding via Grounded Tuning | Oct 25, 2024 | EgoSchemaHallucination | CodeCode Available | 2 |
| Multilingual Hallucination Gaps in Large Language Models | Oct 23, 2024 | HallucinationText Generation | —Unverified | 0 |
| Leveraging the Domain Adaptation of Retrieval Augmented Generation Models for Question Answering and Reducing Hallucination | Oct 23, 2024 | Domain AdaptationHallucination | —Unverified | 0 |
| AVHBench: A Cross-Modal Hallucination Benchmark for Audio-Visual Large Language Models | Oct 23, 2024 | Hallucination | —Unverified | 0 |
| ProveRAG: Provenance-Driven Vulnerability Analysis with Automated Retrieval-Augmented LLMs | Oct 22, 2024 | ChunkingHallucination | CodeCode Available | 0 |
| Privacy-hardened and hallucination-resistant synthetic data generation with logic-solvers | Oct 22, 2024 | Generative Adversarial NetworkHallucination | —Unverified | 0 |
| Do Robot Snakes Dream like Electric Sheep? Investigating the Effects of Architectural Inductive Biases on Hallucination | Oct 22, 2024 | Hallucination | —Unverified | 0 |
| GeoCode-GPT: A Large Language Model for Geospatial Code Generation Tasks | Oct 22, 2024 | Code GenerationCode Summarization | —Unverified | 0 |
| IPL: Leveraging Multimodal Large Language Models for Intelligent Product Listing | Oct 22, 2024 | HallucinationRAG | —Unverified | 0 |
| Navigating Noisy Feedback: Enhancing Reinforcement Learning with Error-Prone Language Models | Oct 22, 2024 | HallucinationLanguage Modeling | CodeCode Available | 0 |
| SG-FSM: A Self-Guiding Zero-Shot Prompting Paradigm for Multi-Hop Question Answering Based on Finite State Machine | Oct 22, 2024 | HallucinationMulti-hop Question Answering | —Unverified | 0 |
| Fine-Tuning Large Language Models to Appropriately Abstain with Semantic Entropy | Oct 22, 2024 | FormHallucination | —Unverified | 0 |
| Towards a Reliable Offline Personal AI Assistant for Long Duration Spaceflight | Oct 21, 2024 | HallucinationKnowledge Graphs | —Unverified | 0 |
| Large language models enabled multiagent ensemble method for efficient EHR data labeling | Oct 21, 2024 | Hallucination | —Unverified | 0 |
| ToW: Thoughts of Words Improve Reasoning in Large Language Models | Oct 21, 2024 | Data AugmentationHallucination | CodeCode Available | 0 |
| Mitigating Object Hallucination via Concentric Causal Attention | Oct 21, 2024 | HallucinationObject | CodeCode Available | 2 |
| Mitigating Hallucinations of Large Language Models in Medical Information Extraction via Contrastive Decoding | Oct 21, 2024 | Hallucination | —Unverified | 0 |
| Learning to Generate and Evaluate Fact-checking Explanations with Transformers | Oct 21, 2024 | Fact CheckingHallucination | —Unverified | 0 |
| Can Knowledge Editing Really Correct Hallucinations? | Oct 21, 2024 | Hallucinationknowledge editing | CodeCode Available | 1 |
| NetSafe: Exploring the Topological Safety of Multi-agent Networks | Oct 21, 2024 | HallucinationMisinformation | —Unverified | 0 |