| Safety challenges of AI in medicine in the era of large language models | Sep 11, 2024 | Hallucination | —Unverified | 0 |
| MEDIC: Towards a Comprehensive Framework for Evaluating LLMs in Clinical Applications | Sep 11, 2024 | EthicsHallucination | —Unverified | 0 |
| Mitigating Hallucination in Visual-Language Models via Re-Balancing Contrastive Decoding | Sep 10, 2024 | HallucinationImage Captioning | —Unverified | 0 |
| LLMs Will Always Hallucinate, and We Need to Live With This | Sep 9, 2024 | Fact CheckingHallucination | —Unverified | 0 |
| Detecting Buggy Contracts via Smart Testing | Sep 6, 2024 | Hallucination | —Unverified | 0 |
| Generating Faithful and Salient Text from Multimodal Data | Sep 6, 2024 | HallucinationKnowledge Graphs | CodeCode Available | 0 |
| Combining LLMs and Knowledge Graphs to Reduce Hallucinations in Question Answering | Sep 6, 2024 | HallucinationKnowledge Graphs | —Unverified | 0 |
| Vietnamese Legal Information Retrieval in Question-Answering System | Sep 5, 2024 | HallucinationInformation Retrieval | —Unverified | 0 |
| Hallucination Detection in LLMs: Fast and Memory-Efficient Fine-Tuned Models | Sep 4, 2024 | GPUHallucination | CodeCode Available | 0 |
| CLUE: Concept-Level Uncertainty Estimation for Large Language Models | Sep 4, 2024 | HallucinationSentence | —Unverified | 0 |
| Improved Single Camera BEV Perception Using Multi-Camera Training | Sep 4, 2024 | Autonomous DrivingHallucination | —Unverified | 0 |
| Multi-Source Knowledge Pruning for Retrieval-Augmented Generation: A Benchmark and Empirical Study | Sep 3, 2024 | BenchmarkingHallucination | CodeCode Available | 0 |
| What does it take to get state of the art in simultaneous speech-to-speech translation? | Sep 2, 2024 | HallucinationManagement | —Unverified | 0 |
| Understanding Multimodal Hallucination with Parameter-Free Representation Alignment | Sep 2, 2024 | HallucinationObject | CodeCode Available | 0 |
| Towards Empathetic Conversational Recommender Systems | Aug 30, 2024 | HallucinationRecommendation Systems | CodeCode Available | 1 |
| LLMs Prompted for Graphs: Hallucinations and Generative Capabilities | Aug 30, 2024 | DiversityHallucination | —Unverified | 0 |
| Pre-Training Multimodal Hallucination Detectors with Corrupted Grounding Data | Aug 30, 2024 | HallucinationPhrase Grounding | —Unverified | 0 |
| UserSumBench: A Benchmark Framework for Evaluating User Summarization Approaches | Aug 30, 2024 | HallucinationRecommendation Systems | —Unverified | 0 |
| Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning | Aug 30, 2024 | Hallucination | CodeCode Available | 1 |
| VLM4Bio: A Benchmark Dataset to Evaluate Pretrained Vision-Language Models for Trait Discovery from Biological Images | Aug 28, 2024 | Hallucination | CodeCode Available | 0 |
| LLaVA-MoD: Making LLaVA Tiny via MoE Knowledge Distillation | Aug 28, 2024 | Computational EfficiencyHallucination | CodeCode Available | 3 |
| Negation Blindness in Large Language Models: Unveiling the NO Syndrome in Image Generation | Aug 27, 2024 | HallucinationImage Generation | —Unverified | 0 |
| Measuring text summarization factuality using atomic facts entailment metrics in the context of retrieval augmented generation | Aug 27, 2024 | HallucinationRetrieval-augmented Generation | —Unverified | 0 |
| Evidence-Enhanced Triplet Generation Framework for Hallucination Alleviation in Generative Question Answering | Aug 27, 2024 | Generative Question AnsweringHallucination | —Unverified | 0 |
| Genetic Approach to Mitigate Hallucination in Generative IR | Aug 25, 2024 | Answer GenerationHallucination | CodeCode Available | 0 |
| Towards Reliable Medical Question Answering: Techniques and Challenges in Mitigating Hallucinations in Language Models | Aug 25, 2024 | Decision MakingHallucination | —Unverified | 0 |
| ConVis: Contrastive Decoding with Hallucination Visualization for Mitigating Hallucinations in Multimodal Large Language Models | Aug 25, 2024 | Hallucination | CodeCode Available | 1 |
| Can LLM be a Good Path Planner based on Prompt Engineering? Mitigating the Hallucination for Path Planning | Aug 23, 2024 | HallucinationPrompt Engineering | —Unverified | 0 |
| Internal and External Knowledge Interactive Refinement Framework for Knowledge-Intensive Question Answering | Aug 23, 2024 | HallucinationQuestion Answering | —Unverified | 0 |
| SLM Meets LLM: Balancing Latency, Interpretability and Consistency in Hallucination Detection | Aug 22, 2024 | HallucinationLanguage Modeling | CodeCode Available | 1 |
| MedDiT: A Knowledge-Controlled Diffusion Transformer Framework for Dynamic Medical Image Generation in Virtual Simulated Patient | Aug 22, 2024 | DiagnosticHallucination | —Unverified | 0 |
| Improving Factuality in Large Language Models via Decoding-Time Hallucinatory and Truthful Comparators | Aug 22, 2024 | HallucinationMixture-of-Experts | CodeCode Available | 0 |
| RoVRM: A Robust Visual Reward Model Optimized via Auxiliary Textual Preference Data | Aug 22, 2024 | Hallucination | CodeCode Available | 0 |
| GRATR: Zero-Shot Evidence Graph Retrieval-Augmented Trustworthiness Reasoning | Aug 22, 2024 | Decision MakingHallucination | CodeCode Available | 0 |
| RAG-Optimized Tibetan Tourism LLMs: Enhancing Accuracy and Personalization | Aug 21, 2024 | HallucinationRAG | —Unverified | 0 |
| Towards Analyzing and Mitigating Sycophancy in Large Vision-Language Models | Aug 21, 2024 | HallucinationPrompt Engineering | —Unverified | 0 |
| Enhanced document retrieval with topic embeddings | Aug 19, 2024 | HallucinationRAG | —Unverified | 0 |
| MAPLE: Enhancing Review Generation with Multi-Aspect Prompt LEarning in Explainable Recommendation | Aug 19, 2024 | DiversityExplainable Recommendation | —Unverified | 0 |
| CLIP-DPO: Vision-Language Models as a Source of Preference for Fixing Hallucinations in LVLMs | Aug 19, 2024 | Hallucinationzero-shot-classification | —Unverified | 0 |
| Reefknot: A Comprehensive Benchmark for Relation Hallucination Evaluation, Analysis and Mitigation in Multimodal Large Language Models | Aug 18, 2024 | AttributeHallucination | CodeCode Available | 1 |
| Cognitive LLMs: Towards Integrating Cognitive Architectures and Large Language Models for Manufacturing Decision-making | Aug 17, 2024 | Decision MakingHallucination | —Unverified | 0 |
| Lower Layer Matters: Alleviating Hallucination via Multi-Layer Fusion Contrastive Decoding with Truthfulness Refocused | Aug 16, 2024 | HallucinationTruthfulQA | —Unverified | 0 |
| Large Language Models Might Not Care What You Are Saying: Prompt Format Beats Descriptions | Aug 16, 2024 | DescriptiveHallucination | —Unverified | 0 |
| Graph Retrieval-Augmented Generation: A Survey | Aug 15, 2024 | HallucinationRAG | CodeCode Available | 3 |
| Plan with Code: Comparing approaches for robust NL to DSL generation | Aug 15, 2024 | Code GenerationHallucination | —Unverified | 0 |
| CodeMirage: Hallucinations in Code Generated by Large Language Models | Aug 14, 2024 | Code GenerationHallucination | —Unverified | 0 |
| Training Language Models on the Knowledge Graph: Insights on Hallucinations and Their Detectability | Aug 14, 2024 | Hallucination | —Unverified | 0 |
| Audit-LLM: Multi-Agent Collaboration for Log-based Insider Threat Detection | Aug 12, 2024 | Common Sense ReasoningHallucination | —Unverified | 0 |
| SSL: A Self-similarity Loss for Improving Generative Image Super-resolution | Aug 11, 2024 | HallucinationImage Super-Resolution | CodeCode Available | 2 |
| Reference-free Hallucination Detection for Large Vision-Language Models | Aug 11, 2024 | HallucinationQuestion Answering | —Unverified | 0 |