| A Novel Approach to Eliminating Hallucinations in Large Language Model-Assisted Causal Discovery | Nov 16, 2024 | Causal DiscoveryHallucination | —Unverified | 0 |
| ViBe: A Text-to-Video Benchmark for Evaluating Hallucination in Large Multimodal Models | Nov 16, 2024 | HallucinationVideo Generation | —Unverified | 0 |
| Chain-of-Programming (CoP) : Empowering Large Language Models for Geospatial Code Generation | Nov 16, 2024 | Code GenerationData Visualization | —Unverified | 0 |
| Thinking Before Looking: Improving Multimodal LLM Reasoning via Mitigating Visual Hallucination | Nov 15, 2024 | HallucinationMultimodal Reasoning | CodeCode Available | 1 |
| Mitigating Hallucination in Multimodal Large Language Model via Hallucination-targeted Direct Preference Optimization | Nov 15, 2024 | HallucinationHallucination Evaluation | —Unverified | 0 |
| Layer Importance and Hallucination Analysis in Large Language Models via Enhanced Activation Variance-Sparsity | Nov 15, 2024 | Contrastive LearningHallucination | —Unverified | 0 |
| Seeing Clearly by Layer Two: Enhancing Attention Heads to Alleviate Hallucination in LVLMs | Nov 15, 2024 | Hallucination | —Unverified | 0 |
| DAHL: Domain-specific Automated Hallucination Evaluation of Long-Form Text through a Benchmark Dataset in Biomedicine | Nov 14, 2024 | FormHallucination | CodeCode Available | 0 |
| LLM Hallucination Reasoning with Zero-shot Knowledge Test | Nov 14, 2024 | Hallucination | —Unverified | 0 |
| On the Limits of Language Generation: Trade-Offs Between Hallucination and Mode Collapse | Nov 14, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| Bridging the Visual Gap: Fine-Tuning Multimodal Models with Knowledge-Adapted Captions | Nov 13, 2024 | DescriptiveHallucination | CodeCode Available | 0 |
| Confidence-aware Denoised Fine-tuning of Off-the-shelf Models for Certified Robustness | Nov 13, 2024 | Adversarial RobustnessDenoising | CodeCode Available | 0 |
| SHARP: Unlocking Interactive Hallucination via Stance Transfer in Role-Playing Agents | Nov 12, 2024 | General KnowledgeHallucination | —Unverified | 0 |
| DecoPrompt : Decoding Prompts Reduces Hallucinations when Large Language Models Meet False Premises | Nov 12, 2024 | Hallucination | CodeCode Available | 0 |
| Trustful LLMs: Customizing and Grounding Text Generation with Knowledge Bases and Dual Decoders | Nov 12, 2024 | DecoderHallucination | —Unverified | 0 |
| Verbosity Veracity: Demystify Verbosity Compensation Behavior of Large Language Models | Nov 12, 2024 | Hallucination | CodeCode Available | 0 |
| Invar-RAG: Invariant LLM-aligned Retrieval for Better Generation | Nov 11, 2024 | HallucinationInformation Retrieval | —Unverified | 0 |
| AssistRAG: Boosting the Potential of Large Language Models with an Intelligent Information Assistant | Nov 11, 2024 | Decision MakingHallucination | CodeCode Available | 1 |
| Evaluating the Accuracy of Chatbots in Financial Literature | Nov 11, 2024 | ChatbotHallucination | —Unverified | 0 |
| Prompt-Efficient Fine-Tuning for GPT-like Deep Models to Reduce Hallucination and to Improve Reproducibility in Scientific Text Generation Using Stochastic Optimisation Techniques | Nov 10, 2024 | Hallucinationparameter-efficient fine-tuning | —Unverified | 0 |
| Mitigating Hallucination with ZeroG: An Advanced Knowledge Management Engine | Nov 8, 2024 | Computational EfficiencyHallucination | —Unverified | 0 |
| Seeing Through the Fog: A Cost-Effectiveness Analysis of Hallucination Detection Systems | Nov 8, 2024 | DiagnosticHallucination | —Unverified | 0 |
| AMSnet-KG: A Netlist Dataset for LLM-based AMS Circuit Auto-Design Using Knowledge Graph RAG | Nov 7, 2024 | Bayesian OptimizationHallucination | —Unverified | 0 |
| Prompt-Guided Internal States for Hallucination Detection of Large Language Models | Nov 7, 2024 | Domain GeneralizationHallucination | —Unverified | 0 |
| LLM-R: A Framework for Domain-Adaptive Maintenance Scheme Generation Combining Hierarchical Agents and RAG | Nov 7, 2024 | HallucinationRAG | —Unverified | 0 |
| H-POPE: Hierarchical Polling-based Probing Evaluation of Hallucinations in Large Vision-Language Models | Nov 6, 2024 | HallucinationObject | —Unverified | 0 |
| Fine-Tuning Vision-Language Model for Automated Engineering Drawing Information Extraction | Nov 6, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| Fine-Grained Guidance for Retrievers: Leveraging LLMs' Feedback in Retrieval-Augmented Generation | Nov 6, 2024 | HallucinationRAG | —Unverified | 0 |
| Automated, LLM enabled extraction of synthesis details for reticular materials from scientific literature | Nov 5, 2024 | HallucinationIn-Context Learning | —Unverified | 0 |
| VERITAS: A Unified Approach to Reliability Evaluation | Nov 5, 2024 | Fact CheckingHallucination | —Unverified | 0 |
| Leveraging Vision-Language Models for Manufacturing Feature Recognition in CAD Designs | Nov 5, 2024 | Few-Shot LearningHallucination | —Unverified | 0 |
| V-DPO: Mitigating Hallucination in Large Vision Language Models via Vision-Guided Direct Preference Optimization | Nov 5, 2024 | HallucinationLanguage Modeling | CodeCode Available | 2 |
| Benchmarking Multimodal Retrieval Augmented Generation with Dynamic VQA Dataset and Self-adaptive Planning Agent | Nov 5, 2024 | BenchmarkingHallucination | CodeCode Available | 3 |
| DDFAV: Remote Sensing Large Vision Language Models Dataset and Evaluation Benchmark | Nov 5, 2024 | Data AugmentationHallucination | CodeCode Available | 0 |
| HtmlRAG: HTML is Better Than Plain Text for Modeling Retrieved Knowledge in RAG Systems | Nov 5, 2024 | HallucinationRAG | CodeCode Available | 3 |
| CleAR: Robust Context-Guided Generative Lighting Estimation for Mobile Augmented Reality | Nov 4, 2024 | HallucinationLighting Estimation | —Unverified | 0 |
| Robust plug-and-play methods for highly accelerated non-Cartesian MRI reconstruction | Nov 4, 2024 | compressed sensingDenoising | —Unverified | 0 |
| Improving Scientific Hypothesis Generation with Knowledge Grounded Large Language Models | Nov 4, 2024 | Experimental DesignHallucination | —Unverified | 0 |
| Rate, Explain and Cite (REC): Enhanced Explanation and Attribution in Automatic Evaluation by Large Language Models | Nov 3, 2024 | HallucinationInstruction Following | CodeCode Available | 0 |
| Towards Multi-Source Retrieval-Augmented Generation via Synergizing Reasoning and Preference-Driven Retrieval | Nov 1, 2024 | HallucinationRAG | —Unverified | 0 |
| RadFlag: A Black-Box Hallucination Detection Method for Medical Vision Language Models | Nov 1, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| Improbable Bigrams Expose Vulnerabilities of Incomplete Tokens in Byte-Level Tokenizers | Oct 31, 2024 | Hallucination | —Unverified | 0 |
| Exploring the Knowledge Mismatch Hypothesis: Hallucination Propensity in Small Models Fine-tuned on Data from Larger Models | Oct 31, 2024 | HallucinationMisinformation | —Unverified | 0 |
| EF-LLM: Energy Forecasting LLM with AI-assisted Automation, Enhanced Sparse Prediction, Hallucination Detection | Oct 30, 2024 | Continual LearningHallucination | —Unverified | 0 |
| Beyond Ontology in Dialogue State Tracking for Goal-Oriented Chatbot | Oct 30, 2024 | ChatbotDialogue State Tracking | CodeCode Available | 0 |
| VisAidMath: Benchmarking Visual-Aided Mathematical Reasoning | Oct 30, 2024 | BenchmarkingHallucination | —Unverified | 0 |
| Unified Triplet-Level Hallucination Evaluation for Large Vision-Language Models | Oct 30, 2024 | HallucinationHallucination Evaluation | CodeCode Available | 0 |
| Distinguishing Ignorance from Error in LLM Hallucinations | Oct 29, 2024 | HallucinationQuestion Answering | CodeCode Available | 1 |
| MARCO: Multi-Agent Real-time Chat Orchestration | Oct 29, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |
| FactBench: A Dynamic Benchmark for In-the-Wild Language Model Factuality Evaluation | Oct 29, 2024 | HallucinationLanguage Modeling | —Unverified | 0 |