| Circuit Transformer: A Transformer That Preserves Logical Equivalence | Mar 14, 2024 | Hallucination | CodeCode Available | 1 |
| The First to Know: How Token Distributions Reveal Hidden Knowledge in Large Vision-Language Models? | Mar 14, 2024 | Hallucinationimage-classification | CodeCode Available | 1 |
| Federated Recommendation via Hybrid Retrieval Augmented Generation | Mar 7, 2024 | HallucinationPrivacy Preserving | CodeCode Available | 1 |
| InterrogateLLM: Zero-Resource Hallucination Detection in LLM-Generated Answers | Mar 5, 2024 | Hallucination | CodeCode Available | 1 |
| CR-LT-KGQA: A Knowledge Graph Question Answering Dataset Requiring Commonsense Reasoning and Long-Tail Knowledge | Mar 3, 2024 | Claim VerificationGraph Question Answering | CodeCode Available | 1 |
| DiaHalu: A Dialogue-level Hallucination Evaluation Benchmark for Large Language Models | Mar 1, 2024 | HallucinationHallucination Evaluation | CodeCode Available | 1 |
| All in an Aggregated Image for In-Image Learning | Feb 28, 2024 | AllHallucination | CodeCode Available | 1 |
| Detecting Machine-Generated Texts by Multi-Population Aware Optimization for Maximum Mean Discrepancy | Feb 25, 2024 | HallucinationSentence | CodeCode Available | 1 |
| Citation-Enhanced Generation for LLM-based Chatbots | Feb 25, 2024 | ChatbotCitation Prediction | CodeCode Available | 1 |
| A Data-Centric Approach To Generate Faithful and High Quality Patient Summaries with Large Language Models | Feb 23, 2024 | Hallucination | CodeCode Available | 1 |
| Seeing is Believing: Mitigating Hallucination in Large Vision-Language Models via CLIP-Guided Decoding | Feb 23, 2024 | HallucinationObject | CodeCode Available | 1 |
| Visual Hallucinations of Multi-modal Large Language Models | Feb 22, 2024 | DiversityHallucination | CodeCode Available | 1 |
| TofuEval: Evaluating Hallucinations of LLMs on Topic-Focused Dialogue Summarization | Feb 20, 2024 | HallucinationNews Summarization | CodeCode Available | 1 |
| Logical Closed Loop: Uncovering Object Hallucinations in Large Vision-Language Models | Feb 18, 2024 | HallucinationObject | CodeCode Available | 1 |
| EFUF: Efficient Fine-grained Unlearning Framework for Mitigating Hallucinations in Multimodal Large Language Models | Feb 15, 2024 | HallucinationObject Hallucination | CodeCode Available | 1 |
| Uncertainty Quantification for In-Context Learning of Large Language Models | Feb 15, 2024 | HallucinationIn-Context Learning | CodeCode Available | 1 |
| Into the Unknown: Self-Learning Large Language Models | Feb 14, 2024 | HallucinationSelf-Learning | CodeCode Available | 1 |
| Gemini Goes to Med School: Exploring the Capabilities of Multimodal Large Language Models on Medical Challenge Problems & Hallucinations | Feb 10, 2024 | DiagnosticHallucination | CodeCode Available | 1 |
| Introspective Planning: Aligning Robots' Uncertainty with Inherent Task Ambiguity | Feb 9, 2024 | Conformal PredictionHallucination | CodeCode Available | 1 |
| Training Language Models to Generate Text with Citations via Fine-grained Rewards | Feb 6, 2024 | HallucinationQuestion Answering | CodeCode Available | 1 |
| INSIDE: LLMs' Internal States Retain the Power of Hallucination Detection | Feb 6, 2024 | DiversityHallucination | CodeCode Available | 1 |
| Unified Hallucination Detection for Multimodal Large Language Models | Feb 5, 2024 | Hallucination | CodeCode Available | 1 |
| Skip : A Simple Method to Reduce Hallucination in Large Vision-Language Models | Feb 2, 2024 | Hallucination | CodeCode Available | 1 |
| K-QA: A Real-World Medical Q&A Benchmark | Jan 25, 2024 | HallucinationIn-Context Learning | CodeCode Available | 1 |
| How well can a large language model explain business processes as perceived by users? | Jan 23, 2024 | HallucinationLanguage Modeling | CodeCode Available | 1 |