| Are Large Language Models Good at Utility Judgments? | Mar 28, 2024 | Answer GenerationBenchmarking | CodeCode Available | 0 |
| FACTOID: FACtual enTailment fOr hallucInation Detection | Mar 28, 2024 | AvgHallucination | —Unverified | 0 |
| Rejection Improves Reliability: Training LLMs to Refuse Unknown Questions Using RL from Knowledge Feedback | Mar 27, 2024 | Hallucination | —Unverified | 0 |
| Mechanistic Understanding and Mitigation of Language Model Non-Factual Hallucinations | Mar 27, 2024 | AttributeDiagnostic | CodeCode Available | 0 |
| "Sorry, Come Again?" Prompting -- Enhancing Comprehension and Diminishing Hallucination with [PAUSE]-injected Optimal Paraphrasing | Mar 27, 2024 | Hallucination | —Unverified | 0 |
| Chain-of-Action: Faithful and Multimodal Question Answering through Large Language Models | Mar 26, 2024 | HallucinationInformation Retrieval | CodeCode Available | 0 |
| DGoT: Dynamic Graph of Thoughts for Scientific Abstract Generation | Mar 26, 2024 | Abstract generationHallucination | CodeCode Available | 0 |
| Visual Hallucination: Definition, Quantification, and Prescriptive Remediations | Mar 26, 2024 | HallucinationImage Captioning | —Unverified | 0 |
| Dyna-LfLH: Learning Agile Navigation in Dynamic Environments from Learned Hallucination | Mar 25, 2024 | HallucinationImitation Learning | —Unverified | 0 |
| Hallucination Detection in Foundation Models for Decision-Making: A Flexible Definition and Review of the State of the Art | Mar 25, 2024 | Common Sense ReasoningDecision Making | —Unverified | 0 |