| Multi-Reference Preference Optimization for Large Language Models | May 26, 2024 | GSM8KTruthfulQA | —Unverified | 0 |
| Harmonic LLMs are Trustworthy | Apr 30, 2024 | HallucinationTruthfulQA | —Unverified | 0 |
| Student Data Paradox and Curious Case of Single Student-Tutor Model: Regressive Side Effects of Training LLMs for Personalized Learning | Apr 23, 2024 | ARCCommon Sense Reasoning | —Unverified | 0 |
| When Hindsight is Not 20/20: Testing Limits on Reflective Thinking in Large Language Models | Apr 14, 2024 | TruthfulQA | CodeCode Available | 0 |
| PoLLMgraph: Unraveling Hallucinations in Large Language Models via State Transition Dynamics | Apr 6, 2024 | BenchmarkingHallucination | CodeCode Available | 0 |
| PRobELM: Plausibility Ranking Evaluation for Language Models | Apr 4, 2024 | Question AnsweringTruthfulQA | —Unverified | 0 |
| SaGE: Evaluating Moral Consistency in Large Language Models | Feb 21, 2024 | Decision MakingHellaSwag | CodeCode Available | 0 |
| Self-Alignment for Factuality: Mitigating Hallucinations in LLMs via Self-Evaluation | Feb 14, 2024 | TruthfulQA | —Unverified | 0 |
| LLMAuditor: A Framework for Auditing Large Language Models Using Human-in-the-Loop | Feb 14, 2024 | HallucinationTruthfulQA | —Unverified | 0 |
| GRATH: Gradual Self-Truthifying for Large Language Models | Jan 22, 2024 | TruthfulQA | —Unverified | 0 |